This is a very complicated topic.
First: Grok sucks, and is controlled by a fascist asshole. Full stop. I will not argue this.
Second: GitHub CoPilot is, as Mr. Roddy noted, almost decent as a code completion tool. But Microsoft have so badly brain damaged the models by forcing a 128K context window it makes AI look really bad. And the tools are so bad, they constantly make mistakes.
Third: The OSR team have spent a great deal of time on this, and are finding AI in general and Claude Code In particular, very helpful. Good prompt engineering is essential. Focused prompting, even for code reviews, can improve the depth of results. Writing code with one model and getting a code review from another model (even if a typically “inferior” one) seems to be very effective.
We have had luck with code reviews of mini-filter code. Like most static tools, you wade through a pile of false positives to fine one or two gems. But, find them we have.
In terms of actually generating new KM code for Windows: Not that much success. It seems to me that the models don’t have a big enough base on which to train. As OSR’s Scott Noone recently noted (words to the effect of): “The more I know about a topic, the less happy I am with the answers I get from an AI.” In short, based on my personal experience and the experience of the whole OSR team over the past six months, I would absolutely not trust AI to author my kernel mode code that needed to work.
OTOH, we have had repeated and spectacular luck with having Claude Code generate user mode utilities, and test cases, in Modern C++ 20. I mean, complete semi-complex utilities in domains where none of us know a single thing. As an example, we wrote a set of utilities to build software kits (think of a crazy powerful modern version of binplace), that includes moving files around, signing executables, building driver packages, uploading the package to the hardware dashboard for attestation signing, tracking the submission’s progress, downloading the results, and kitting the resulting signed driver binaries into a zip archive. One command at the command line, go get coffee, come back and release. 90% AI designed, built, tested (including test framework designed and built and test cases created). Yes, it took engineer oversight and guidance, but… just really impressive results, in a domain where folks here know sweet fuck all.
I could go on about this at great length, and already have. But bottom line: Lots of uses for AI in our world. But not yet for actually authoring KM code for Windows. No way, no how, not even close when you really look closely at the code that’s been written.