

Microsoft CEO Says Up To 30% of the Company's Code Was Written by AI (techcrunch.com) 126
Microsoft CEO Satya Nadella said that 20%-30% of code inside the company's repositories was "written by software" -- meaning AI -- during a fireside chat with Meta CEO Mark Zuckerberg at Meta's LlamaCon conference on Tuesday. From a report: Nadella gave the figure after Zuckerberg asked roughly how much of Microsoft's code is AI-generated today. The Microsoft CEO said the company was seeing mixed results in AI-generated code across different languages, with more progress in Python and less in C++.
BS (Score:5, Interesting)
Re: (Score:2)
Re:BS (Score:5, Insightful)
Because I've tried using LLMs to generate code and I've seen the results. They are not usable. They *resemble* valid code, but they typically throw exceptions and raise errors, they can't pass unit tests, and they don't correctly handle edge cases. AI-generated code is a mess that *superficially looks right* but isn't fit to purpose.
There is a meme going around about the fact that you can tackle a normal coding task by spending 3 hours to write code and 1 hour to debug and test it, or you can use CoPilot to spend 15 minutes to write the code and 8 hours to debug and test it. That matches my experience.
Re: BS (Score:2, Insightful)
It's possible you're just doing it wrong.
Re: (Score:2)
Re: BS (Score:5, Funny)
Re: (Score:2)
Given, its tough to maintain such bloat ware that should be just a simple OS and not integrated with terrible options like web browsing in your fucking file system, or self masturbatory "AI" disguised as spyware by the Microsoft that proports to help y
Re: (Score:2)
FWIW, I don't particularly disagree with you.
Re: (Score:2)
Re: (Score:2)
I, know what a comma, is used for do, you?
Re: (Score:2)
Re: (Score:2)
I know right! The other one that drives me nut's is misused apostrophe's. :-)
Re: (Score:2)
Is "detection of a misplaced comma" something it can even handle?
If that is something it's supposed to do, and did not do here, then it is broken.
Re: (Score:3, Informative)
>> They are not usable.
Maybe you just aren't doing it right. I use LLM's to generate code for me every day and usually it is entirely adequate. The AI does require some handholding at times. There is a learning curve for working with it effectively but once you figure out what it can and can't do it is a brute.
Meanwhile, Nadella said that 20%-30% of code inside the company's repositories was "written by software" and that seems entirely possible.
Re: (Score:2, Insightful)
If you think that AI is doing a good job writing code, you probably shouldn't be writing code. You're clearly not qualified.
Re: (Score:2)
Nobody said it had to be good. Only more cost effective.
Re: (Score:2)
you probably shouldn't be writing code. You're clearly not qualified.
Being unqualified never stopped anyone else, why would it stop him?
That's one of the best and the worst things about coding, any fool with a PC can do it. Many of those that keep doing it, eventually get better at it.
Re: (Score:2)
Meanwhile, Nadella said that 20%-30% of code inside the company's repositories was "written by software" and that seems entirely possible.
The statement "20%-30% of code inside the company’s repositories was 'written by software'" was a response to the question of "how much of Microsoft’s code is AI-generated today." I assume that not all of Microsoft's repositories were created today, so the statement about repositories cannot possibly be true. I'm guessing that the article writer incorrectly paraphrased what Nadella said. Then again, I suppose it's possible that maybe Microsoft has been secretly using AI to generate code for t
Re: (Score:2)
Good point, probably he was talking about the proportion of code being generated by AI today.
Re: (Score:2)
I'm going to take a view in between the "AI can do it all", versus "AI can't do it".
As of now, perhaps it may get better over time, if I ask a chatbot to make some code for me, I'll get it.
However, I then have to debug it, unit test it, change stuff to match the coding conventions of everything else. This might be minimal, or I might have spent more time fixing AI code than if I wrote things from scratch.
So, it is sort of like OCR was in the 1990s... it sort of helps, but you have to go through it with a f
Re: BS (Score:3, Insightful)
Not denying your experience at all, but I've had a great experience using Claude to do the busy-work of "open a serial port on any of 3 platforms", or "set up the basics of a tkinter GUI" in Python, such that it let me get to the real work of manipulating data, addressing the device on the serial port, designing the UI much quicker.
I wouldn't hand an AI the whole task, at least not yet. But the thankless grunt work? Yes please. Away with it.
The one time I had trouble with Claude was when every suggestion ra
Re: BS (Score:2)
Re: (Score:2)
I've had a great experience using Claude to do the busy-work of "open a serial port on any of 3 platforms", or "set up the basics of a tkinter GUI" in Python, such that it let me get to the real work of manipulating data, addressing the device on the serial port, designing the UI much quicker.
I wouldn't hand an AI the whole task, at least not yet. But the thankless grunt work? Yes please. Away with it.
AI can be useful at cutting and pasting that kind of boilerplate code that it has seen on Stackoverflow many times. It will still mess it up and is not neatly as reliable as cutting and pasting it yourself.
I have found that it cannot write trivial SQL queries that work. The code references table and columns that do not even exist. Looking it over and trying to debug it takes longer than writing it myself.
Whenever someone tells me how helpful the AI is at writing code for them, all I hear is: "I can't progr
Re: BS (Score:2)
I can write, and have written both the examples I provided here before. Claude did a great job, and provided clear, legible working code.
These were little jobs, not big ones. I wouldn't try having it land a plane or prescribe meds...well, at all right now. Maybe in future with much human supervision and review.
It's also important to keep in mind that until very recently, ALL programming errors with catastrophic results came from humans, not AI.
We're not uniformly good, and not at all foolproof.
Re: (Score:2)
This is also my experience. AI can do the simplest of tasks - poorly. Generating a Makefile or a Maven pom.xml file, sure. I was struggling to get simple unit test cases.
I was given a defect to fix. I found the problem quick enough. So I asked the AI.... It was completely useless.
Re: (Score:2)
HOW you ask the AI tool is a very large part of the result validity.
Re: (Score:2)
Well, if I hold its hand long enough, it will eventually get the correct answer. Along the way, it will very confidently provide an incorrect answer.
Re: (Score:2)
Re: (Score:2)
I've done some experiments recently with LLM-backed tools to try to understand the current state of the art. FWIW, my own experience has been that for relatively simple boilerplate-generation jobs they can often produce useful code, but their limit is roughly the capabilities of a junior developer. They make mistakes fairly often. Maybe more importantly, even when their code technically produces the right answer, they rarely infer much about any existing design or coding standards and their code often doesn
Re: (Score:2)
With experience, you definitely acquire a "feeling" that continuing to engage with it on this particular problem is not going to be fruitful. It took me a few wasted hours to figure that out.
They do make mistakes fairly often (but oddly, are usually highly capable of fixing them when they're pointed out, either by you, or a compiler).
As for their ability to infer, I couldn't agree less with that.
Dump an entire code base into
Re: (Score:2)
Sounds like a perfectly good description of Windows code to me!
Re: (Score:2)
reposited code doesn't not necessarily mean actually being used.
Re: (Score:2)
I have been working with various LLM solutions, as well as more specialized coding solutions being developed for certain specialized cases.
They are not perfect, not by a long shot, but they yield pretty nice code, if properly prompted to do so.
Prompting is an art, and it becomes more complex as time passes. There are already attempts to develop AI tools which generates complex prompts for, well, other AI tools.
Re: (Score:2)
There are already attempts to develop AI tools which generates complex prompts for, well, other AI tools.
Attempts? This is the magic sauce of all agentic coding platforms. We're well past attempts- we have paid products.
Re: (Score:2)
Well, yes, but they are not always successful.
Re: (Score:2)
Do you know why you have to route all traffic through them, even when the end-model is something you could access directly like ChatGPT or Claude? So that they can hide the actual prompt engineering they do from you. That's their proprietary sauce- the thing they're actually selling, because as you noted- prompt engineering is 99.9% of the game with getting really good stuff out of LLMs. An amateur can fumble around, and rely on the LLM to fill in all the gaps
Re: (Score:2)
Re: (Score:2)
Makes sense. I have been using Cursor, but to be frank, I never had thought about their reprompting :)
Re: (Score:3)
Because I've tried using LLMs to generate code and I've seen the results. They are not usable. They *resemble* valid code, but they typically throw exceptions and raise errors, they can't pass unit tests, and they don't correctly handle edge cases. AI-generated code is a mess that *superficially looks right* but isn't fit to purpose.
There is a meme going around about the fact that you can tackle a normal coding task by spending 3 hours to write code and 1 hour to debug and test it, or you can use CoPilot to spend 15 minutes to write the code and 8 hours to debug and test it. That matches my experience.
This has been my experience for long code snippets. If you give it a very specific function you need, sometimes it can produce something valid quickly. The problem is trying to break down large tasks into very small functions so that the AI generator can actually handle the request. Sometimes it's possible, sometimes user requirements are so wildly speculative to begin with you really can't break it down to small functions playing against each other.
Re: (Score:3)
From my experience, it's a question of scope and how detailed your prompt is. If you ask "write me a program in X that does Y" you aren't going to get nearly as good of an answer as "write me a function in X that takes A, B, C as inputs; and gives Y as the return"
And yes, just like if you were doing a code review of a junior engineer, don't rubberstamp what it gives you back. And if what it gives you back is garbage, oh no you've sadly wasted 30 seconds and can always fall back on writing it yourself.
Re: (Score:2)
AI doesn't just mean LMM. If its something as broad as the autogenerated code when you create a project in visual studio than that 20-30% could easily be true.
AI is like reference books ... (Score:2)
Because I've tried using LLMs to generate code and I've seen the results. They are not usable. They *resemble* valid code, but they typically throw exceptions and raise errors, they can't pass unit tests, and they don't correctly handle edge cases. AI-generated code is a mess that *superficially looks right* but isn't fit to purpose.
That's what the other 70% of the code is for, the human written code.
There is a meme going around about the fact that you can tackle a normal coding task by spending 3 hours to write code and 1 hour to debug and test it, or you can use CoPilot to spend 15 minutes to write the code and 8 hours to debug and test it. That matches my experience.
It depends entirely on the complexity of the code you asked for. For simple things, like very short code samples that might appear in a textbook or reference book, the state of the art seems to be about equal to these references. Note these references also lack the defensive programming you mention.
AI is, currently, more of an alternative to reference books than programmers. And reference books have always provided code to real work pro
Re: BS (Score:2)
I just used an AI for re-skilling and it's helped immensely with a project of mine.
With a bit of skill and the right direction it's not overly difficult to get good results.
YMMV.
Re: (Score:2)
Your entire post is bullshit.
Everyone one of your posts is bullshit. Listen to yourself.
Re: (Score:2)
20-30% of code in the repositories. That is 20% to 30% of the code they are shipping, or even code the represents a branch anyone is doing work on that will ever land in a deliverable.
Could be just a bunch of devs branching stuff to try and quick experiment in and then forgetting about entirely 10min later.
LLMs are good at generating a lot of code, and version control systems are good at storing it; none of this suggests a large portion of it isn't useless garbage.
Re: (Score:2)
I think there is a strange kind of delusion that exists among the people that frequent this site that LLM usage equates to "vibe coding" or whatever the fuck they call that shit today where people who have no fucking idea what they're doing generate programs with LLMs.
LLMs are becoming more and more integrated into actual professional developer workflows, and they're quite good at what they do in that capacity.
Re:BS (Score:5, Funny)
Re: (Score:2)
Re: (Score:2)
MS still has a shitton of legacy code in their repos.
"AI" doesn't explain Windows 8, Windows ME, Windows CE, Windows Vista, the Office Ribbon, or Microsoft Bob.
Re: BS (Score:2)
That was RAI (resident alien intelligence)
Re: (Score:2)
Seems plausible, given what shit their software is.
Re: (Score:3)
Dodgy wording - "written by software" could just mean boilerplate code generation that uses no LLM at all. It sounds impressive as long as you don't think about it.
Re: (Score:2)
It sounds impressive as long as you don't think about it.
That sounds like most of silicon valley for the last 5 years.
Re: (Score:2)
written by software
He means an editor software, like a text editor? Or he's attributing any usage of auto-complete, while someone is writing code? Or is he counting template code, where the compiler emits code based on a template definition? Maybe he is summing it all up and making it sound like AI?
Re: (Score:2)
Re: (Score:2)
Then it's 100% code at MS is written using text editors (software), and then 70% is thrown away because it was prototype-PoC-ware (doesn't make it into the repositories). That's what he really meant.
the hype train goes "WOO WOO"! (Score:2)
How much of that is actually used? And correct?
In a generation, kids won't believe you when you tell them we used computers to do calculations instead of hallucinate rationalizations for the decrease in the chocolate ration.
Re: (Score:2)
You're talking out your ass, so it's like you're engaging in a PR war against the big bad LLM specter. To what end? Self delusion, or the hope of misinforming people that happen by your post?
Re: (Score:2)
It could also mean boilerplate code generation that uses an LLM. They're pretty good at that.
Dear Copilot, somebody wants another damn web form. Make it just like this existing one except change the label to "Fire idiot?" and the submission address to http://hr.microsoft.com/fireid... [microsoft.com].
Tnks.
And there you go, 1000 lines of code.
Re: (Score:2)
It depends upon how it is measured. If it is lines of code, I can easily imagine that AI spits out huge amounts of low quality code. There isn't any comment about what percent actually passes into production... just how much is in their repositories. You can always lie with statistics. CEO are masters at it.
Re: (Score:2)
AI less capable of copy-paste than human? (Score:2)
Does he really believe this?
What makes you think an AI is any less capable of copy-paste'ing code from the internet than a modern human programmer?
Re: (Score:2)
Does he really believe this?
What makes you think an AI is any less capable of copy-paste'ing code from the internet than a modern human programmer?
Because the human programmer understands what they are reading, and the AI is just stringing shit together based on statistical likelihood of one meaningless (to it) token following another?
Unless you mean to suggest that most "coders" don't actually understand what they are reading. In which case I think you might be right. In the early 2000s I worked with "coders" (mainly outsourced from India) that were like that. We had to rewrite pretty much everything they did. Every day.
Re: BS (Score:2)
Re: (Score:2)
Probably, The guy is definitely not competent with regards to technology ...
Leap (Score:5, Insightful)
You may be putting words in his mouth by assuming "AI" when he said "software".
Open up Visual studio and start drawing out a GUI app and a ton of the background code is then generated by software. Its been that way for ~30 years. That is was written by software doesn't necessarily mean it was AI, unless we're jumping on the train of calling basically everything in the computer "AI" these days.
Re: (Score:3)
"basically everything in the computer "AI" these days" - this seems to be the narrative these companies are pushing doesn't it. Another buzzword for bullshit as for as I'm concerned.
Re: (Score:2)
If they convince you that their crap, hallucinating AI today is "just like what we've been using for years, only better," maybe people will spend money on their crap, hallucinating AI that they're pushing today.
It's an admission of failure, when they push their new crap by saying it's more of the same, instead of the newest, bestest thing since sliced bread.
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
... everything began with a small "e".
Re: (Score:2)
Re: (Score:3)
"It's all ball bearings these days." -Fletch
Re: (Score:2)
And before that everything was "blockchain"
And before that everything was "enterprise"
I'm waiting for them to release the Blockchain-enabled Web-Scale Enterprise AI Agent Vibe Grid -aaS", with Quantum Guardeails to prevent incorrect code.
Until then I'll just write it myself TYVM.
Re: (Score:2)
Re: (Score:3)
Re: (Score:2)
You may be putting words in his mouth by assuming "AI" when he said "software".
Open up Visual studio and start drawing out a GUI app and a ton of the background code is then generated by software. Its been that way for ~30 years. That is was written by software doesn't necessarily mean it was AI, unless we're jumping on the train of calling basically everything in the computer "AI" these days.
You've hit the nail on the head there with the last sentence.
Up to 30% (Score:5, Informative)
But realistically there's no way anything like 30% of the lines of code were written by GenAI. The vast majority of Windows Server 2025's code is identical to Server 2022, 2019, and 2016. Win11 24H2 is almost identical to Win10 1501. At least when you consider individual lines.
Now, it's possible that 30% of the titles/programs they offer have been touched in one way or another by GenAI. I'm sure there's some stuff in Office and Windows and Exchange Online that was spewed by CoPilot.
That said, I think the pride here is... misplaced. Want me to buy AI? Tell me it's reliable. I get that. Thats what he's doing. Want me to buy everything else? Tell me it's reliable... which GenAI code is not. So... this announcement disincentivizes me to buy MS products except CoPilot, which I don't want in the first place. It's kind of "we've smeared shit sandwich on all of the non-shit-sandwich meals at our restaurant. It's that good!"
Re: (Score:3)
That includes 0.000000000000000001% But realistically there's no way anything like 30% of the lines of code were written by GenAI. The vast majority of Windows Server 2025's code is identical to Server 2022, 2019, and 2016. Win11 24H2 is almost identical to Win10 1501. At least when you consider individual lines. Now, it's possible that 30% of the titles/programs they offer have been touched in one way or another by GenAI. I'm sure there's some stuff in Office and Windows and Exchange Online that was spewed by CoPilot. That said, I think the pride here is... misplaced. Want me to buy AI? Tell me it's reliable. I get that. Thats what he's doing. Want me to buy everything else? Tell me it's reliable... which GenAI code is not. So... this announcement disincentivizes me to buy MS products except CoPilot, which I don't want in the first place. It's kind of "we've smeared shit sandwich on all of the non-shit-sandwich meals at our restaurant. It's that good!"
It could be as simple as AI can generating a *LOT* of code very, very quickly. I could believe 30% of code in a tech company that's created is created by AI. That doesn't mean 30% of code making it to release candidates is generated by AI.
Re: (Score:2)
It's kind of cute you think Windows would constitute a significant portion of their code footprint.
Re: (Score:2)
It's kind of cute you think Windows would constitute a significant portion of their code footprint.
Microsoft says there are 50 million lines of code in Windows 10, and it's guaranteed there are more in Windows 11. They also say there are 45-50 million lines of code in Office. These are believed to be their two largest codebases. You don't think 50 million+ lines, and being their single biggest codebase, is significant?
Re: Up to 30% (Score:2)
That's not what he's saying. The wording is terrible. He's saying 30% of new code is software written. Zuck was asking about how much of their code is CURRENTLY being written by software
Re: (Score:2)
That includes 0.000000000000000001%
"
You May Already Have Won Up To 1,000,000 Commits!*
*Hallucinations may vary by state.
Yeah... (Score:5, Insightful)
That explains a lot about the current state of Windows.
and the result is... (Score:2)
Recall. Oops, nvm.
they've been testing it for awhile now (Score:2)
probably started with Windows ME
The Creature with Two Assholes (Score:5, Funny)
Microsoft CEO Satya Nadella said [...] during a fireside chat with Meta CEO Mark Zuckerberg
I don't see a fire in the picture in the article, although I'd like to.
Re: (Score:2)
Microsoft CEO Satya Nadella said [...] during a fireside chat with Meta CEO Mark Zuckerberg
I don't see a fire in the picture in the article, although I'd like to.
I'd like to see both of them on fire. Maybe we can toss on Matthew Prince [wikipedia.org] on as some extra kindling.
Re: (Score:2)
I don't see a fire in the picture in the article, although I'd like to.
Here ya go! [giphy.com]
That explains a lot (Score:5, Insightful)
Oh yeah me too (Score:2)
n/t
... that's not what that means (Score:3)
Interesting (Score:2)
Now we know (Score:4, Informative)
It's all AI generated CR@Pola!
Maybe... (Score:2)
AI wrote huge volumes of throw-away code? Just keep the engine churning, soon it will write a volume of code equal to 20-30%. That doesn't mean the code is actually used in something real.
It is 100% .. if he meant machine code. (Score:2)
Maybe he meant machine code? This is 100% written by software - by compilers and assemblers.
Does that mean that their copyrights are 30% fraud (Score:2)
AI product doesn't qualify for copyright, isn't that the legal standard.
Re: So They Don't Use Version Control? (Score:2)
That's not what he's saying. He's saying 30% of new code is written by software.
Re: (Score:2)
In a former position, well over 90% of the code I committed in a few projects was generated from automated processes. As an example, we had OpenAPI-ish documents describing the data structures on our network appliances. I consumed that and generated things like Ansible modules (Python) and statically typed API clients (C#).
"Written by software" takes a lot of forms and I'd bet my last dollar on the notion that the portion described above is almost entirely generated by automated processes.