Catch up on stories from the past week (and beyond) at the Slashdot story archive

 



Forgot your password?
typodupeerror
×
Programming AI Microsoft

Does GitHub Copilot Improve Code Quality? (github.blog) 76

Microsoft-owned GitHub published a blog post asking "Does GitHub Copilot improve code quality? Here's what the data says."

Its first paragraph includes statistics from past studies — that GitHub Copilot has helped developers code up to 55% faster, leaving 88% of developers feeling more "in the flow" and 85% feeling more confident in their code.

But does it improve code quality? [W]e recruited 202 [Python] developers with at least five years of experience. Half were randomly assigned GitHub Copilot access and the other half were instructed not to use any AI tools... We then evaluated the code with unit tests and with an expert review conducted by developers.

Our findings overall show that code authored with GitHub Copilot has increased functionality and improved readability, is of better quality, and receives higher approval rates... Developers with GitHub Copilot access had a 56% greater likelihood of passing all 10 unit tests in the study, indicating that GitHub Copilot helps developers write more functional code by a wide margin. In blind reviews, code written with GitHub Copilot had significantly fewer code readability errors, allowing developers to write 13.6% more lines of code, on average, without encountering readability problems. Readability improved by 3.62%, reliability by 2.94%, maintainability by 2.47%, and conciseness by 4.16%. All numbers were statistically significant... Developers were 5% more likely to approve code written with GitHub Copilot, meaning that such code is ready to be merged sooner, speeding up the time to fix bugs or deploy new features.

"While GitHub's reports have been positive, a few others haven't," reports Visual Studio magazine: For example, a recent study from Uplevel Data Labs said, "Developers with Copilot access saw a significantly higher bug rate while their issue throughput remained consistent."

And earlier this year a "Coding on Copilot" whitepaper from GitClear said, "We find disconcerting trends for maintainability. Code churn — the percentage of lines that are reverted or updated less than two weeks after being authored — is projected to double in 2024 compared to its 2021, pre-AI baseline. We further find that the percentage of 'added code' and 'copy/pasted code' is increasing in proportion to 'updated,' 'deleted,' and 'moved 'code. In this regard, AI-generated code resembles an itinerant contributor, prone to violate the DRY-ness [don't repeat yourself] of the repos visited."

This discussion has been archived. No new comments can be posted.

Does GitHub Copilot Improve Code Quality?

Comments Filter:
  • No. (Score:5, Funny)

    by Local ID10T ( 790134 ) <ID10T.L.USER@gmail.com> on Saturday November 23, 2024 @06:41PM (#64967449) Homepage

    /betteridge

  • by Kelxin ( 3417093 ) on Saturday November 23, 2024 @06:44PM (#64967457)
    100% more likely to be stolen and used by other people for monetary gain.
  • by Misagon ( 1135 ) on Saturday November 23, 2024 @06:59PM (#64967473)

    Every account of using it I've read online has been negative about code quality.

    • The primary positive accounts I've heard about Copilot is that it works as a substitute for Stackoverflow. It helps you figure out what to do when documentation is crappy or nonexistent (but you still need to test it because Stackoverflow can really lead you astray).
      • by gweihir ( 88907 ) on Sunday November 24, 2024 @12:43AM (#64967793)

        That would be rally bad. Because Stackoverflow usually includes a discussion of alternatives and advantages and drawbacks. This a) serves so that a competent (!) coder can understand the problem better and make an adequate selection of a solution and b) does contribute to developer education and experience. Yes, it takes more time, but that time is well-spent.

        • Yeah, if Stackoverflow goes away (or diminishes) then the LLMs are not going to keep up with the latest technology.
    • It works when the LSP is too slow or does not load properly, but there is no way I will trust copilot to fill out a code block for me.

      The irony, I suppose that is the correct word - might be coincidence, its suggestions that it uses are my own code so I guess I should feel flattered in those instances.

  • by stuff-n-things ( 89988 ) on Saturday November 23, 2024 @07:02PM (#64967477) Homepage

    No conflict of interest at Github/Micro$oft either. ;-)

    • I'm sure Microsoft is being 1000% ethical and if there was any evidence that AI actually makes code worse they would definitely let Github publish stuff about that despite Microsoft investing $100 billion or more in AI and AI related stuff.

      • by gweihir ( 88907 )

        Indeed. Obviously Microsoft would fall on their sword to protect us all and make the world a better place! Right? Right?

        Man, I really hope I am retired when all this AI crap has to be ripped out everywhere...

    • Nope, never. Their marketing has never lied before, so why would it lie now. After all, everyone knows if you train AI on average crap on the internet that you get diamonds as a result.

      • by gweihir ( 88907 )

        You can turn crap into diamonds. Just takes a lot of heat and pressure. Marketing can also do it, using a similar approach.

  • by rapjr ( 732628 ) on Saturday November 23, 2024 @07:15PM (#64967487)
    so the common mistakes of people will become pasted into your code by AI. Since training an AI is a statistical process it seems likely that the most average code examples will be what it generates. This might be ok for some things, but it may lower overall code quality. Imagine working with someone who is ok but average in your project and they keep inserting average code into your codebase, perhaps reverting it to less desirable states. Security in particular is not well represented in public code bases. This stuff should be considered experimental until it is better vetted. Yeah it may make your life easier, but what happens later? Any analysis like this needs to follow the side effects from generated code for years.
    • & some languages have more gotchas..& some languages are more common for beginners, who make more mistakes.

      • One thing which gives me the greatest cause for concern is that internet tech is changing continuously yet the AIs seem to consider the version of each piece of software to be immaterial or it is version aware yet its knowledge cut-off isn't aware that for the version being used, the recommendations are no-longer appropriate.

    • by gweihir ( 88907 )

      Exactly. This will also likely lead to common security mistakes becoming more prevalent, decreasing attacker effort. And as a bonus on top, it will be really hard to prevent an LLM from continuing to recommend some crap code once it is known it is crap.

  • Yes and no (Score:5, Insightful)

    by Drethon ( 1445051 ) on Saturday November 23, 2024 @07:25PM (#64967503)

    Based on absolutely no studies or anything but my own opinion... I suspect AI will make good coders better and bad coders worse. Good coders will consider the suggestions, take the good ones and reject the bad ones. Bad coders will take everything.

    • I would agree with that. It's anecdotal but I've noticed when using Copilot at my job that it usually gets me a "mostly" proper solution. But even getting you mostly to a solution can save you an hour or more of digging through documentation. "Hey Copilot, I have an Excel workbook in a memory stream. Load it up with the Open XML library, open up the Summary spreadsheet, and copy out the contents of cell D:3." AI bots are pretty good at crawling through lots of information and summarizing it; I've been

    • by gweihir ( 88907 )

      Probably, although I am doubtful on the impact on good coders. Since most coders are crap (just look at the flood of security vulnerabilities we see every day), that part of the impact will dominate anyways.

    • I've seen this in other fields. Lazy and unskilled workers use AI badly and make their work worse, faster. Skilled, careful workers will use AI as a starting point and research/confirm/expand on it. We saw this with the legal system last year when a bunch of lazy lawyers submitted documents to the court with fake precedents. I've seen it with research articles. AI doesn't so much improve or worsen a person's work as magnifies what is already there.
  • by davecb ( 6526 ) <davecb@spamcop.net> on Saturday November 23, 2024 @07:46PM (#64967517) Homepage Journal

    The best idea is Advait Sarkar's. He noticed how bad LLMs are at anything creative, and instead suggested we use them for things they're good at, predicting what humans would say. Especially if they were asked what a critic would say. See https://leaflessca.wordpress.c... [wordpress.com]

    Trying Pull Requests with CodeRabbit. One of the things I think LLMs can do well is compare my text with a whole body of other people's work. In that vein, CodeRabbit now offers to review git pull requests. https://leaflessca.wordpress.c... [wordpress.com]

    In the search for true artificial intelligence, large language models are a horrible failure which look like a success. https://leaflessca.wordpress.c... [wordpress.com]

    • Yeah, but they are pretty cool. And someone else paid for it.
    • by gweihir ( 88907 )

      In the search for true artificial intelligence, large language models are a horrible failure which look like a success. https://leaflessca.wordpress.c... [wordpress.com]

      You need to have some actual insight to see that though. One thing we are finding out with the current AI craze is how many people actually lack natural insight and typically do not use whatever general intelligence they may actually have available. If you yourself are dumb that way, AI may look like something that can perform on your level or better. That this level can be and often is really bad gets overlooked.

  • Confidence (Score:5, Insightful)

    by phantomfive ( 622387 ) on Saturday November 23, 2024 @08:00PM (#64967529) Journal

    "85% feeling more confident in their code."

    Imagine having such low confidence in the quality of your own code that you feel an LLM is doing you better.

    • Oooo. Burn.

    • by gweihir ( 88907 )

      Indeed. Imagine being this bad at your job. And then ask why that is and does not seem to change. Obviously, incompetent coders (the vast majority) always look for some magic language or tool or approach that makes their code not suck. Obviously that does not work and cannot work because the tooling and the processes are not the problem.

      • by dvice ( 6309704 )

        Compilers are also tools that improves code quality.

        Personally I don't use AI yet for coding. I have tried it, but it is like asking a junior developer for advice. It can do some small things, but for the most part, it is just faster to do it myself than explain how to do it.

        The area where I would like to see AI usage to increase is testing. I think testing is much more better suited for AI than programming, because testing does not require anything except trying all sorts of things, and it doesn't really m

        • by gweihir ( 88907 )

          What about when in testing the "AI" claims a test was successful when it was not, either by test design or by misinterpretation of results? I think mistakes matter a lot in testing.

  • The success you'll have with copilot will depend on the language used. Python is the language that CoPilot generates more useful code between java, Angular et C#.

    • Interesting. Some languages have more "gotcha"s, so there are surely more examples our there of code that falls into those traps, and so the AI surely uses those in its answers too. Languages with fewer "gotchas" result in better AI code...

      No?

    • The success you'll have with copilot will depend on the language used. Python is the language that CoPilot generates more useful code between java, Angular et C#.

      Hmm, I tried it 2 weeks ago with a question of "for /aaa/bbb/.../xxx[12334]yyy write me a RegEx in Java that replaces 1234 with abcd" (roughly...can't share the details).

      1. The RegEx was declared on a Java String with newlines, so it didn't even compile.
      2. The RegEx was wrong. It didn't work if I had fixed it for them...it just completely fucked up the RegEx.
      3. The RegEx they tried to do was about 10x more complicated than it needed to be
      4. The Java API they used was really outdated
      5. Their general

      • by Ksevio ( 865461 )

        Yeah they don't work for everything. Regex's seem to be a pretty big weak point in particular. After using them for a while you get a feel for their strengths and weaknesses.

      • by gweihir ( 88907 )

        Yep, pretty much. The thing is generating a RegEx requires insight. Obviously an LLM can only give you a RegEx has seen before or incompetently try to combine some. That will not work. And your example was _really_ simple.

        • by dvice ( 6309704 )

          On the other hand, I think it would be possible to make an AI that is specifically educated to create regex, because it is quite easy to verify and score the results. And I think this would also be pretty good idea as at least I personally quite often need to find and replace something in hundreds of files. If I could get a regex for that in few seconds, by just asking a question, it would be nice.

  • This whole thing seems super-self-serving and suspect so in that setting:

    Developers with GitHub Copilot access had a 56% greater likelihood of passing all 10 unit tests in the study

    I'm no expert but isn't the idea that you tweak your code until 100% pass is reached and don't stop until then - regardless whether microsoft are watching everything you type? ProTip: Yes.

    Our findings overall show that code authored with GitHub Copilot has increased functionality and improved readability, is of better quality, an

  • by NobleNobbler ( 9626406 ) on Saturday November 23, 2024 @09:28PM (#64967631)

    It's literally impossible to tell the level developers are at with AI. Juniors are abusing it so hard (and hiding it) that they do the most inscrutable, wrong, zero-context solutions, and when I called them on it-- my manager received a report I was being mean. I feel like it's time to get my hose and spray the kids to keep them off my lawn with how this is coming off, but what in the hell is going on. Total circus.

    • and when I called them on it-- my manager received a report I was being mean.

      Use more emojis and memes. You can't say someone is mean when they send you a positive cat GIF.

      Like, Stop using AI, you fucker! [giphy.com]

    • Are you saying junior engineers have gotten worse? Or that junior engineers have always been bad, but now they're just using more advanced tools?
    • Seniors lean on stackoverflow

    • by gweihir ( 88907 )

      And what is worse is that these juniors will never grow into seniors (except by aging), because harder stuff "AI" cannot help them with and they never really learn the simple stuff now.

    • by sfcat ( 872532 )
      That should make this year's performance reviews fun. Make sure to work in the words, "unreliable", "overly-sensitive", "disruptive", and "counter-productive". For bonus points, use "liability".
  • by Felix Baum ( 6314928 ) on Saturday November 23, 2024 @10:16PM (#64967685)
    "GitHub Copilot, an AI-powered coding assistant, has shown potential in improving code quality based on recent studies and user feedback. Research highlights that it enhances several aspects of code quality, including readability, reusability, maintainability, conciseness, and resilience. Developers reported feeling more confident in their work when using Copilot, thanks to its ability to provide clean, error-resilient, and easily understandable code. These qualities make debugging and collaboration more efficient. Specifically, tools like GitHub Copilot Chat, which integrates directly into IDEs, help streamline coding processes. Developers who used Copilot Chat completed code reviews 15% faster and found their reviews to be more actionable compared to traditional methods. Additionally, the assistant helps users maintain focus and reduces frustration, particularly for repetitive or research-intensive tasks However, critiques exist. Some studies suggest that while Copilot speeds up development, it doesn't always guarantee flawless output, and users must still review and adapt its suggestions carefully. In summary, GitHub Copilot appears to boost both productivity and code quality, though it works best when paired with human oversight to address its occasional shortcomings.
  • Including repos that are created to demonstrate vulnerable code.

  • It's obviously shit, but the point is to accelerate code completion faster than what came before. By that measure, the answer is a resounding "yes". Next story.
  • The problem is not to write code faster, the problem is to write better code, and even if Copilot would consistently write excellent code, the fact alone that it becomes easier and faster to write code, will cause there to be more code. More code however means more complexity, more errors and higher costs maintaining that code.

  • I work as a developer and am the only person on my team to shun Copilot. In terms of experience I have more years of experience than half the team and less than the other half. Since my coworkers have adopted Copilot Github says that the number of comments I have left on their PRs has roughly doubled. Even the more experienced developers no write code that looks like it was copy pasted from stackoverflow without regard to correctness. This increased burden of review and the increased number of bug fixes I h

Promising costs nothing, it's the delivering that kills you.

Working...