Skip to content

Study finds AI tools made open source software developers 19 percent slower

Technology
37 31 0
  • Their sample size was 16 people...

    I'm not really sure why it was such a small sample size. It definitely casts doubt on some of their conclusions. I also have issues with some methodology used. I think a better study that came out a week or two ago was the one that showed visible neurological decline from AI use.

  • Great as an assistant for boring tasks. Still needs checking.

    Can also help suggest improvements, but still needs checking.

    Have to learn when to stop interacting with it and do it yourself.

    A "junior" project manager at my company vibe coded an entire full stack web app with one of those LLM IDEs. His background is industrial engineering and claims to have basically no programming experience.

    It "works", as in, it does what it's meant to, but as you can guess, it relies on calls to LLM APIs where it really doesn't have to, and has several critical security flaws, inconsistencies in project structure and convention, and uses deprecated library features.

    He already pitched it to one of our largest clients, and they're on board. They want to start testing at the end of the month.

    He's had one junior dev who's been managing to keep things somewhat stable, but the poor dude really had his work cut out for him. I only recently joined the project because "it sounded cool", so I've been trying to fix some flaws while adding new requested features.

    I've never worked with the frameworks and libraries before, so it's a good opportunity to upskill, but god damn I don't know if I want my name on this project.

    A similar thing is happening with my brother at a different company. An executive vibe coded a web application, but this thing absolutely did not work.

    My brother basically had one night to get it into a working state. He somehow (ritalin) managed to do it. The next day they presented it to one of their major clients. They really want it.

    These AI dev tools absolutely have a direct negative impact on developer productivity, but they also have an indirect impact where non-devs use them and pass their Eldritch abominations to the actual devs to fix, extend and maintain.

    Two years ago, I was worried about AI taking dev jobs, but now it feels like, to me, we'll need more human devs than ever in the long run.

    Like, weren't these things supposed to exponentially get better? Like, cool, gh copilot can fuck up my project files now.

  • Coders spent more time prompting and reviewing AI generations than they saved on coding. On the surface, METR's results seem to contradict other benchmarks and experiments that demonstrate increases in coding efficiency when AI tools are used. But those often also measure productivity in terms of total lines of code or the number of discrete tasks/code commits/pull requests completed, all of which can be poor proxies for actual coding efficiency. These factors lead the researchers to conclude that current AI coding tools may be particularly ill-suited to "settings with very high quality standards, or with many implicit requirements (e.g., relating to documentation, testing coverage, or linting/formatting) that take humans substantial time to learn." While those factors may not apply in "many realistic, economically relevant settings" involving simpler code bases, they could limit the impact of AI tools in this study and similar real-world situations.

    True and not true at the same time. Using agents indeed often don't work, mostly when I'm trying to do the wrong thing. Because then, AI agent does not say "the way you do it is overly complicated, it does not make any sense", but instead it says: "excellent idea, here are X steps I need to do to make it happen". It wasted my time many times, but it also guided me quickly though some problems that would take hours to research. Some of my projects wouldn't have been finished without AI.

  • Their sample size was 16 people...

    Who are in the process of learning to do something new, versus the workflow that they've been trained in and have a lot of experience in.

    Where was the sample of non-coders tasked with doing the same thing, using AI to help or learning without assistance?

    Where was the sample of coders prohibited from looking anything up and having to rely solely on their prior knowledge to do the job?

    It might help refine what's actually being tested.

  • Sounds reasonable. The time and energy ive lost on trying very confident chat gpt suggestions that doesnt work must be weeks at this point.

    Sometimes its very good though and really helps, which is why its so frustrating. You never know if its going to work before you go through the process.

    It has changed how me and coworkers work now also. We just talk to chat gpt instead of even trying to look something up in the docs and trying to understand it. Too slow to do that now, it feels like. There is a pressure to solve anything quickly now that chat gpt exists.

    You have to ignore the obsequious optimism bias LLM's often have. It all comes down to their training set and if they have seen more than you have.

    I don't generally use them on projects I'm already familiar with unless it's for fairly boring repetitive work that would be fiddly with search and replace, e.g. extract the common code out of these functions and refactor.

    When working with unfamiliar code they can have an edge so if I needed a simple mobile app I'd probably give the LLM a go and then tidy up the code once it's working.

    At most I'll give it 2 or 3 attempts to correct the original approach before I walk away and try something else. If it starts making up functions it APIs that don't exist that is usually a sign out didn't know so time to cut your losses and move on.

    Their real strengths come in when it comes to digesting large amounts of text and sumerising. Great for saving you reading all the documentation on a project just to try a small thing. But if your going to work on the project going forward your going to want to invest that training data yourself.

  • Coders spent more time prompting and reviewing AI generations than they saved on coding. On the surface, METR's results seem to contradict other benchmarks and experiments that demonstrate increases in coding efficiency when AI tools are used. But those often also measure productivity in terms of total lines of code or the number of discrete tasks/code commits/pull requests completed, all of which can be poor proxies for actual coding efficiency. These factors lead the researchers to conclude that current AI coding tools may be particularly ill-suited to "settings with very high quality standards, or with many implicit requirements (e.g., relating to documentation, testing coverage, or linting/formatting) that take humans substantial time to learn." While those factors may not apply in "many realistic, economically relevant settings" involving simpler code bases, they could limit the impact of AI tools in this study and similar real-world situations.

    The main issue i have with AI coding, hasn't been the code. Its a bit ham fisted and overly naive, it is as if it's speed blind.

    The main issue is that some of the code is out of date using functions that are deprecated etc, and it seems to be mixing paradigms and styles across languages in a very frustrating? way.

  • True and not true at the same time. Using agents indeed often don't work, mostly when I'm trying to do the wrong thing. Because then, AI agent does not say "the way you do it is overly complicated, it does not make any sense", but instead it says: "excellent idea, here are X steps I need to do to make it happen". It wasted my time many times, but it also guided me quickly though some problems that would take hours to research. Some of my projects wouldn't have been finished without AI.

    Some of my projects wouldn’t have been finished without AI.

    This says way more about you than it says about AI tools

  • True and not true at the same time. Using agents indeed often don't work, mostly when I'm trying to do the wrong thing. Because then, AI agent does not say "the way you do it is overly complicated, it does not make any sense", but instead it says: "excellent idea, here are X steps I need to do to make it happen". It wasted my time many times, but it also guided me quickly though some problems that would take hours to research. Some of my projects wouldn't have been finished without AI.

    Just make sure you're validating everything you produce with it.

  • Coders spent more time prompting and reviewing AI generations than they saved on coding. On the surface, METR's results seem to contradict other benchmarks and experiments that demonstrate increases in coding efficiency when AI tools are used. But those often also measure productivity in terms of total lines of code or the number of discrete tasks/code commits/pull requests completed, all of which can be poor proxies for actual coding efficiency. These factors lead the researchers to conclude that current AI coding tools may be particularly ill-suited to "settings with very high quality standards, or with many implicit requirements (e.g., relating to documentation, testing coverage, or linting/formatting) that take humans substantial time to learn." While those factors may not apply in "many realistic, economically relevant settings" involving simpler code bases, they could limit the impact of AI tools in this study and similar real-world situations.

    Studies show that the electric drills drill faster than a manual, hand-cranked drill.

  • Coders spent more time prompting and reviewing AI generations than they saved on coding. On the surface, METR's results seem to contradict other benchmarks and experiments that demonstrate increases in coding efficiency when AI tools are used. But those often also measure productivity in terms of total lines of code or the number of discrete tasks/code commits/pull requests completed, all of which can be poor proxies for actual coding efficiency. These factors lead the researchers to conclude that current AI coding tools may be particularly ill-suited to "settings with very high quality standards, or with many implicit requirements (e.g., relating to documentation, testing coverage, or linting/formatting) that take humans substantial time to learn." While those factors may not apply in "many realistic, economically relevant settings" involving simpler code bases, they could limit the impact of AI tools in this study and similar real-world situations.

    On a different note: is it just me or do images with this color scheme (that blue and black) also have a weird 3d look to them to you?

  • Their sample size was 16 people...

    Where the most experienced minority only had a few weeks of using AI inside an IDE like Cursor.

  • The main issue i have with AI coding, hasn't been the code. Its a bit ham fisted and overly naive, it is as if it's speed blind.

    The main issue is that some of the code is out of date using functions that are deprecated etc, and it seems to be mixing paradigms and styles across languages in a very frustrating? way.

    Yep I've got a working iOS app, a v.2 branched and on the way, with a ton of MapKit integrations. Unfortunately I'm getting depreciation errors and having to constantly remind the AI that it's using old code, showing it examples of new code, and then watching it forget as we keep talking.

    Still, I have a working iOS app, which only took a few hours. When Jack Dorsey said he'd vibe coded his new app in a long weekend, I'm like, hey me too.

  • Coders spent more time prompting and reviewing AI generations than they saved on coding. On the surface, METR's results seem to contradict other benchmarks and experiments that demonstrate increases in coding efficiency when AI tools are used. But those often also measure productivity in terms of total lines of code or the number of discrete tasks/code commits/pull requests completed, all of which can be poor proxies for actual coding efficiency. These factors lead the researchers to conclude that current AI coding tools may be particularly ill-suited to "settings with very high quality standards, or with many implicit requirements (e.g., relating to documentation, testing coverage, or linting/formatting) that take humans substantial time to learn." While those factors may not apply in "many realistic, economically relevant settings" involving simpler code bases, they could limit the impact of AI tools in this study and similar real-world situations.

    I don't doubt this is true. I've been playing with an A.I and some fairly simple python scripts and it's so tedious to get the A.I. to actually do something to the script correctly. Learning to prompt is a skill all it's own.

    In my experience it's much more useful for doing things like in AWS like create a Cloudformation template or look through user permissions for excess privileges or setup a backup schedule, like at scale when you have lots of accounts and users, etc.

  • A "junior" project manager at my company vibe coded an entire full stack web app with one of those LLM IDEs. His background is industrial engineering and claims to have basically no programming experience.

    It "works", as in, it does what it's meant to, but as you can guess, it relies on calls to LLM APIs where it really doesn't have to, and has several critical security flaws, inconsistencies in project structure and convention, and uses deprecated library features.

    He already pitched it to one of our largest clients, and they're on board. They want to start testing at the end of the month.

    He's had one junior dev who's been managing to keep things somewhat stable, but the poor dude really had his work cut out for him. I only recently joined the project because "it sounded cool", so I've been trying to fix some flaws while adding new requested features.

    I've never worked with the frameworks and libraries before, so it's a good opportunity to upskill, but god damn I don't know if I want my name on this project.

    A similar thing is happening with my brother at a different company. An executive vibe coded a web application, but this thing absolutely did not work.

    My brother basically had one night to get it into a working state. He somehow (ritalin) managed to do it. The next day they presented it to one of their major clients. They really want it.

    These AI dev tools absolutely have a direct negative impact on developer productivity, but they also have an indirect impact where non-devs use them and pass their Eldritch abominations to the actual devs to fix, extend and maintain.

    Two years ago, I was worried about AI taking dev jobs, but now it feels like, to me, we'll need more human devs than ever in the long run.

    Like, weren't these things supposed to exponentially get better? Like, cool, gh copilot can fuck up my project files now.

    These AI dev tools absolutely have a direct negative impact on developer productivity, but they also have an indirect impact where non-devs use them and pass their Eldritch abominations to the actual devs to fix, extend and maintain.

    Sounds like the next evolution of the Excel spreadsheet macro. Or maybe it's convergent evolution toward the same niche. (I still have nightmares about Excel spreadsheet macros.)

  • I like to think typos like that confirm my humanity 🙂

  • Yep I've got a working iOS app, a v.2 branched and on the way, with a ton of MapKit integrations. Unfortunately I'm getting depreciation errors and having to constantly remind the AI that it's using old code, showing it examples of new code, and then watching it forget as we keep talking.

    Still, I have a working iOS app, which only took a few hours. When Jack Dorsey said he'd vibe coded his new app in a long weekend, I'm like, hey me too.

    LLMs can't forget things because they are not capable of memory.

  • I don't doubt this is true. I've been playing with an A.I and some fairly simple python scripts and it's so tedious to get the A.I. to actually do something to the script correctly. Learning to prompt is a skill all it's own.

    In my experience it's much more useful for doing things like in AWS like create a Cloudformation template or look through user permissions for excess privileges or setup a backup schedule, like at scale when you have lots of accounts and users, etc.

    So it's like talking to women...

  • I like to think typos like that confirm my humanity 🙂

    shhh don’t let the bots in on our secret

    also now I’m hungry for phở

  • shhh don’t let the bots in on our secret

    also now I’m hungry for phở

    With enough training data from me and chatbots will spell like shit. Bad grammar as well.

  • Coders spent more time prompting and reviewing AI generations than they saved on coding. On the surface, METR's results seem to contradict other benchmarks and experiments that demonstrate increases in coding efficiency when AI tools are used. But those often also measure productivity in terms of total lines of code or the number of discrete tasks/code commits/pull requests completed, all of which can be poor proxies for actual coding efficiency. These factors lead the researchers to conclude that current AI coding tools may be particularly ill-suited to "settings with very high quality standards, or with many implicit requirements (e.g., relating to documentation, testing coverage, or linting/formatting) that take humans substantial time to learn." While those factors may not apply in "many realistic, economically relevant settings" involving simpler code bases, they could limit the impact of AI tools in this study and similar real-world situations.

    Slowing you down is the main benefit!

    It helps you to keep more brain time on solving the actual problem, and less on boring syntax crap. Of course, then it gets the syntax crap wrong and you need to waste a lot of time fixing it.

  • Worth it to note that you don’t need an app to do this.

    Technology
    1
    2 Stimmen
    1 Beiträge
    0 Aufrufe
    Niemand hat geantwortet
  • You know all those Cyberpunk books and movies?

    Technology
    3
    2 Stimmen
    3 Beiträge
    0 Aufrufe
    S
    Tech bros actually think it’s something to aspire to. Saw some tech moron on Xitter say that cyberpunk is a utopia we can achieve. Then he started arguing with people who told him it’s a dystopia. Fascist tech bros think they will be the elites in Harlan’s World and not some downtrodden servant.
  • DIY experimental Redox Flow Battery kit

    Technology technology
    3
    1
    17 Stimmen
    3 Beiträge
    0 Aufrufe
    C
    The roadmap defines 3 milestone batteries. The first is released, it's a benchtop device that you can relatively easily build on your own. It has an electrode side of 2 x 2cm2. It does not store any significant amount of energy. The second one is being developed right now, it has a cell the size of a small 3d printer bed (20x20cm) and will also not store practical amounts of energy. It will hopefully prove though that they are on the right track and that they can scale it up. The third battery only will store significant amounts of energy but in only due end of the year (probably later). Current Vanadium systems cost approx. 300-600$/kWh according to some random website I found. The goal of this project is to spread the knowledge about Redox Flow Batteries and in the medium term only make them commercially viable. The aniolyth and catholyth are based on the Zink-Iodine system in an aqueous solution. There are a bunch of other systems though, each with their trade offs. The anode and cathode are both graphite felt in the case of the dev kit.
  • 138 Stimmen
    23 Beiträge
    0 Aufrufe
    A
    industries tend to be more centralized in China. It's not that that's indicative of every city, more that Shenzhen already has easy access to the kind of manufacturing and products that a robotics company would find ideal.
  • 344 Stimmen
    77 Beiträge
    0 Aufrufe
    mhad1@lemmy.zipM
    Nah its not even always about profit, sometimes its just pure sloppy showoff like a page where I am supposed to sign up should not be promoting the company, if Ive already got onto that page why do I need to scroll all the way down to the join/sign up button!
  • So what's left?

    Technology
    33
    17 Stimmen
    33 Beiträge
    0 Aufrufe
    povoq@slrpnk.netP
    Look, if you run the server you have access to metadata of clients connecting to it. That is networking 101. And that Signal shares phone numbers and connection timestamps is well established by court documents. The security audits are of the code and encryption algorithm, not the infrastructure.
  • Giving Up on Element & Matrix.org

    Technology
    78
    1
    241 Stimmen
    78 Beiträge
    0 Aufrufe
    swelter_spark@reddthat.comS
    The CSAM spam is so annoying. I don't understand who is doing this or why.
  • 24 Stimmen
    6 Beiträge
    48 Aufrufe
    I
    They want to become the new tax collectors. They want to buy our toy democracy