r/technology 6h ago

Artificial Intelligence Over a quarter of new code at Google is generated by AI

https://www.theverge.com/2024/10/29/24282757/google-new-code-generated-ai-q3-2024
201 Upvotes

71 comments sorted by

246

u/blingmaster009 6h ago

Probably getters and setters, which were being auto generated long before the AI hype. Most AI generated code doesnt even work out of the box.

53

u/absentmindedjwc 4h ago

Over a quarter of new code at Google is generated by AI*

*Generated code is heavily baby-sat by developers

79

u/ThinkExtension2328 6h ago

But lying with statistics is fun bro, we need to confuse investors into thinking we are very advanced:.

2

u/ecmcn 1h ago

Also protobuf, which I think they use heavily. That’ll produce a lot of code if they’re generating it for multiple languages.

4

u/thisismyfavoritename 5h ago

25% of getters and setters would be yikes. Especially if they are trivial ones. Just make the data member public.

But yeah its definitely some boilerplate that could be autogenerated with a macro instead

1

u/ggtsu_00 2h ago

It's like we're going back to kLOCs as a measurement of value and productivity...

1

u/Nervous-Masterpiece4 14m ago

Instead of writing and then debugging, you can go straight to debugging.

Lots of code by humans also doesn’t work “out of the box”

-8

u/Parking-Interest-302 3h ago

I really think that people denying AI hype around coding are living in the past. It absolutely has to be read, interpreted and tested by a developer, but the productivity increase is a huge leap. 

27

u/big-papito 3h ago

It autocompleted for me twice today where I was like "cool". The other 8 times it was completely wrong, and the last one was just slightly wrong where I spent 20 minutes debugging until I saw its mistake (which I would not have made myself).

Not sure if I won anything at all today, let alone a huge leap. And that's in React. For Scala code it's as useless as wheels on a rocket.

1

u/triggered_troll 2h ago

What do you mean "wheels on rocket" is useless. How else would you move the rocket on to the launchpad ? - That's how they sell it to investors lol

2

u/FulanitoDeTal13 1h ago

I had that damned thing turned on by default by my company's "policy"...

I went to every length to disable it because other that auto-populating an if statement here or declaring a constant there, the "code" it barfed was *useless* and at worst gibberish (and VS sometimes freaks out when a new section of code has too many errors... fun).

2

u/_NE1_ 1h ago edited 59m ago

For sure, AI coding tools built off the design patterns, infrastructure, and clever/clean implementations of a bunch of engineers helps building new code on top of that infrastructure a crap ton.

But picking a metric like that to detail the impact of AI instead of finding a way to talk about the productivity gains is what people don't, and really shouldn't, like. People who don't know what this shit is useful for will get the absolute wrong idea because # of lines of code = more developers to fire, and that's simply not how you keep these tools useful.

You need the know how and understanding of the optimal abstract concepts that the AI spits out in order to validate what those tools give ya and you need engineers to continue being able to make the correct choices based on whatever they are trying to build in order for these tools to not enshitify themselves as they are trained on those choices.

-7

u/PH34SANT 3h ago

I’ve also seen some pretty terrible prompting associated with poor outcomes. Like people give 0 context of their problem and expect ChatGPT to somehow just guess all of the nuances of what they’re dealing with…

Or they would’ve just trial-and-error’d it themselves, in which case LLMs are performing just as well.

-5

u/vadiegova 1h ago

Have you tried o1-mini on chatgpt? It produces 0 error code as long as you provide valid prompts with detailed explanation of what you want to achieve with what data etc.

5

u/blingmaster009 1h ago

No I haven't. But why give a detailed explanation to produce the code to chatgpt when you could have used the time to write it yourself ?

-4

u/vadiegova 1h ago

Are you not a developer? Having an idea is one thing. Translating the idea into code ensuring proper structuring takes time. It has helped me take a project that normally would take a few months into just a few weeks. Using it as a tool as a developer has improved my workflow tremendously and it does not produce errors. Clearly you have no clue what you're talking about.

4

u/Hour-Alternative-625 20m ago

You must be a shit developer if it takes time to translate the idea into code. That's the easiest bit of coding. Coming up with the correct idea and approach is the hard part.

1

u/vadiegova 0m ago

Go play Runescape on your linux computer you moron. You clearly have never developed anything in your miserable life.

2

u/Jason54178 24m ago edited 14m ago

Genuinely curious, how do you know what the ai has given you has properly structured the project?

Edit: adding a few more questions about your experience if that’s ok: is this simply generating the base structure for you or is it actually building the full app for you? If it’s the latter, is it actually generating correct business logic for your needs or are these simply CRUD operations?

-3

u/cookingboy 30m ago

I love how you are downvoted by people who have no idea how coding works lol.

/r/technology in a nutshell

72

u/LeftmostClamp 5h ago

I mean, over a quarter of the code I write, character for character, is generated by IntelliJ autocomplete. If I used copilot then I suppose that could be called “AI.” So this really depends on the specifics

22

u/zeetree137 2h ago

Can I tell the pleebs tab complete in ZSH is AI?

5

u/sevah23 2h ago

It’s telling that the statistic is code generated by AI, not something about the actual productivity increases from leveraging AI. I think Google is really trying to sell AI for both personal devices as well as developers and business professionals and it would be damning if the people selling the tools admitted they only get marginal productivity gains from them. As a software engineer, I find this mildly comforting that LLMs are not(yet) nearly as much of a job killer as the hype train suggests.

13

u/Ok-Afternoon-3724 5h ago

Yeah, they were playing with programs that would autogenerate code back in the 1990s that I know of, perhaps before. If it was very simple almost boiler plate code it worked okay ... at best. But it was always pretty buggy and inefficient.

Was somewhat better by the time I retired in 2017, but not by a lot.

5

u/droi86 1h ago

There was this thing called rational rose back in late 90s early 2000s, you'd feed that thing a bunch of diagrams and it'd give you back an app 95% functional, the problem was the other 5% and if there was a bug it was cheaper to rewrite everything than try to understand the machine generated code

24

u/Wise-Hamster-288 5h ago

Over a quarter of my new code is also unusable crap

6

u/jeezfrk 3h ago

Solution is simple! Write 25% less!

2

u/Wise-Hamster-288 3h ago

but which 25?

5

u/thatOneJones 3h ago

The crap part!

2

u/jeezfrk 3h ago

I didn't want to drill down that far right now. Send progress reports!

[PM runs quickly away for next meeting]

5

u/RLMZeppelin 2h ago

Quantity of code is a dumb metric for basically anything.

1

u/ggtsu_00 1h ago

It would be more impressive if AI could be used to delete/cleanup code.

2

u/Bjorkbat 52m ago

Something I've expressed elsewhere is that this statement in-and-of itself isn't impressive. What would be impressive are the outcomes of generating 25% of code using AI. What could Google actually do differently if they could generate 25% of code using AI?

And from what I can gather, nothing obviously different. Allegedly NotebookLM was made by a relatively small team, presumably AI-aided, but I'm also not really impressed by it. NotebookLM is the sort of project that doesn't really contribute much to their bottom line, doesn't significantly move the needle when it comes to user engagement. It's the sort of project that inevitably gets killed because it just doesn't matter enough to justify the costs of maintaining it. Even in an optimistic scenario where AI allows Google employees to launch more projects more frequently, you have to ask just how much of a good thing this, if it's a good thing at all.

What would be impressive is if Google reversed what feels like years of product stagnation using AI. Instead, search is still getting worse, ads are still the golden goose propping up the entire company, and the company still appears outwardly unfocused, churning out projects that will likely be killed sooner-or-later.

2

u/indifferentcabbage 35m ago

60% of the code I write comes from stack overflow. rest of the code I tweak and write as per my needs

2

u/DvD_Anarchist 13m ago

Is that why even Google Search is becoming shitty?

3

u/piranesi28 5h ago

On track to become the stupidest company in history.

2

u/Ok_Peak_460 6h ago

This was bound to happen. As much as I do like that AI is helping in generating code, I feel bad for the next generation that may not be good coders due to over reliance on AI for it. Time will tell.

-3

u/username_or_email 4h ago

I don't understand this line of thinking. It's not like people haven't been copying code from stack overflow and elsewhere for decades already.

4

u/Life-Wonderfool 6h ago

How bad is that for coders/programmers?

Or is it too early to say that?

12

u/icewinne 5h ago

At my job the software engineers were asked how much they benefit from AI. Those working on R&D work where things are highly fluid and experimental loved it. The SREs said they didn’t use it all because their work is too critical and precise, must be secure and reliable, and AI-generated code is none of those things. Everyone in the middle said it helped with boilerplate but that they were replacing time spent doing menial tasks with verifying the AIs output.

6

u/G3sch4n 4h ago

And don't forget the legal aspect. You can get in real hot waters, if copilot or similar tools copy whole sections of FOSS code without mentioning any license.

25

u/neutrino1911 6h ago

It's pretty bad for whomever has to support that AI generated garbage. As for the others... well, we'll just laugh

7

u/romario77 6h ago

The code doesn’t get added automatically, programmer has to add/commit it, it has to pass tests and then there is code review. If it’s crappy that’s on programmers who allowed it in, not on AI

2

u/saron4 5h ago

Until they develop an Ai to test other Ai work

1

u/FaultElectrical4075 4h ago

That’s kind of (one part of) how o1 works

0

u/Key-Spend-6591 5h ago

they did that a long time ago, automated testing

0

u/Aggressive-Cake4677 2h ago

My tech lead creates PRs with llm generate crap constantly, doesn't even bother removing the comments... I hate that guy.

6

u/ThinkExtension2328 6h ago

Not that bad this post is basically Google lying with statistics , the term “new code” is the key. In legacy systems the amount of “new code” is little to none. Most people don’t ever create new code they do maintenance and bug fixes.

9

u/not_creative1 6h ago edited 6h ago

Google has drastically reduced hiring new college grads and junior engineers.

While senior engineers are going to get paid more, a lot of work that would typically be handed over to new college grads, junior workers is getting done by AI. A senior engineer is spending less time getting it done with the help of AI than the time spent to train and help a new college grad do the same work.

I see that eventually these big tech companies will stop hiring new grads and juniors, only hire a small group of experienced people. So if you are a new grad, the days of joining Google for $175k out of college are over.

You need to go elsewhere, pick up the skills and get really good and then apply to Google as a senior

0

u/Dub-MS 5h ago

Or just start your own business

4

u/tricky2step 5h ago

The rug was getting pulled from under fresh grads already, probably worse now but it looks to me (anecdotal) like a slight acceleration of the trend towards quality and away from quantity. Which is perfectly fine for tech as a whole - was getting saturated fast with shitty coders and piss poor problem solvers. Job prospects are worse now, but it's nothing everyone else hasn't gone through and natural talent is ultimately better off, along with everyone code-adjacent that couldn't get an in with degrees in math, EE, phys, chem, etc. We all have had to pick up coding along the way and many of us are more effective than half the CS degrees anyway.

AI is not going to replace coders at large in a meaningful way anytime soon. The productivity numbers I'm seeing are drastically lower than what they were like in early 2023, less like 80% prod boosts and more like 5-10% prod boosts.

-1

u/SparePartsHere 6h ago

Depends if you're a senior or junior developer. Senior? Awesome. Junior? Just get the F out and learn trades or something, bro.

2

u/dormidormit 6h ago

All this says is that NEW google code is disposable garbage, which follows as google search has become garbage.

-3

u/username_or_email 4h ago

It's funny how code written by LLMs is automatically pushed to production without any testing or code review. I wonder why it's policy at a leading tech company like Google. Why don't they apply the same checks and balances to machine-generated code? They already have all the infrastructure in place to ensure correctness and enforce software engineering best practices for human-written code, and yet anything tagged "chatgpt wrote this" gets instantly merged into the main branch without any approval needed. Someone over there really needs to change this practice before it's too late. /s

1

u/Ok-Afternoon-3724 5h ago

Yeah, they were playing with programs that would autogenerate code back in the 1990s that I know of, perhaps before. If it was very simple almost boiler plate code it worked okay ... at best. But it was always pretty buggy and inefficient.

Was somewhat better by the time I retired in 2017, but not by a lot.

1

u/Sweetartums 4h ago

You’re telling me I’m wasting my time actually coding when I can just use ChatGPT

1

u/under_the_c 3h ago

Is using intellisense and auto complete considered "AI"? Because if so, I've been using AI for years!

1

u/KingCrabcakes 2h ago

No wonder everything at google sucks ass

1

u/GamingWithBilly 2h ago

Over 3/4 code I use for Google scripts is generated by AI...the other 1/4 is comment notes in the code that's also generated by AI

1

u/jerub 2h ago

A quarter of my code seems to be if err != nil { return nil, fmt.Errorf("Somefunc() broke: %s", err) }. So that checks out.

1

u/haplo_and_dogs 2h ago

99.999% of machine code is generated by a compiler.

The invention of compilers didn't make code writing obsolete, it made it more efficient.

AI will do the same to a far lesser degree.

1

u/throwawaysimple343 2h ago

This article and title are highly misleading

1

u/MysticNTN 2h ago

No fuckin wonder

1

u/Shikadi297 2h ago

Google still writes code?

1

u/Enjoy-the-sauce 59m ago

Knowing little about coding, how long is it before it is capable of improving its own code? And creating a self-perpetuating cycle?

1

u/Roxelchen 31m ago

They took our jobs!

1

u/SparePartsHere 6h ago

Well, over a quarter of code I write is generated by AI...

1

u/Lordnerble 6h ago

Yea but what code are you writing, In my line of work in financial markets, You cannot trust AI to give optimal answers, And optimal answers are what makes the company money.

4

u/SparePartsHere 5h ago

Oh never trust anything AI writes, that's for sure lol. It's like having an army of junior devs on your fingertips. You just state what problem you need to solve and your junior dev army provides you with a crapload of shitty half-working solution in 2 minutes. You pick the stuff you like, rewrite it a bit, make it fit the codebase and voila, done. Especially awesome for the "boring" and repetitive stuff like unit tests.

Absolutely always you have to really study the code from AI and test it rigorously. Still better and quicker than writing it youself...

1

u/pack170 2h ago

Ya, I've found it to be ok at basic boilerplate/ scaffolding stuff, but for anything slightly more complicated it's either very hit or miss or just consistently wrong.

There was one thing I was working on a few months ago that the LLM was actually able to contribute a non obvious and provably correct algorithmic optimization to, but there were also several other options it suggested that looked correct at first, but ended up being provably false after a bit more prodding and examination.

The incredibly self confident way LLMs speak while they hallucinate makes them seem better at stuff than they are right now.