Not a skeptic, I use AI for coding daily and am working on a custom agent setup because, through my experience for more than a year, they are not up to hard tasks.
This is well known I thought, as even the people who build the AIs we use talk about this and acknowledge their limitations.
I'm pretty sure at this point more than half of Anthropic's new production code is LLM-written. That seems incompatible with "these agents are not up to the task of writing production level code at any meaningful scale".
how are you pretty sure? What are you basing that on?
If true, could this explain why Anthropics APIs are less reliable than Gemini's? (I've never gotten a service overloaded response from Google like I did from Anthropic)
My current understanding (based on this text and other sources) is:
- There exist some teams at Anthropic where around 90% of lines of code that get merged are written by AI, but this is a minority of teams.
- The average over all of Anthropic for lines of merged code written by AI is much less than 90%, more like 50%.
> I've never gotten a service overloaded response from Google like I did from Anthropic
They're Google, they out-scale everyone. They run more than 1.3 quadrillion tokens per month through LLMs!
This is well known I thought, as even the people who build the AIs we use talk about this and acknowledge their limitations.