Anthropic tightens usage limits for Claude Code without telling users
88 comments
·July 17, 2025thr0waway001
> One user, who asked not to be identified, said it has been impossible to advance his project since the usage limits came into effect.
Vibe limit reached. Gotta start doing some thinking.
m4rtink
Who would have though including a hard depedency on third part service with unclear long term availability would be a problem!
Paid compilers and remotely acessible mainframes all over again - people apparently never learn.
dude250711
He did not pass the vibe check.
bGl2YW5j
Came to comment on the same quote.
I'm surprised, but know I shouldn't be, that we're at this point already.
mrits
First one was free
mattigames
I would be a little disappointed if that wasn't the case, after all we have been there quite a while in regards to the art models.
mrits
I honestly feel sorry for these vibe coders. I'm loving AI in a similar way that I loved google or IDE magic. This seems like a far worst version of those developers that tried to build an entire app with Eclipse or Visual Studio GUI drag and drop from the late 90s
ants_everywhere
The other day I was doing major refactorings on two projects simultaneously while doing design work for two other projects. It occurred to me to check my API usage for Gemini and I had spent $200 that day already.
Users are no doubt working these things even harder than I am. There's no way they can be profitable at $200 a month with unlimited usage.
I think we're going to evolve into a system that intelligently allocates tasks based on cost. I think that's part of what openrouter is trying to do, but it's going to require a lot of context information to do the routing correctly.
pembrook
The funny thing is Claude 4.0 isn't even that 'smart' from a raw intelligence perspective compared to the other flagship models.
They've just done the work to tailor it specifically for proper tool using during coding. Once other models catch up, they will not be able to be so stingy on limits.
Google has the advantage here given they're running on their own silicon; can optimize for it; and have nearly unlimited cashflows they can burn.
I find it amusing nobody here in the comments can understand the scaling laws of compute. It seems like people have a mental model of Uber burned into their head thinking that at some point the price has to go up. AI is not human labor.
Over time the price of AI will fall, not rise. Losing money in the short term betting this will happen is not a dumb strategy given it's the most likely scenario.
I know everybody really wants this bubble to pop so they can make themselves feel smart for "calling it" and I'm sure it will, but in the long term this is all correct.
alphager
Even if Moore's law was still in effect and the computer resources required stayed the same and compute stayed as efficient per watt (neither is true), it would just halve compute costs every 18 months. You're able to read about people hitting $4000 costs/month on the $200 plan upthread. That's 8 years until it's cost effective.
Are they really ready to burn money for 8 years?
pembrook
Uber operated at a loss for 9 years. They're now a profitable, market-winning business.
Amazon operated at a loss for 9 years, and barely turned a profit for over a decade longer than that. They're now one of the greatest businesses of all time.
Spotify operated at a loss for 17 years until becoming profitable. Tesla operated at a loss for 17 years before turning a profit. Palantir operated at a loss for 20 years before turning a profit.
And this was before the real age of Big Tech. Google can burn more than any of these companies ever raised, combined.
klik99
Those aren’t good comparisons.
Uber operated at a loss to destroy competition and raised prices after they did that.
Amazon (the retailer) did the same and leveraged their position to enter new more lucrative markets.
Dunno about Spotify, but Tesla and palantir both secured lucrative contracts and subsidies.
Anthropic is against companies with deeper pockets and can’t spend to destroy competition, their current business model can only survive if they reduce costs or raise prices. Something’s got to give
sothatsit
I think people also expect models to be optimised over time. For example, the 5x drop in cost of o3 was probably due to some optimisation on OpenAI's end (although I'm sure they had business reasons for dropping the price as well).
Small models have also been improving steadily in ability, so it is feasible that a task that needs Claude Opus today could be done by Sonnet in a year's time. This trend of model "efficiency" will add on top of compute getting cheaper.
Although, that efficiency would probably be quickly eaten up by increased appetites for higher performance, bigger, models.
andix
The thing is, all the models are not that 'smart'. None of them is AGI.
Currently it's much more important to manage context, split tasks, retry when needed, not getting stuck in an infinite loop, expose the right tools (but not too many), ...
macinjosh
Prices for yesterday's frontier models will fall but there will always be the next big model. similar to how game graphics get ever better but ever more demanding at the bleeding edge.
carlhjerpe
Yes but games also look an awful lot better (fidelity wise) than not so many years ago.
Aurornis
I played with Claude Code using the basic $20/month plan for a toy side project.
I couldn't believe how many requests I could get in. I wasn't using this full-time for an entire workweek, but I thought for sure I'd be running into the $20/month limits quickly. Yet I never did.
To be fair, I spent a lot of time cleaning up after the AI and manually coding things it couldn't figure out. It still seemed like an incredible number of tokens were being processed. I don't have concrete numbers, but it felt like I was easily getting $10-20 worth of tokens (compared to raw API prices) out of it every single day.
My guess is that they left the limits extremely generous for a while to promote adoption, and now they're tightening them up because it’s starting to overwhelm their capacity.
I can't imagine how much vibe coding you'd have to be doing to hit the limits on the $200/month plan like this article, though.
eddythompson80
Worth noting that a lot of these limits are changing very rapidly (weekly if not daily) and also depend on time of day, location, account age, etc.
dawnerd
I hit the limits within an hour with just one request in CC. Not even using opus. It’ll chug away but eventually switch to the nearing limit message. It’s really quite ridiculous and not a good way to upsell to the higher plans without definitive usage numbers.
ChadMoran
If you aren't hitting the limits you aren't writing great prompts. I can write a prompt and have it go off and work for about an hour and hit the limit. You can have it launch sub-agents, parallelize work and autonomously operate for long periods of time.
Think beyond just saying "do this one thing".
stogot
How is that a great prompt having it run for an hour without your input? Sounds like it’s just generating wasteful output.
mrits
That clears up a lot for me. I don't think I've ever had it take for than a couple of minutes. If it takes more than a minute I usually freak out and press stop
cladopa
Thinking is extremely inefficient compared with the usual query in Chat.
If you think a lot, you can spend hundreds of dollars easily.
buremba
They're likely burning money so I can't be pissed off yet, but we see the same Cursor as well; the pricing is not transparent.
I'm paying for Max, and when I use the tooling to calculate the spend returned by the API, I can see it's almost $1k! I have no idea how much quota I have left until the next block. The pricing returned by the API doesn't make any sense.
neom
We just came out of closed alpha yesterday and have been trying to figure out how best to price, if you'd be willing to provide any feedback I'd certainly appreciate it: https://www.charlielabs.ai/pricing - Thank you!! :)
roxolotl
A coworker of mine claimed they've been burning $1k a week this month. Pretty wild it’s only costing the company $200 a month.
gerdesj
Crikey. Now I get the business model:
I hire someone for say £5K/mo. They then spend $200/mo or is it a $1000/wk on Claude or whatevs.
Profit!
AtheistOfFail
The model is "outspend others until they're bankrupt".
Known as the Uber model or Amazon vs Diapers.com
null
dfsegoat
Can you clarify which tooling you are using? Is it cursor-stats?
adamtaylor_13
That’s funny I literally started the $200/month plan this week because I routinely spend $300+/month on API tokens.
And I was thinking to myself, “How does this make any sense financially for Anthropic to let me have all of this for $200/month?”
And then I kept getting hit with those overloaded api errors so I canceled my plan and went back to API tokens.
I still have no idea what they’re doing over there but I’ll happily pay for access. Just stop dangling that damn $200/month in my face if you’re not going to honor it with reasonable access.
bad_haircut72
I went from pro to max because I hve been hitting limits, I could tell they were reducing it because I used to go multiple hours on pro but now its like 3. Congrats Anthropic you got $100 more out of me, at the cost of irrecoverable goodwill
hellcow
For what it's worth, when Cursor downgraded their Claude limits in the middle of my annual subscription term, I emailed them to ask for a pro-rated refund, and it was granted. You may be able to do something similar with Claude Code.
Changing the terms of the deal midway through a subscription to make it much less valuable is a really shady business practice, and I'm not sure it's legal.
Ataraxic
I need to see a video of what people are doing to hit the max limits regularly.
I find sonnet really useful for coding but I never even hit basic limits. at $20/mo. Writing specs, coming up with documentation, doing wrote tasks for which many examples exist in the database. Iterate on particular services etc.
Are these max users having it write the whole codebase w/ rewrites? Isn't it often just faster to fix small things I find incorrect than type up why I think it's wrong in English and have it do a whole big round trip?
sothatsit
I can tell you how I hit it: Opus and long workflows.
I have two big workflows: plan and implement. Plan follows a detailed workflow to research an idea and produce a planning document for how to implement it. This routinely takes $10-30 in API credits to run in the background. I will then review this 200-600 line document and fix up any mistakes or remove unnecessary details.
Then implement is usually cheaper, and it will take that big planning document, make all the changes, and then make a PR in GitHub for me to review. This usually costs $5-15 in API credits.
All it takes is for me to do 3-4 of these in one 5-hour block and I will hit the rate-limit of the $100 Max plan. Setting this up made me realise just how much scaffolding you can give to Opus and it handles it like a champ. It is an unbelievably reliable model at following detailed instructions.
It is rare that I would hit the rate-limits if I am just using Claude Code interactively, unless I am using it constantly for hours at a time, which is rare. Seems like vibe coders are the main people who would hit them regularly.
Ensorceled
> Isn't it often just faster to fix small things I find incorrect than type up why I think it's wrong in English and have it do a whole big round trip?
This is my experience: at some point the AI isn't converging to a final solution and it's time to finish the rest by hand.
bluefirebrand
My experience is that if the AI doesn't oneshot it, it's faster to do it myself
If you find yourself going back and forth with the AI, you're probably not saving time over a traditional google search
Edit: and it basically never oneshots anything correctly
pinoy420
[dead]
nh43215rgb
Are you using claude code for coding with sonnet? Just claude web use alone is indeed fairly relaxed i think.
adamtaylor_13
I couldn’t even get it to do simple tasks for me this week on the max plan. It’s not just max users overloading it. It feels like they’re randomly rate limiting users.
One day my very first prompt in the morning was blocked. Super strange.
jmartrican
I have the $100 plan and now quickly get downgraded to Sonnet. But so far have not hit any other limits. I use it more on the weekends over several hours, so lets see what this weekend has in store.
I suspected that something like this might happen, where the demand will outstrip the supply and squeeze small players out. I still think demand is in its infancy and that many of us will be forced to pay a lot more. Unless of course there are breakthroughs. At work I recently switched to non-reasoning models because I find I get more work done and the quality is good enough. The queue to use Sonnet 3.7 and 4.0 is too long. Maybe the tools will improve reduce token count, e.g. a token reducing step (and maybe this already exists).
j45
Off hour usage seems to be different for sure.
Also there's likely only so much fixed compute available, and it might be getting re allcoated for other uses behind the scene from time to time as more compute arrives.
martinald
I'm not sure this is "intentional" per se or just massively overloaded servers because of unexpected demand growth and they are cutting rate limits until they can scale up more. This may become permanent/worse if the demand keeps outstripping their ability to scale.
I'd be extremely surprised if Anthropic picked now of all times to decide on COGS optimisation. They potentially can take a significant slice of the entire DevTools market with the growth they are seeing, seems short sighted to me to nerf that when they have oodles of cash in bank and no doubt people hammering at their door to throw more cash at them.
andix
A lot of people switched away from Cursor within the blink of an eye. Switching IDEs is a big deal for me - it takes a lot of effort, which is why I never switched to Cursor in the first place.
I think Claude Code is a much better concept, the coding agent doesn't need to be connected to the IDE at all. Which also means you can switch even faster to a competitor. In that sense, Claude Code may have been a huge footgun. Gaining market share might turn out to be completely worthless.
mattnewton
I think in the case of Cursor, they are one of may VScode forks, so a switch is not really very challenging. I agree there is little to keep me on any individual app or model (which is one reason I think cursor's reported 9b valuation is a little crazy!)
andix
Only if you're using VS code in the first place. VS code is fine for web dev and js/ts/python. But I really don't like it for Java, C#, C++, SQL, and many more.
bgwalter
“It just stopped the ability to make progress,” the user told TechCrunch. “I tried Gemini and Kimi, but there’s really nothing else that’s competitive with the capability set of Claude Code right now.”
This is probably another marketing stunt. Turn off the flow of cocaine and have users find out how addicted they are. And they'll pay for the purest cocaine, not for second grade.
ceejayoz
It was always gonna be the Uber approach. Cheap and great turns to expensive and mediocre when they have to turn the money spigot on.
I made a quick site so you can see what tools are using the most context and help control it, totally free and in your browser.
https://claude-code-analysis.pages.dev/