Skip to content(if available)orjump to list(if available)

Zed's Pricing Has Changed: LLM Usage Is Now Token-Based

indymike

I'm going through this with pricing for our product and the battle is:

Customer: wants predictable spend Sales/Marketing: wants an amazing product that is easy to sell (read does the mostest for the leastest) Advisors: want to see the SaaS model in all it's glory (i.e. 85% margin primarily from oversubscription of infrastructure) Finance: wants to make money and is really worried about variable costs ruining SaaS profit margin

A couple of thoughts:

1. AI-based cost is mostly variable. Also, things like messaging, phone minutes and so on are variable. Cloud expenses are also variable... There's a theme and it's different.

2. The unit of value most ai software is delivering is work that is being done by people or should be being done and is not.

3. Seems like it's time to make friends with a way to make money other than subscription pricing.

hmokiguess

Zed was supposed to be the answer to Atom / Sublime Text in my opinion, and I kinda do want to use it as my main driver, but it just isn’t there yet for me. It’s shameful because I like its aesthetics as a product more than the competition out there.

Just this other day I tried using it for something it sort of advertised itself as the superior thing, which was to load this giant text file I had instantly and let me work on it.

I then tried opening this 1GB text file to do a simple find/replace on it only to find macOS run out of system memory with Zed quickly using 20gb of memory for that search operation.

I then switched to vscode, which, granted opened it in a buffered sort of way and limited capability, but got the job done.

Maybe that was a me issue I don’t know, but aside from this one-off, it doesn’t have a good extensions support in the community for my needs yet. I hope it gets there!

klaussilveira

I feel like Zed stopped working on the editor itself since AI was rolled out. I also wanted it to be the open-source alternative to Sublime, but nothing comes close.

hakanensari

I think the recent push to delegate to CLI agents in the agent panel is the right direction. You can now run Claude Code has been running in Zed for the past month. Sure, there are SDK limitations and kinks to iron out, but it’s moving quickly. I’m into it.

hmokiguess

Yeah exactly this! I get they want to stay in the game and follow the market, but I’m sad they’re not being more aggressive on that original vision. I still think there could be a huge payoff for them if they invested more on their brand and aesthetics of a more polished and comfy editor.

The way I see it, we’re sort of living in a world where UX is king. (Looking at you Cursor)

I feel like there’s a general sentiment where folks just want a sense of home with their tools more than anything. Yes they need to work, but they also need to work for you in your way. Cursor reinvented autocomplete with AI and that felt like home for most, what’s next? I see so much focus on Agents but to me personally that feels more like it should live on the CI/CD layer of things. Editors are built for humans, something isn’t quite there yet, excited to see how it unfolds.

bombcar

If the full magnitude of products that stopped working on the main product and started to try to shoehorn AI in was known, the economy would collapse overnight.

laweijfmvo

Similar experience: I added a folder to my zed project that was too big, causing zed to lock up and eventually crash. But because the default setting was to re-open projects on launch, I was stuck in a loop where I couldn’t remove the folder either. Eventually found a way to clear the recent projects, load an empty editor, and change the setting to avoid it in the future.

mr90210

Ok, how big was your project?

My JetBrains IDEs (RustRover, Goland) probably would have choked out too.

j_bum

VSCode is my go to for large text file interaction on macOS.

TextEdit may be worth looking into as well? Haven’t tested it for large files before.

dewey

I have Sublime Text installed for the onlu use case of opening large files. Nothing comes close.

CharlesW

Googling around a bit, Sublime Text doesn't seem to be particularly good at this: https://forum.sublimetext.com/t/unable-to-open-a-large-text-...

In my experience, BBEdit will open files that kill other editors: "Handling large files presents no intrinsic problems for BBEdit, though some specific operations may be limited when dealing with files over 2GB in size."

mr90210

Speaking of aesthetics, I switched back to VSCode but I ended up installing the theme "Zed One Theme" and switched the editor's font to "IBM Plex Mono".

I know it's not Zed, but I am pretty satisfied with the results.

null

[deleted]

bluehatbrit

Token based pricing generally makes a lot of sense for companies like Zed, but it sure does suck for forecasting spend.

Usage pricing on something like aws is pretty easy to figure out. You know what you're going to use, so you just do some simple arithmetic and you've got a pretty accurate idea. Even with serverless it's pretty easy. Tokens are so much harder, especially when using it in a development setting. It's so hard to have any reasonable forecast about how a team will use it, and how many tokens will be consumed.

I'm starting to track my usage with a bit of a breakdown in the hope that I'll find a somewhat reliable trend.

I suspect this is going to be one of the next big areas in cloud FinOps.

garrickvanburen

My rant on token-based pricing is primarily based on the difficulty in consistently forecasting spend.....and also that the ongoing value of a token is controlled by the vendor...."the house always wins"

https://forstarters.substack.com/p/for-starters-59-on-credit...

coder543

[delayed]

scuff3d

Also seems like a great idea to create a business models where the companies aren't incentivised to provide the best product possible. Instead they'll want to create a product just useful enough to not drive away users, but just useless enough to temp people to go up a tier, "I'm so close, just one more prompt and it will be right this time!"

Edit: To be clear, I'm not talking about Zed. I'm talking about the companies make the models.

potlee

While Apple is incentivized to ship a smaller battery to cut costs, it is also incentivized to make their software efficient as possible to make the best use of the battery they do ship

Spartan-S63

> I suspect this is going to be one of the next big areas in cloud FinOps.

It already is. There’s been a lot of talk and development around FinOps for AI and the challenges that come with that. For companies, forecasting token usage and AI costs is non-trivial for internal purposes. For external products, what’s the right unit economic? $/token, $/agentic execution, etc? The former is detached from customer value, the latter is hard to track and will have lots of variance.

With how variable output size can be (and input), it’s a tricky space to really get a grasp on at this point in time. It’ll become a solved problem, but right now, it’s the Wild West.

prasoon2211

This is partially why, at least for LLM-assisted coding workloads, orgs are going with the $200 / mo Claude Code plans and similar.

jsheard

Until the rug inevitably gets pulled on those as well. It's not in your interest buy a $200/mo subscription unless you use >$200 of tokens per month, and long term it's not in their interest to sell you >$200 of tokens for a flat $200.

baq

meanwhile me hiding from accounting for spending $500 on cursor max mode in a day

Hamuko

The pricing model works as long as people (on average) think they need >$200 worth of tokens per month but actually do something less, like $170/month. Is that happening? No idea.

mdasen

I agree that tokens are a really hard metric for people. I think most people are used to getting something with a certain amount of capacity per time and dealing with that. If you get a server from AWS, you're getting a certain amount of capacity per time. You still might not know what it's going to cost you to do what you want - you might need more capacity to run your website than you think. But you understand the units that are being billed to you and it can't spiral out of control (assuming you aren't using autoscaling or something).

When you get Claude Code's $20 plan, you get "around 45 messages every 5 hours". I don't really know what that means. Does that mean I get 45 total conversations? Do minor followups count against a message just as much as a long initial prompt? Likewise, I don't know how many messages I'll use in a 5 hour period. However, I do understand when I start bumping up against limits. If I'm using it and start getting limited, I understand that pretty quickly - in the same way that I might understand a processor being slower and having to wait for things.

With tokens, I might blow through a month's worth of tokens in an afternoon. On one hand, it makes more sense to be flexible for users. If I don't use tokens for the first 10 days, they aren't lost. If I don't use Claude for the first 10 days, I don't get 2,160 message credits banked up. Likewise, if I know I'm going on vacation later, I can't use my Claude messages in advance. But it's just a lot easier for humans to understand bumping up against rate limits over a more finite period of time and get an intuition for what they need to budget for.

Filligree

Both prefill and decode count against Claude’s subscriptions; your conversations are N^2 in conversation length.

My mental model is they’re assigning some amount of API credits to the account and billing the same way as if you were using tokens, shutting off at an arbitrary point. The point also appears to change based on load / time of day.

jklinger410

Token based pricing works for the company, but not for the user.

dinobones

Making this prediction now, LLM pricing will eventually be priced in bytes.

Why: LLMs are increasingly becoming multimodal, so an image "token" or video "token" is not as simple as a text token. Also, it's difficult to compare across competitors because tokenization is different.

Eventually prices will just be in $/Mb of data processed. Just like bandwidth. I'm surprised this hasn't already happened.

jermaustin1

The problem is that tokens don't all equate to the same size. A megabyte of some random json is a LOT more tokens than a megabyte of "Moby Dick".

dragonwriter

> Why: LLMs are increasingly becoming multimodal, so an image "token" or video "token" is not as simple as a text token.

For autoregressive token-based multimodal models, image tokens are as straightforward as text tokens, and there is no reason video tokens wouldn’t also be. (If models also switch architecture and multimodal diffusion models, say, become more common, then, sure, a different pricing model more tied to actual compute cost drivers for that architecture are likely but... even that isn’t likely to be bytes.)

> Also, it's difficult to compare across competitors because tokenization is different.

That’s a reason for incumbents to prefer not to switch, though, not a reason for them to switch.

> Eventually prices will just be in $/Mb of data processed.

More likely they would be in floatint point operations expended processing them, but using tokens (which are the primary drivers for the current LLM architectures) will probably continue as long as the architecture itself is doninant.

oblio

> For autoregressive token-based multimodal models, image tokens are as straightforward as text tokens, and there is no reason video tokens wouldn’t also be.

In classical computing, there is a clear hierarchy: text < images <<< video.

Is there a reason why video computing using LLMs shouldn't be much more intensive and therefore costly than text or image output?

dragonwriter

No, it'll certainly be more expensive in any conceivable model that handles all three modalities, but if the model uses an architecture like current autoregressive, token-based multimodal LLMs/VLMs, tokens will make just as much sense as the basis for pricing, and be similarly straightforward, as with text and images.

Filligree

Of course it’s more expensive. It’s still tokens, but considerably more of them.

mhuffman

No one is going to give up token-based pricing. The main players can twiddle their models to make anything any amount of tokens they choose.

vtail

Hm... why not tokens as reported by each LLM provider? They already handle pricing for images etc.

null

[deleted]

jstummbillig

Why this instead of cpu/gpu time?

genshii

I'm personally looking forward to this change because I currently pay $20/month just to get edit prediction. I use Claude Code in my terminal for everything else. I do wish I could just pay for edit prediction at an even lower price, but I can understand why that's not an option.

I'm curious if they have plans to improve edit prediction though. It's honestly kind of garbage compared to Cursor, and I don't think I'm being hyperbolic by calling it garbage. Most of the time it's suggestions aren't helpful, but the 10-20% of the time it is helpful is worth the cost of the subscription for me.

chewz

I have never used Zed predictions but $20 for 500 prompts is quite a good deal. I use it mostly with Opus for some hard cases.

morgankrey

We have a significant investment underway in edit predictions. We hear you, more soon.

sippeangelo

This is the one thing keeping me from switching from Cursor. I much prefer Zed in every other way. Exciting!

hombre_fatal

Yeah, Cursor tab completion is basically in the realm of magical mind reading and might still be the most insane productivity demonstration of LLM tech in software.

It obsoleted making Vim macros and multiline editing for example. Now you just make one change and the LLM can derive the rest; you just press tab.

It's interesting that the Cursor team's first iteration is still better than anything I've seen in their competitors. It's been an amazing moat for a year(?) now.

pkilgore

This is very very exciting.

genshii

That's great to hear, thanks!

inerte

That's been my workflow also. Claude Code / OpenAI Codex most of the time, when I have to edit files Cursor's auto-complete is totally worth the $20.

WD-42

Good change. I’m not a vibe coder, I use Zed Pro llm integration more like glorified stack overflow. I value Zed more for being an amazing editor for the code I actually write and understand.

I suspect I’m not alone on this. Zed is not the editor for hardcore agentic editing and that’s fine. I will probably save money on this transition while continuing to support this great editor for what it truly shines at: editing source code.

sharkjacobs

This whole business model of trying to shave off or arbitrage a fraction of the money going to OpenAI and Anthropic just sucks. And it seems precarious. There's no honest way to resell tokens at a profit, and everyone knows it.

thelastbender12

Sorry, how is this new pricing anything but honest? They provide an editor you can use to - optimize the context you send to the LLM services - interact with the output that comes out of them

Why does not justify charging a fraction of your spend on the LLM platform? This is pretty much how every service business operates.

hu3

There's now greater incentive for Zed to stuff more content in the prompts to inflate tokens used and thus profit more. Or at least be less zealous.

This is not a new concern. And is not unique to Zed.

drakythe

For companies where that is their entire business model I absolutely agree. Zed is a solid editor with additional LLM integration features though, so this move would seem to me to just cover their costs + some LLM integration development funds. If their users don't want to use the LLM then no skin off Zed's back unless they've signed some guaranteed usage contract.

Havoc

>There's no honest way to resell tokens at a profit, and everyone knows it.

Agree with the sentiment, but I do think there are edge cases.

e.g. I could see a place like openrouter getting away with a tiny fractional markup based on the value they provide in the form of having all providers in one place

Lalabadie

The issue with a model like this (fixed small percentage) is that your biggest clients are the most incentivized to move away.

At scale, OpenRouter will instead get you the lower high-volume fees they themselves get from their different providers.

dinvlad

The whole business model even for OAI/Anthropic is unsustainable.. they are already running it at a huge loss atm, and will do for the foreseeable future. The economics simply doesn't work, unfortunately or not

vtail

Prediction: the only remaining providers of AI-assisted tools in a few years will be the LLM companies themselves (think claude code, codex, gemini, future xai/Alibaba/etc.), via CLIs + integrations such as ASP.

There is very little value that a company that has to support multiple different providers, such as Cursor, can offer on top of tailored agents (and "unlimited" subscription models) by LLM providers.

rudedogg

If you look at even the Claude/OpenAI chat UIs, they kind of suck. Not sure why you think someone else can't/won't do it better. Yes, the big players will copy what they can, but they also need to chase insane growth and getting every human on earth paying for an LLM subscription.

A tool that is good for everyone is great for no one.

Also, I think we're seeing the limits on "value" of a chat interface already. Now they're all chasing developers since there's a real potential to improve productivity (or sadly cut-costs) there. But even that is proving difficult.

serbuvlad

I recently started using Codex (OpenAI's Claude Code) and it has a VSCode extension that works like a charm. I tried out Windsurf a while ago. And the Codex extension simply does everything that Windsurf did. I guess it doesn't show changes at well, (it shows diffs in it's own window instead of in the file), but I can just check a git diff graphically (current state vs. HEAD) if I really wanted that.

I am really tempted to buy ChatGPT Pro, and probably would have if I lived in a richer country (unfortunetley purchase power parity doesn't equalize for tech products). The problem with Windsurf (and presumably Cursor and others) is that you buy the IDE subscription and then still have to worry about usage costs. With Codex/Claude Code etc., yeah, it's expensive, but, as long as you're within the usage limits, which are hopefully reasonable for the most expensive prices, you don't have to worry about it. AND you get the web and phone apps with GPT 5 Pro, etc.

computerex

I don't know. Foundation models are very good, and you can get a surprising amount of mileage from them by using them with low level interfaces. But personally I think companies building development tools of the future will use LLMs to build systems with increasing capabilities. I think a lot of engineering challenges remain in scaling LLM's to take over day to day in programming, and the current tools are scratching the surface of what's possible when you combine LLMs with traditional systems engineering.

prymitive

I can imagine the near future where companies “sponsor” open source projects by donating tokens to “mine” a PR for a feature they need.

hombre_fatal

But the reason LLMs aren't used to build features isn't because they are expensive.

The hard work is the high level stuff like deciding on the scope of the project, how it should fit in to the project, what kind of extensibility the feature might need to be built with, what kind of other components can be extended to support it, (and more), and then reviewing all the work that was done.

ebrescia

I love this! Finally a more direct way for companies to sponsor open source development. GitHub Sponsors helps, but it is often so vague where the funding is going.

scuff3d

If companies want to help they can just... I don't know... give projects some money

drakythe

Unless companies also donate money to sponsor the code review that will be required to be done by real human being I could see this idea being a problem for maintainers. Yes you have to code review a human being as well but a human being is capable of learning and carrying that learning forward and their next PR will be better, as well as being able to look at past PRs to evaluate whether the user is a troll/bad actor or someone who genuinely wants to assist with the project. An LLM won't learn and will always spit out valid _looking_ code.

bsnnkv

More often than not, for individuals, it's barely contributing to their living costs

giancarlostoro

I was just thinking this morning about how I think Zed should rethink their subscription because its a bit pricey if they're going to let you just use Claude Code. I am in the process of trying out Claude and figured just going to them for the subscriptions makes more sense.

I think Zed had a lot of good concepts where they could make paid AI benefits optional longer term. I like that you can join your devs to look at different code files and discuss them. I might still pay for Zed's subscription in order to support them long term regardless.

I'm still upset so many hosted models dont just let you use your subscription on things like Zed or JetBrains AI, what's the point of a monthly subscription if I can only use your LLM in a browser?

hamandcheese

> I'm still upset so many hosted models dont just let you use your subscription on things like Zed or JetBrains AI, what's the point of a monthly subscription if I can only use your LLM in a browser?

This is her another reason why CLI-based coding agents will win. Every editor out there trying to be the middle man between you and an AI provider is nuts.

oblio

Wouldn't the last step just be an API? That would allow direct integration from everywhere.

oakesm9

I completely get why this pricing is needed and it seems fair. There’s a major flaw in the announcement though.

I get that the pro plan has $5 of tokens and the pricing page says that a token is roughly 3-4 characters. However, it is not clear:

- Are tokens input characters, output characters, or both?

- What does a token cost? I get that the pricing page says it varies by model and is “ API list price +10%”, but nowhere does it say what these API list prices are. Am I meant to go to The OpenAI, Anthropic, and other websites to get that pricing information? Shouldn’t that be in a table on that page which each hosted model listed?

I’m only a very casual user of AI tools so maybe this is clear to people deep in this world, but it’s not clear to me just based on Zelda pricing page exactly how far $5 per month will get me.

morgankrey

List here: https://zed.dev/docs/ai/models. Thanks for the feedback, we'll make sure this is linked from the pricing page. Think it got lost in the launch shuffle.

oakesm9

All makes sense. I presumed it was an oversight.

It’s hard for me to conceptualise what a million tokens actually looks like, but I don’t think there’s a way around that aside from making proving some concrete examples of inputs, outputs, and the number of tokens that actually is. I guess it would become clearer after using it a bit.

morgankrey

Now live: https://zed.dev/pricing#what-is-a-token. Thanks for the feedback

rdtsc

I am wondering why they couldn't have foreseen this. Was it really a failure to predict the need to charge for tokens eventually, or was planned from the start that way -- get people to use the unlimited option for a bit, they get hooked, then switch them to per-token subscriptions.