Grok 4 Fast
28 comments
·September 20, 2025turblety
barrell
I would guess that it is essentially just a “grok 4 mini”, but if you use mini as the qualifier then most people will be inclined not to use it. If you call it fast then it gives people a reason to select it.
ActorNightly
Has it been "fixed" from its "liberal bias"?
glenstein
I was wondering that too. Anything reflecting human institutional knowledge writ large will, from Elons perspective, have "liberal bias" which is why he's also attacking Wikipedia and mainstream knowledge across the board. But other attempts at making non "biased" AI have barely done more than have custom instructions requesting that it "be conservative".
Unless you train it on Conservapedia or some equivalent corpus I'm not sure you'll be able to make it agree that "the Irish were the real slaves", that the D's and R's never realigned after the civil war, that the 2020 election was stolen and that gamergate was truly about ethics in journalism.
null
zone411
Matches Grok 4 at the top of the Extended NYT Connections leaderboard: https://github.com/lechmazur/nyt-connections/
RayVR
A faster model that outperforms its slower version on multiple benchmarks? Can anyone explain why that makes sense? Are they simply retraining on the benchmark tests?
yorwba
It doesn't outperform uniformly across benchmarks. It's worse than Grok 4 on GPQA Diamond and HLE (Humanity's Last Exam) without tools, both of which require the model to have memorized a large number of facts. Large (and thus slow) models typically do better on these.
The other benchmarks focus on reasoning and tool use, so the model doesn't need to have memorized quite so many facts, it just needs to be able to transform them from one representation to another. (E.g. user question to search tool call; list of search results to concise answer.) Larger models should in theory also be better at that, but you need to train them for those specific tasks first.
So I don't think they simply trained on the benchmark tests, but they shifted their training mix to emphasize particular tasks more, and now in the announcement they highlight benchmarks that test those tasks and where their model performs better.
You could also write an anti-announcement by picking a few more fact recall benchmarks and highlighting that it does worse at those. (I assume.)
NitpickLawyer
> Can anyone explain why that makes sense?
Can be anything from different arch, more data, RL, etc. It's probably RL. In recent months top tier labs seem to have "cracked" RL to a level not seen yet in open models, and by a large margin.
raincole
Just two different models branded under similar names. That's it. Grok 4 is not the slower version of Grok 4 Fast, just like gpt-4 is not the slower version of gpt-4o.
johnisgood
I think we all want fast AND accurate, is "AND accurate" true for this model? I would rather prefer to wait a few seconds more if the result is much more accurate.
mrklol
Pricing is really good for this benchmark value. Let’s see how it holds against people testing it.
NitpickLawyer
If this is sonoma-dusk that was on preview on openrouter, it's pretty cool. I've tested it with some code reverse engineering tasks, and it is at or above gpt5-mini level, while being faster. Works well till about 110-130k tokens tasks, then it gets the case of "getthereitis" and finishes the task even if not all constraints are met (i.e. will say I've solved x/400 tests, the rest can be done later)
mrklol
I can imagine, no model so far could actually use those context sizes…
hi_hi
I'm waiting for the Tesla FSD playbook to be rolled out for Grok. That is, launch something named like Grok AGI 1, wait for it to become obvious it isn't infact AGI, create a narrative redefining AGI, promise new AGI is 1 year away, and repeat for many years.
null
zozbot234
For the fastest performance, run it on Groq. /s
oulipo2
[flagged]
holoduke
Oh man. Life must be hard
kubb
I mean Elon is some kid of fascist adjacent pop figure. Let’s not pretend he’s not.
He’s open about wanting to use it to influence people’s behavior, and most likely disseminate his views.
AtlasBarfed
Standing on the precipice of AI assisted total information awareness and total authoritarian oppression?
Life is frightening right now.
andrewinardeer
So will you make it your daily driver then?
echelon
Musk / Politics question -
If half of the developers in the world hate Musk and refuse to use his company's tools, does that mean his tools won't be adopted enough to gain valuable market share?
Microsoft hate in the 90's didn't stop it. Could it stop X.ai now that the internet world is a bigger, more political place? Where ICs supposedly have more power to shape outcomes?
As another counter point, Amazon, Apple, Meta, and Google are often derided by loud voices (I'm occasionally one of them), but clearly none of that detracts from their combined trillion plus dollars of annual revenue.
I'm really curious to see how this shakes out.
RestartKernel
Did using Microsoft's tools ever feel like a political standpoint? Because I won't even consider pitching Grok to my employers/clients for that very reason.
mongol
Grok / X does not have the moat that Microsoft had.
raincole
> If half of the developers in the world hate Musk and refuse to use his company's tools
In general, developers use what tools their employers paid for.
AtlasBarfed
The banality of evil and the milgram experiment showed that employees will happily shoot the people that they're told to as well.
And the milgram experiment didn't even have subhuman classes and other such psychological manipulation and pre-biasing
Why would a model called "Fast" not advertise the tokens per second speed it performs at? Is "Fast" not representing speed, but another meaning? Is it too variable?