Cursor IDE support hallucinates lockout policy, causes user cancellations
634 comments
·April 14, 2025nerdjon
There is a certain amount of irony that people try really hard to say that hallucinations are not a big problem anymore and then a company that would benefit from that narrative gets directly hurt by it.
Which of course they are going to try to brush it all away. Better than admitting that this problem very much still exists and isn’t going away anytime soon.
lynguist
https://www.anthropic.com/research/tracing-thoughts-language...
The section about hallucinations is deeply relevant.
Namely, Claude sometimes provides a plausible but incorrect chain-of-thought reasoning when its “true” computational path isn’t available. The model genuinely believes it’s giving a correct reasoning chain, but the interpretability microscope reveals it is constructing symbolic arguments backward from a conclusion.
https://en.wikipedia.org/wiki/On_Bullshit
This empirically confirms the “theory of bullshit” as a category distinct from lying. It suggests that “truth” emerges secondarily to symbolic coherence and plausibility.
This means knowledge itself is fundamentally symbolic-social, not merely correspondence to external fact.
Knowledge emerges from symbolic coherence, linguistic agreement, and social plausibility rather than purely from logical coherence or factual correctness.
emn13
While some of what you say is an interesting thought experiment, I think the second half of this argument has, as you'd put it, a low symbolic coherence and low plausibility.
Recognizing the relevance of coherence and plausibility does not need to imply that other aspects are any less relevant. Redefining truth merely because coherence is important and sometimes misinterpreted is not at all reasonable.
Logically, a falsehood can validly be derived from assumptions when those assumptions are false. That simple reasoning step alone is sufficient to explain how a coherent-looking reasoning chain can result in incorrect conclusions. Also, there are other ways a coherent-looking reasoning chain can fail. What you're saying is just not a convincing argument that we need to redefine what truth is.
learningstud
Validity is not soundness. Wonder why people are just beginning to realize what logicians have been studying for more than a century. This goes to show that most programming was never based on logic but vibes. People have been vibe coding with themselves before AI became prominent.
dcow
For this to be true everyone must be logically on the same page. They must share the same axioms. Everyone must be operating off the same data and must not make mistakes or have bias evaluating it. Otherwise inevitably sometimes people will arrive at conflicting truths.
In reality it’s messy and not possible with 100% certainty to discern falsehoods and truthoods. Our scientific method does a pretty good job. But it’s not perfect.
You can’t retcon reality and say “well retrospectively we know what happened and one side was just wrong”. That’s called history. It’s not useful or practical working definition of truth when trying to evaluate your possible actions (individually, communally, socially, etc) and make a decision in the moment.
I don’t think it’s accurate to say that we want to redefine truth. I think more accurately truth has inconvenient limitations and it’s arguably really nice most of the time to ignore them.
jimbokun
> Knowledge emerges from symbolic coherence, linguistic agreement, and social plausibility rather than purely from logical coherence or factual correctness.
This just seems like a redefinition of the word "knowledge" different from how it's commonly used. When most people say "knowledge" they mean beliefs that are also factually correct.
indigo945
As a digression, the definition of knowledge as justified true belief runs into the Gettier problems:
> Smith [...] has a justified belief that "Jones owns a Ford". Smith
> therefore (justifiably) concludes [...] that "Jones owns a Ford, or Brown
> is in Barcelona", even though Smith has no information whatsoever about
> the location of Brown. In fact, Jones does not own a Ford, but by sheer
> coincidence, Brown really is in Barcelona. Again, Smith had a belief that
> was true and justified, but not knowledge.
Or from 8th century Indian philosopher Dharmottara: > Imagine that we are seeking water on a hot day. We suddenly see water, or so we
> think. In fact, we are not seeing water but a mirage, but when we reach the
> spot, we are lucky and find water right there under a rock. Can we say that we
> had genuine knowledge of water? The answer seems to be negative, for we were
> just lucky.
More to the point, the definition of knowledge as linguistic agreement is convincingly supported by much of what has historically been common knowledge, such as the meddling of deities in human affairs, or that the people of Springfield are eating the cats.dcow
I don’t think it’s so clear cut… Even the most adamant “facts are immutable” person can agree that we’ve had trouble “fact checking” social media objectively. Fluoride is healthy, meta analysis of the facts reveals fluoride may be unhealthy. The truth of the matter is by and large what’s socially cohesive for doctors’ and dentists’ narrative, that “fluoride is fine any argument to the contrary—even the published meta-analysis—is politically motivated nonsense”.
CodesInChaos
> The model genuinely believes it’s giving a correct reasoning chain, but the interpretability microscope reveals it is constructing symbolic arguments backward from a conclusion.
Sounds very human. It's quite common that we make a decision based on intuition, and the reasons we give are just post-hoc justification (for ourselves and others).
RansomStark
> Sounds very human
well yes, of course it does, that article goes out of its way to anthropomorphize LLMs, while providing very little substance
jimbokun
Isn't the point of computers to have machines that improve on default human weaknesses, not just reproduce them at scale?
learningstud
Exactly, most of us behave in almost the same as AI does. We finally have a mirror to reflect upon.
throwway120385
The other very human thing to do is invent disciplines of thought so that we don't just constantly spew bullshit all the time. For example you could have a discipline about "pursuit of facts" which means that before you say something you mentally check yourself and make sure it's actually factually correct. This is how large portions of the populace avoid walking around spewing made up facts and bullshit. In our rush to anthropomorphize ML systems we often forget that there are a lot of disciplines that humans are painstakingly taught from birth and those disciplines often give rise to behaviors that the ML-based system is incapable of like saying "I don't know the answer to that" or "I think that might be an unanswerable question."
jerf
In a way, the main problem with LLMs isn't that they are wrong sometimes. We humans are used to that. We encounter people who are professionally wrong all the time. Politicians, con-men, scammers, even people who are just honestly wrong. We have evaluation metrics for those things. Those metrics are flawed because there are humans on the other end intelligently gaming those too, but generally speaking we're all at least trying.
LLMs don't fit those signals properly. They always sound like an intelligent person who knows what they are talking about, even when spewing absolute garbage. Even very intelligent people, even very intelligent people in the field of AI research are routinely bamboozled by the sheer swaggering confidence these models convey in their own results.
My personal opinion is that any AI researcher who was shocked by the paper lynguist mentioned ought to be ashamed of themselves and their credulity. That was all obvious to me; I couldn't have told you the exact mechanism the arithmetic was being performed (though what is was doing was well in the realm of what I would have expected from a linguistic AI trying to do math), but the fact that its chain of reasoning bore no particular resemblance to how it drew its conclusions was always obvious. A neural net has no introspection on itself. It doesn't have any idea "why" it is doing what it is doing. It can't. There's no mechanism for that to even exist. We humans are not directly introspecting our own neural nets, we're building models of our own behavior and then consulting the models, and anyone with any practice doing that should be well aware of how those models can still completely fail to predict reality!
Does that mean the chain of reasoning is "false"? How do we account for it improving performance on certain tasks then? No. It means that it is occurring at a higher level and a different level. It is quite like humans imputing reasons to their gut impulses. With training, combining gut impulses with careful reasoning is actually a very, very potent way to solve problems. The reasoning system needs training or it flies around like an unconstrained fire hose uncontrollably spraying everything around, but brought under control it is the most powerful system we know. But the models should always have been read as providing a rationalization rather than an explanation of something they couldn't possibly have been explaining. I'm also not convinced the models have that "training" either, nor is it obvious to me how to give it to them.
(You can't just prompt it into a human, it's going to be more complicated than just telling a model to "be carefully rational". Intensive and careful RHLF is a bare minimum, but finding humans who can get it right will itself be a challenge, and it's possible that what we're looking for simply doesn't exist in the bias-set of the LLM technology, which is my base case at this point.)
jmaker
I haven’t used Cursor yet. Some colleagues have and seemed happy. I’ve had GitHub Copilot on for what feels like a couple years, a few days ago VS Code was extended to provide an agentic workflow, MCP, bring-your-own-key, it interprets instructions in a codebase. But the UX and the outputs are bad in over 3/4 of cases. It’s a nuisance to me. It injects bad code even though it has the full context. Is Cursor genuinely any better?
To me it feels like people that benefit from or at least enjoy that sort of assistance and I solve vastly different problems and code very differently.
I’ve done exhausting code reviews on juniors’ and middles’ PRs but what I’ve been feeling lately is that I’m reviewing changes introduced by a very naive poster. It doesn’t even type-check. Regardless of whether it’s Claude 3.7, o1, o3-mini, or a few models from Hugging Face.
I don’t understand how people find that useful. Yesterday I literally wasted half an hour for a test suite setup a colleague of mine introduced to the codebase that wasn’t good, and I tried delegating that fix to several of the Copilot models. All of them missed the point, some even introduced security vulnerabilities in the process invalidating JWT validation, I tried “vide coding” it till it works, until I gave up in frustration and just used an ordinary search engine, which led me to the docs, in which I immediately found the right knob. I reverted all that crap and did the simple and correct thing. So my conclusion was simple: vibe coding and LLMs made the codebase unnecessarily more complicated and wasted my time. How on earth do people code whole apps with that?
trilbyglens
I think it works until it doesn't. The nature of technical debt of this kind means you can sort of coast on things until the complexity of the system reaches such a level that it's effectively painted into a corner, and nothing but a massive teardown will do as a fix.
ScottBurson
> The model genuinely believes it’s giving a correct reasoning chain
The model doesn't "genuinely believe" anything.
nickledave
Yes
https://link.springer.com/article/10.1007/s10676-024-09775-5
> # ChatGPT is bullshit
> Recently, there has been considerable interest in large language models: machine learning systems which produce human-like text and dialogue. Applications of these systems have been plagued by persistent inaccuracies in their output; these are often called “AI hallucinations”. We argue that these falsehoods, and the overall activity of large language models, is better understood as bullshit in the sense explored by Frankfurt (On Bullshit, Princeton, 2005): the models are in an important way indifferent to the truth of their outputs. We distinguish two ways in which the models can be said to be bullshitters, and argue that they clearly meet at least one of these definitions. We further argue that describing AI misrepresentations as bullshit is both a more useful and more accurate way of predicting and discussing the behaviour of these systems.
skrebbel
Offtopic but I'm still sad that "On Bullshit" didn't go for that highest form of book titles, the single noun like "Capital", "Sapiens", etc
mvieira38
Starting with "On" is cooler in philosophical tradition, though, starting in classical and medieval times, e.g. On Interpretation, On the Heavens, etc by Aristotle, De Veritate, De Malo, etc. by Aquinas. Capital is actually "Das Kapital", too
tomaskafka
[flagged]
ModernMech
It's a huge problem. I just can't get past it and I get burned by it every time I try one of these products. Cursor in particular was one of the worst; the very first time I allowed it to look at my codebase, it hallucinated a missing brace (my code parsed fine), "helpfully" inserted it, and then proceeded to break everything. How am I supposed to trust and work with such a tool? To me, it seems like the equivalent of lobbing a live hand grenade into your codebase.
Don't get me wrong, I use AI every day, but it's mostly as a localized code complete or to help me debug tricky issues. Meaning I've written and understand the code myself, and the AI is there to augment my abilities. AI works great if it's used as a deductive tool.
Where it runs into issues is when it's used inductively, to create things that aren't there. When it does this, I feel the hallucinations can be off the charts -- inventing APIs, function names, entire libraries, and even entire programming languages on occasion. The AI is more than happy to deliver any kind of information you want, no matter how wrong it is.
AI is not a tool, it's a tiny Kafkaesque bureaucracy inside of your codebase. Does it work today? Yes! Why does it work? Who can say! Will it work tomorrow? Fingers crossed!
yodsanklai
You're not supposed to trust the tool, you're supposed to review and rework the code before submitting for external review.
I use AI for rather complex tasks. It's impressive. It can make a bunch of non-trivial changes to several files, and have the code compile without warnings. But I need to iterate a few times so that the code looks like what I want.
That being said, I also lose time pretty regularly. There's a learning curve, and the tool would be much more useful if it was faster. It takes a few minutes to make changes, and there may be several iterations.
ryandrake
> You're not supposed to trust the tool, you're supposed to review and rework the code before submitting for external review.
It sounds like the guys in this article should not have trusted AI to go fully open loop on their customer support system. That should be well understood by all "customers" of AI. You can't trust it to do anything correctly without human feedback/review and human quality control.
schmichael
> You're not supposed to trust the tool
This is just an incredible statement. I can't think of another development tool we'd say this about. I'm not saying you're wrong, or that it's wrong to have tools we can't just, just... wow... what a sea change.
gtirloni
1) Once you get it to output something you like, do you check all the lines it changed? Is there a threshold after which you just... hope?
2) No matter what the learning curve, you're using a statistical tool that outputs in probabilities. If that's fine for your workflow/company, go for it. It's just not what a lot of developers are okay with.
Of course it's a spectrum with the AI deniers in one corner and the vibe coders in the other. I personally won't be relying 100% on a tool and letting my own critical thinking atrophy, which seems to be happening, considering recent studies posted here.
bigstrat2003
> You're not supposed to trust the tool, you're supposed to review and rework the code before submitting for external review.
Then it's not a useful tool, and I will decline to waste time on it.
jorvi
> But I need to iterate a few times so that the code looks like what I want.
The LLM too. You can get a pretty big improvement by telling the LLM to "iterate 4 times on whichever code I want you to generate, but only show me the final iteration, and then continue as expected".
I personally just inject the request for 4 iterations into the system prompt.
mrheosuper
If i dont trust my tool, i would never use it, or use something else better
null
e3bc54b2
> You're not supposed to trust the tool, you're supposed to review and rework the code before submitting for external review.
The vibe coding guy said to forget the code exists and give in to vibes, letting the AI 'take care' of things. Review and rework sounds more like 'work' and less like 'vibe'.
/s
mediaman
I'd add that the deductive abilities translate to well-defined spec. I've found it does well when I know what APIs I want it to use, and what general algorithmic approaches I want (which are still sometimes brainstormed separately with an AI, but not within the codebase). I provide it a numbered outline of the desired requirements and approach to take, and it usually does a good job.
It does poorly without heavy instruction, though, especially with anything more than toy projects.
Still a valuable tool, but far from the dreamy autonomous geniuses that they often get described as.
Mountain_Skies
Versioning in source control for even personal projects just got far more important.
AdrianEGraphene
It's wild how people write without version control... Maybe I'm missing something.
o11c
Thankfully modern source control doesn't reuse user-supplied filenames for its internals. In the dark ages, I destroyed more than one checkout using commands of the form:
find -name '*somepattern*' -exec clobbering command ...
skissane
> the very first time I allowed it to look at my codebase, it hallucinated a missing brace (my code parsed fine), "helpfully" inserted it, and then proceeded to break everything.
This is not an inherent flaw of LLMs, rather it is a flaw of a particular implementation-if you use guided sampling, so during sampling you only consider tokens allowed by the programming language grammar at that position, it becomes impossible for the LLM to generate ungrammatical output
> When it does this, I feel the hallucinations can be off the charts -- inventing APIs, function names, entire libraries,
They can use guided sampling for this too - if you know the set of function names which exist in the codebase and its dependencies, you can reject tokens that correspond to non-existent function names during sampling
Another approach, instead of or as well as guided sampling, is to use an agent with function calling - so the LLM can try compiling the modified code itself, and then attempt to recover from any errors which occur.
simeonww
[flagged]
theonething
> it hallucinated a missing brace (my code parsed fine), "helpfully" inserted it, and then proceeded to break everything.
Your tone is rather hyperbolic here, making it sound like an extra brace resulted in a disaster. It didn't. It was easy to detect and easy to fix. Not a big deal.
ModernMech
It's not a big deal in the sense that it's easily reversed, but it is a big deal in that it means the tool is unpredictably unhelpful. Of the properties that good tools in my workflow possess, "unpredictably unhelpful" does not make the top 100.
When a tool starts confidently inserting random wrong code into my 100% correct code, there's not much more I need to see to know it's not a tool for me. That's less like a tool and more like a vandal. That's not something I need in my toolbox, and I'm certainly not going to replace my other tools with it.
cryptoegorophy
I think that’s why Apple is very slow at rolling out AI if it ever actually will. Downside is way too big than the upside.
saintfire
You say slowly, but in my opinion Apple made an out of character misstep by releasing a terrible UX to everyone. Apple intelligence is a running joke now.
Yes they didn't push it as hard as, say, copilot. I still think they got in way too deep way too fast.
devmor
This is not the first time that Apple has released a terrible UX that very few users liked, and it certainly wont be the last.
I don’t necessarily agree with the post you’re responding to, but what I will give Apple credit for is making their AI offering unobtrusive.
I tried it, found it unwanted and promptly shut it off. I have not had to think about it again.
Contrast that with Microsoft Windows, or Google - both shoehorning their AI offering into as many facets of their products as possible, not only forcing their use, but in most cases actively degrading the functionality of the product in favor of this required AI functionality.
throwaway2037
> Apple made an out of character misstep by releasing a terrible UX to everyone
What about Apple Maps? That roll-out was awful.manmal
Remember „You are a bad user, I am a good bing“? Apple is just slower in fixing and improving things.
stogot
Fast!? They were two years slow and still fell face flat, and then rolled back the software
miki123211
Apple made a huge mistake by keeping their commitment to "local first" in the age of AI.
The models and devices just aren't quite there yet.
Once Google gets its shit together and starts deploying (cloud--based) AI features to Android devices en masse, Apple is going to have a really big problem on their hands.
Most users say that they want privacy, but if privacy comes in the way of features or UX, they choose the latter. Successful privacy-respecting companies (Apple, Signal) usually understand this, it's why they're successful, but I think Apple definitely chose the wrong tradeoff here.
zdragnar
Investors seem to be starved for novelty right now. Web 2.0 is a given, web 3.0 is old, crypto has lost the shine, all that's left to jump on at the moment is AI.
Apple fumbled a bit with Siri, and I'm guessing they're not too keen to keep chasing everyone else, since outside of limited applications it turns out half baked at best.
Sadly, unless something shinier comes along soon, we're going to have to accept that everything everywhere else is just going to be awful. Hallucinations in your doctor's notes, legal rulings, in your coffee and laundry and everything else that hasn't yet been IoT-ified.
timr
> we're going to have to accept that everything everywhere else is just going to be awful. Hallucinations in your doctor's notes, legal rulings, in your coffee and laundry and everything else that hasn't yet been IoT-ified.
I installed a logitech mouse driver (sigh) the other day, and in addition to being obtrusive and horrible to use, it jams an LLM into the UI, for some reason.
AI has reached crapware status in record time.
pcthrowaway
> Hallucinations in your doctor's notes, legal rulings, in your coffee
"OK Replicator, make me one espresso with creamer"
"Making one espresso with LSD"
VenturingVole
"all that's left to jump on at the moment is AI" -> No, it's the effective applications of AI. It's unprecedented.
I was in the VC space for a while previously, most pitch decks claimed to be using AI: But doing even the briefest of DD - it was generally BS. Now it's real.
With respect to everything being awful: One might say that's always been the case. However, now there's a chance (and requirement) to build in place safeguards/checks/evals and massively improve both speed and quality of services through AI.
Don't judge for the problems: Look at the exponential curve, think about how to solve the problems. Otherwise, you will get left behind.
sillyfluke
They already rolled out an "AI" product. Got humiliated pretty bad, and rolled it back. [0]
VenturingVole
They had an opportunity to actually adapt, to embrace getting rapid feedback/iterating: But they are not equipped for it culturally. Major lost opportunity as it could have been a driver of internal change.
I'm certain they'll get it right soon enough though. People were writing off Google in terms of AI until this year.. and oh how attitudes have changed.
furyofantares
They also have text thread and email summaries. I still think it counts as a slow rollout.
jmaker
Even the iOS and macOS typing correction engine has been getting worse for me over the past few OS updates. I’m now typing this on iOS, and it’s really annoying how it injects completely unrelated words, replaces minor typos with completely irrelevant words. Same in Safari on macOS. The previous release felt better than now, but still worse than a couple years ago.
TylerE
It’s not just you. iOS auto correct has gotten damn near malicious. E seen it insert entire words out of nowhere
gambiting
>>if it ever actually will.
If they don't then I'd hope they get absolutely crucified by trade comissions everywhere, currently there are bilboards in my city advertising Apple AI even though it doesn't even exist yet - if it's never brought to the market then it's a serious case of misleading advertising.
poink
Yet Apple has reenabled Apple Intelligence multiple times on my devices after OS updates despite me very deliberately and angrily disabling it multiple times
m3kw9
When you got 1-2billion users a day doing maybe 10 billion prompts a day, it’s risky
anonzzzies
Did anyone say that? They are an issue everywhere, including for code. But with code at least I can have tooling to automatically check and feed back that it hallucinated libraries, functions etc, but with just normal research / problems there is no such thing and you will spend a lot of time verifying everything.
threeseed
I use Scala which has arguably the best compiler/type system with Cursor.
There is no world in which a compiler or tooling will save you from the absolute mayhem it can do. I’ve had it routinely try to re-implement third party libraries, modify code unrelated to what it was asked, quietly override functions etc.
It’s like a developer who is on LSD.
Tainnor
I don't know Scala. I asked cursor to create a tutorial for me to learn Scala. It created two files for me, Basic.scala and Advanced.scala. The second one didn't compile and no matter how often I tried to paste the error logs into the chat, it couldn't fix the actual error and just made up something different.
Terr_
Yeah, everyone wanted a thinking machine, but the best we can do right now is a dreaming machine... And dreams don't have to make sense.
larodi
Developer on LSD is likely to hallucinate less in terms of how weird the LLM hallucinations are sometimes. Besides I know people, not myself, who fare very well on LSD and particularly when micro dosing Adderal style
jmaker
Granted the Scala language is much more complex than Go. To produce something useful it must be capable of an equivalent of parsing the AST.
felipefar
Yes, most people who have an incentive in pushing AI say that hallucinations aren't a problem, since humans aren't correct all the time.
But in reality hallucinations either make people using AI lose a lot of their time trying to stuck the LLMs from dead ends or render those tools unusable.
Gormo
> Yes, most people who have an incentive in pushing AI say that hallucinations aren't a problem, since humans aren't correct all the time.
Humans often make factual errors, but there's a difference between having a process to validate claims against external reality, and occasionally getting it wrong, and having no such process, with all output being the product of internal statistical inference.
The LLM is engaging in the same process in all cases. We're only calling it a "hallucination" when its output isn't consistent with our external expectations, but if we regard "hallucination" as referring to any situation where the output for a wholly endogenous process is mistaken for externally validated information, then LLMs are only ever hallucinating, and are just designed in such a way that what they hallucinate has a greater than chance likelihood of representing some external reality.
jimbokun
> Yes, most people who have an incentive in pushing AI say that hallucinations aren't a problem, since humans aren't correct all the time.
We have legal and social mechanisms in place for the way humans are incorrect. LLMs are incorrect in new ways that our legal and social systems are less prepared to handle.
If a support human lies about a change to policy, the human is fired and management communicates about the rogue actor, the unchanged policy, and how the issue has been handled.
How do you address an AI doing the same thing without removing the AI from your support system?
manmal
You get some superficial checking by the compiler and test cases, but hallucinations that pass both are still an issue.
anonzzzies
Absolutely, but at least you have some lines of defence while with real world info you have nothing. And the most offending stuff like importing a package that doesn't exist or using a function that doesn't exist does get caught and can be auto fixed.
rini17
Except when the hallucinated library exists and it's malicious. This is actually happening. Without AI, by using plain google you are less likely to fall for that (so far).
jmaker
Until the model injects a subtle change to your logic that does type-check and then goes haywire in production. Just takes a colleague of yours under pressure and another one to review the PR, and then you’re on call and they out sick or on vacation.
learningstud
People hallucinate all the time out of pressure or habit. We don't need AI for that. It's hard to tell most people from AI. Most people would fail Turing tests as subjects.
_jonas
I see this fallacy often too.
My company provides hallucination detection software: https://cleanlab.ai/tlm/
But we somehow end up in sales meetings where the person who requested the meeting claims their AI does not hallucinate ...
mntruell
(Cursor cofounder)
Apologies - something very clearly went wrong here. We’ve already begun investigating, and some very early results:
* Any AI responses used for email support are now clearly labeled as such. We use AI-assisted responses as the first filter for email support.
* We’ve made sure this user is completely refunded - least we can do for the trouble.
For context, this user’s complaint was the result of a race condition that appears on very slow internet connections. The race leads to a bunch of unneeded sessions being created which crowds out the real sessions. We’ve rolled out a fix.
Appreciate all the feedback. Will help improve the experience for future users.
nextaccountic
Why did you remove this thread?
https://old.reddit.com/r/cursor/comments/1jyy5am/psa_cursor_...
(For reference, here it is in reveddit https://www.reveddit.com/v/cursor/comments/1jyy5am/psa_curso... - text from post was unfortunately not saved)
It's already locked and with a stickied comment from a dev clarifying what happened
Did you remove it so people can't find about this screwup when searching Google?
Anyway, if you acknowledge it was a mistake to remove the thread, could you please un-remove it?
PrayagS
The whole subreddit is moderated poorly. I’ve seen plenty of users post on r/LocalLlama about how something negative or constructive they said on the Cursor sub was just removed.
AyyEye
Why would anyone trust you?
The best case scenario is that you lied about having people answer support. LLMs pretending to be people (you named it Sam!) and not labeled as such is clearly intended to be deceptive. Then you tried to control the narrative on reddit. So forgive me if I hit that big red DOUBT button.
Even in your post you call it "AI-assisted responses" which is as weaselly as it gets. Was it a chatbot response or was a human involved?
But 'a chatbot messed up' doesn't explain how users got locked out in the first place. EDIT: I see your comment about the race condition now. Plausible but questionable.
So the other possible scenario is that you tried to hose your paying customers then when you saw the blowback blamed it on a bot.
'We missed the mark' is such a trope non-apology. Write a better one.
I had originally ended this post with "get real" but your company's entire goal is to replace the real with the simulated so I guess "you get what you had coming". Maybe let your chatbots write more crap code that your fake software engineers push to paying customers that then get ignored and/or lied to when they ask your chatbots for help. Or just lie to everyone when you see blowback. Whatever. Not my problem yet because I can write code well enough that I'm embarrassed for my entire industry whenever I see the output from tools like yours.
This whole "AI" psyop is morally bankrupt and the world would be better off without it.
PoignardAzur
> The best case scenario is that you lied about having people answer support. LLMs pretending to be people (you named it Sam!) and not labeled as such is clearly intended to be deceptive.
Also, illegal in the EU.
jackaroe420
I don't know who you are but you said this so well!
Azeralthefallen
Hi since i know you will never respond to this or hear this.
We spent almost 2 months fighting with you guys about basic questions any B2B SaaS should be able to answer us. Things such as invoicing, contracts, and security policies. This was for a low 6 figure MRR deal.
When your sales rep responds "I don't know" or "I will need to get back to you" for weeks about basic questions it left us with a massive disappointment. Please do better, however we have moved to Copilot.
dspillett
> Any AI responses used for email support are now clearly labeled as such.
Because we all know how well people pay attention to such clear labels, even seasoned devs not just “end users”⁰.
Also, deleting public view of the issue (locking & hiding the reddit thread) tells me a lot about how much I should trust the company and its products, and as such I will continue to not use them.
--------
[0] though here there the end users are devs
Snakes3727
I do truely love how you guys even went so far to hide and lock the post from Reddit.
This person is not the only one to experiencing this bug. As this thread has pointed out.
KennyBlanken
I wish more people realized that virtually any subreddit for a company or product is run by the company - either directly or via a firm that specializes in 'sentiment analysis and management' or whatever the marketdroids call it these days. Even if they don't remove posts via moderation, they'll just hammer it with downvotes from sockpuppet accounts.
HN goes a step further. It has a function that allows moderators to kill or boost a post by subtracting or adding a large amount to the post's score. HN is primarily a place for Y Combinator to hype their latest venture, and a "safe" place for other startups and tech companies.
Snakes3727
Yes and it irritates the hell out of me. Cursor support is garbage, but issues with billing and other things are so much worse.
The team I work with it took nearly 3 months to get basic questions answered correctly when it came to a sales contract. They never gave our Sec team acceptable answers around privacy and security.
thinkingemote
I've always wondered how Reddit can make money from these companies. I agree they are literally everywhere, even in non-company specific but generic subreddits where if it's big enough you might have multiple shadow marketing firms competing to push their products (e.g. AI, movies, food, porn etc).
Reddit is free to play for marketing firms. Perhaps they could add extra statistics, analytics, promotions for these commercial users.
patcon
Agreed, this is what's infuriating: insistence on control.
They will utterly fail to build for a community of users if they don't have anyone on-hand who can tell them what a terrible idea that was
To the cofounder: hire someone (ideally with some thoughtful reluctance around AI, who understands what's potentially lost in using it) who will tell you your ideas around this are terrible. Hire this person before you fuck up your position in benevolent leadership of this new field
petesergeant
I dunno, that seems pretty reasonable to me simply for stopping the spread of misinformation. The main story will absolutely get written up by some smaller news sources, but is it really a benefit for someone facing a similar issue in the future to find an outdated and probably confusing Reddit post about it?
slotrans
> We use AI-assisted responses as the first filter for email support.
Literally no one wants this. The entire purpose of contacting support is to get help from a human.
fragmede
Sorta? I mean I want my problem fixed, regardless of it it's a person or not. Having a person listen to me complain about my problems might sooth my conscience, but I can't pay my bill or why was it so high; having those answered by a system that is contextualized to my problem sand is empowered to fix it, and not just a talking to a brick wall? I wouldn't say totally fine, but at the end of the day, if my problem is solved or my query, even if it's weird, I can't say I really needed for the voice on the other end of the pHone to come from a human. If a companies business model isn't sustainable without using AI agents, it's not really my problem that it's not, but also if I'm using their product, presumably I don't want that to go away.
conartist6
Isn't the real scary thing here that the AI agent is empowered to control your life?
You're imagining that if you get the answer you want from the AI you hang up the phone, and if you don't you're imagining a human will pick up and have the political power and will to overrule the AI. I think it's more realistic is the way things have played out here: nobody took or had any responsibility because "they made the AI responsible" and second-guessing that choice isn't second-guessing the AI, it's second-guessing the human leaders who decreed that human support had no value. This with the result that the humans would let the AI go about as far as setting fire to the building before some kind of human element imbued with any real accountability steps in. The evidence of this is all the ignored requests presented here.
hartator
> For context, this user’s complaint was the result of a race condition that appears on very slow internet connections.
Seems like you are still blaming the user for his “very slow internet”.
How do you know the user internet was slow? Couldn’t a race condition like this exist anyway with regular 2 fast internet connections competing for the same sessions?
Something doesn’t add up.
mritchie712
huh?
this is a completely reasonable and seemingly quite transparent explaination.
if you want a conspiracy, there are better places to look.
ben0x539
When admitting fault with your a PR hat on after pissing off a decent(?) number of your paying customers, you're supposed to fully fall on your own sword, not assign blame to factors outside of your control.
Instead of saying "race condition that appears on very slow internet connections", you might say "race condition caused by real-world network latencies that our in-office testing didn't reveal" or some shit.
hartator
I don't think you are being transparent.
Like it all sounds like a business decision (limiting 1 device to 1 sub) which is actually that was confirmed both by the actual tech limitation (logging out users from their other devices) and your own support.
Blaming the AI, then the user connection, and then some odd race conditions seem unnecessary. You can just say it was a bad business decision, roll it back, and go on with your day. Instead of treating multiple paying users badly.
eranation
Side note... I'm a paying enterprise customer who moved all my team to cursor and have to say I'm considering canceling due to the non existent support. For example Cursor will create new files instead of edit an existing one when you have a workspace with multiple folders in a monorepo...
geuis
Why in all of hades would you force your entire eng org to only use one LLM provider. It's incredibly easy to run this stuff locally on 4+ year old hardware. Why is this even something you're spending company money on? Investor funds?
mosdl
It's weirdly common at young startups. Interviewed at two places in the past few weeks where cursor was required. Funnily enough the reason for hiring was because they needed people to fix things ...
birdman3131
Tinfoil hat me says that it was a policy change that they are blaming on an "AI Support Agent" and hoping nobody pokes too much behind the curtain.
Note that I have absolutely no knowledge or reason to believe this other than general distrust of companies.
rustc
> Tinfoil hat me says that it was a policy change that they are blaming on an "AI Support Agent" and hoping nobody pokes too much behind the curtain.
Yeah, who puts an AI in charge of support emails with no human checks and no mention that it's an AI generated reply in the response email?
daemonologist
AI companies high on their own supply, that's who. Ultralytics is (in)famous for it.
itissid
Why is Ultralytics yolo famous for it?
recursive
A forward-thinking company that believes in the power of Innovation™.
sitkack
These bros are getting high on their own supply. I vibe, I code, but I don't do VibeOps. We aren't ready.
VibeSupport bots, how well did that work out for Canada Air?
https://thehill.com/business/4476307-air-canada-must-pay-ref...
behnamoh
"It's evolving, but backwards."
p1necone
An AI company dogfooding their own marketing. It's almost admirable in a way.
rangerelf
I worry that they don't understand the limitations of their own product.
nkrisc
This is the future AI companies are selling. I believe they would 100%.
xbar
I worry that the tally of those who do is much higher than is prudent.
conradfr
A lot of company actually, although 100% automation is still rare.
that_guy_iain
100% for first line support is very common. It was common years ago before ChatGPT and ChatGPT made it so much better than before.
pxx
OpenAI seems to do this. I've gotten complete nonsense replies from their support for billing questions.
nimchimpsky
[dead]
babypuncher
Given how incredibly stingy tech companies are about spending any money on support, I would not be surprised if the story about it being a rogue AI support agent is 100% true.
It also seems like a weird thing to lie about, since it's just another very public example of AI fucking up something royally, coming from a company whose whole business model is selling AI.
xienze
Both things can be true. The AI support bot might have been trained to respond with “yup that’s the new policy”, but the unexpected shitstorm that erupted might have caused the company to backpedal by saying “official policy? Ha ha, no of course not, that was, uh, a misbehaving bot!”
arkh
> how incredibly stingy tech companies are about spending any money on support
Which is crazy. Support is part of marketing so it should get the same kind of consideration.
Why do people think Amazon is hard to beat? Price? nope. Product range? nope. Delivery time? In part. The fact if you have a problem with your product they'll handle it? Yes. After getting burned multiple times by other retailers you're gonna pay the Amazon tax so you don't have to ask 10 times for a refund or be redirected to the supplier own support or some third party repair shop.
Everyone knows it. But people are still stuck on the "support is a cost center" way of life so they keep on getting beat by the big bad Amazon.
miyuru
In my products, if a user has payed me, their support tickets get high priority, and I get notified immediately.
Other tickets get replied within the day.
I am also running it by myself; I wonder why big companies with 50+ employees like cursor cheaps out with support.
sitkack
That is because AI runs PR as well.
throwaway314155
Yeah it makes little sense to me that so many users would experience exactly the same "hallucination" from the same model. Unless it had been made deterministic but even then subtle changes in the wording would trigger different hallucinations, not an identical one.
WesolyKubeczek
What if the prompt to the “support assistant” postulates that 1) everything is a user error, 2) if it’s not, it’s a policy violation, 3) if it’s not, it may be our fuckup but we are allowed? This plus the question in the email leading to a particular answer.
Given that LLMs are trained on lots of stuff and not just the policy of this company, it’s not hard to imagine how it could conjure that the policy (plausibly) is “one session per user”, and blame them of violating it.
6510
This is the best idea I read all day. Going to implement AI for everything right now. This is a must have feature.
isaacremuant
I think this would actually make them look worse, not better.
joe_the_user
Weirdly, your conspiracy theory actually makes the turn of events less disconcerting.
The thing is, what the AI hallucinated (if it was an AI-hallucinating), was the kind of sleezy thing companies do do. However, the thing with sleezy license changes is they only make money if the company publicizes them. Of course, that doesn't mean a company actually thinks that far ahead (X many managers really think "attack users ... profit!"). Riddles in enigmas...
jgb1984
LLM anything makes me queasy. Why would any self respecting software developer use this tripe? Learn how to write good software. Become an expert in the trade. AI anything will only dig a hole for software to die in. Cheapens the product, butchers the process and absolutely decimates any hope for skill development for future junior developers.
I'll just keep chugging along, with debian, python and vim, as I always have. No LLM, no LSP, heck not even autocompletion. But damn proud of every hand crafted, easy to maintain and fully understood line of code I'll write.
cachvico
I use it all the time, and it has accelerated my output massively.
Now, I don't trust the output - I review everything, and it often goes wrong. You have to know how to use it. But I would never go back. Often it comes up with more elegant solutions than I would have. And when you're working with a new platform, or some unfamiliar library that it already knows, it's an absolute godsend.
I'm also damn proud of my own hand-crafted code, but to avoid LLMs out of principal? That's just luddite.
20+ years of experience across game dev, mobile and web apps, in case you feel it relevant.
ericwood
I have a hard time being sold on “yea it’s wrong a lot, also you have to spend more time than you already do on code review.”
Getting to sit down and write the code is the most enjoyable part of the job, why would I deprive myself of that? By the time the problem has been defined well enough to explain it to an LLM sitting down and writing the code is typically very simple.
tptacek
You're giving the game away when you talk about the joy LLMs are robbing from you. I think we all intuit why people don't like the idea of big parts of their jobs being automated away! But that's not an argument on the merits. Our entire field is premised on automating people's jobs away, so it's always a little rich to hear programmers kvetching about it being done to them.
pizza
The parts worth thinking about you still think about. The parts that you’ve done a million times before you delegate so you can spend better and greater effort on the parts worth thinking about.
woah
I'm confused when people say that LLMs take away the fun or creativity of programming. LLMs are only really good at the tedious parts.
null
dgs_sgd
For me it's typically wrong not in a fundamental way but a trivial way like bad import paths or function calls, like if I forgot to give it relevant context.
And yet the time it takes me to use the LLM and correct its output is usually faster than not using it at all.
Over time I've developed a good sense for what tasks it succeeds at (or is only trivially wrong) and what tasks it's just not up for.
YeGoblynQueenne
>> I use it all the time, and it has accelerated my output massively.
Like how McDonalds makes a lot of burgers fast and they are very successful so that's all we really care about?
cachvico
Terrible analogy. I don't commit jank. If the LLM comes out with nonsense, I'll fix it first.
timewizard
> "and it has accelerated my output massively."
The folly of single ended metrics.
> but to avoid LLMs out of principal? That's just luddite.
Do you double check that the LLM hasn't magically recreated someone else's copyrighted code? That's just irresponsible in certain contexts.
> in case you feel it relevant.
Of course it's relevant. If a 19 year old with 1 year of driving experience tries to sell me a car using their personal anecdote as a metric I'd be suspicious. If their only salient point is that "it gets me to where I'm going faster!" I'd be doubly suspicious.
xvector
> Do you double check that the LLM hasn't magically recreated someone else's copyrighted code?
I frankly do not care, and I expect LLMs to become such ubiquitous table-stakes that I don't think anyone will really care in the long run.
cachvico
Add "Without compromising quality then"!
callc
I’m pretty much in the same boat as you, but here’s one place that LLMs helped me:
In python I was scanning 1000’s of files each for thousands of keywords. A naive implementation took around 10 seconds, obviously the largest share of execution time after running instrumentation. A quick ChatGPT led me to Aho-Corasick and String searching algorithms, which I had never used before. Plug in a library and bam, 30x speed up for that part of the code.
I could have asked my knowledgeable friends and coworkers, but not at 11PM on a Saturday.
I could have searched the web and probably found it out.
But the LLM basically auto completed the web, which I appreciate.
kovac
This is where education comes in. When we come cross a certain scale, we should know that O(n) comes into play, and study existing literature before trying to naively solve the problem. What would happen if the "AI" and web search didn't return anything? Would you have stuck with your implementation? What if you couldn't find a library with a usable license?
Once I had to look up a research paper to implement a computational geometry algorithm because I couldn't find it any of the typical Web sources. There were also no library to use with a license for our commercial use.
I'm not against use of "AI". But this increasing refusal of those who aspire to work in specialist domains like software development to systematically learn things is not great. That's just compounding on an already diminished capacity to process information skillfully.
callc
In my context, the scale is small. It just passed the threshold where a naive implementation would be just fine.
> What would happen if the "AI" and web search didn't return anything? Would you have stuck with your implementation?
I was fairly certain there must exist some type of algorithm exactly for this purpose. I would have been flabbergasted if I couldn’t find something on the web. But it that failed, I would have asked friends and cracked open the algorithms textbooks.
> I'm not against use of "AI". But this increasing refusal of those who aspire to work in specialist domains like software development to systematically learn things is not great. That's just compounding on an already diminished capacity to process information skillfully.
I understand what you mean, and agree with you. I can also assure you that that is not how I use it.
infoseek12
There is a time and a place for everything. Software development is often about compromise and often it isn’t feasible to work out a solution from foundational principles and a comprehensive understanding of the domain.
Many developers use libraries effectively without knowing every time consideration of O(n) comes into play.
Competently implemented, in the right context, LLMs can be an effective form of abstraction.
klabb3
Yes! This is how AI should be used. You have a question that’s quite difficult and may not score well on traditional keyword matching. An LLM can use pattern matching to point you in the right direction of well written library based on CS research and/or best practices.
valenterry
I mean, even in the absence of knowledge of the existence of text searching algorithms (where I'm from we learn that in university) just a simple web search would have gotten you there as well no? Maybe would have taken a few minutes longer though.
callc
Extremely likely, yes. In this case, since it was an unknown unknown at the time, the LLM nicely explaining that this class of algorithms exists was nice, then I could immediately switch to Wikipedia to learn more (and be sure of the underlying information)
I think of LLMs as an autocomplete of the web plus hallucinations. Sometimes it’s faster to use the LLM initially rather than scour through a bunch of sites first.
mrheosuper
But do you know every important detail of that library. For example, maybe that lib is not thread safe, or it allocates a lot of memory to speed thing up, or it wont work on ARM CPU because it uses some x86 hackery ASM?
callc
Nope. And I don’t need to. That is the beauty of abstractions and information hiding.
Just read the docs and assume the library works as promised.
To clarify, the LLM did not tell me about the specific library I used. I found it the old fashioned way.
mixmastamyk
Sounds like a job for silver/ripgrep and possibly stack exchange. Might take another minute to get it rolling but has other benefits like cost and privacy.
aleph_minus_one
> I could have asked my knowledgeable friends and coworkers, but not at 11PM on a Saturday.
Get friends with weirder daily schedules. :-)
financypants
I think it's best if we all keep the hours from ~10pm to the morning sacred. Even if we are all up coding, the _reason_ I'm up coding at that hour is because no one is pinging me
marcus_holmes
I was with you 150% (though Arch, Golang and Zed) until a friend convinced me to give it a proper go and explained more about how to talk to the LLM.
I've had a long-term code project that I've really struggled with, for various reasons. Instead of using my normal approach, which would be to lay out what I think the code should do, and how it should work, I just explained the problem and let the LLM worry about the code.
It got really far. I'm still impressed. Claude worked great, but ran out of free tokens or whatever, and refused to continue (fine, it was the freebie version and you get what you pay for). I picked it up again in Cursor and it got further. One of my conditions for this experiment was to never look at the code, just the output, and only talk to the LLM about what I wanted, not about how I wanted it done. This seemed to work better.
I'm hitting different problems, now, for sure. Getting it to test everything was tricky, and I'm still not convinced it's not just fixing the test instead of the code every time there's a test failure. Peeking at the code, there are several remnants of previous architectural models littering the codebase. Whole directories of unused, uncalled, code that got left behind. I would not ship this as it is.
But... it works, kinda. It's fast, I got a working demo of something 80% near what I wanted in 1/10 of the time it would have taken me to make that manually. And just focusing on the result meant that I didn't go down all the rabbit holes of how to structure the code or which paradigm to use.
I'm hooked now. I want to get better at using this tool, and see the failures as my failures in prompting rather than the LLM's failure to do what I want.
I still don't know how much work would be involved in turning the code into something I could actually ship. Maybe there's a second phase which looks more like conventional development cleaning it all up. I don't know yet. I'll keep experimenting :)
imhoguy
> never look at the code, just the output, and only talk to the LLM about what I wanted
Sir, you have just passed vibe coding exam. Certified Vibe Coder printout is in the making but AI has difficulty finding a printer. /s
FeepingCreature
Computers don't need AI help to have trouble finding the printer, lol.
SkyPuncher
> Why would any self respecting software developer use this tripe?
Because I can ship 2x to 5x more code with nearly the same quality.
My employer isn't paying me to be a craftsman. They're paying me to ship things that make them money.
ivan_gammel
How do you define code quality in this case and what is your stack?
SkyPuncher
The definition of code quality is irrelevant to my argument as both human and AI written code are held to the same standard by the same measure (however arbitrary that measure is). 100 units of something vs 99 units of something is a 1 unit difference regardless of what the unit is.
By the time the AI is actually writing code, I've already had it do a robust architecture evaluation and review which it documents in a development plan. I review that development plan just like I'd review another engineers dev plan. It's pretty hard for it to write objectively bad code after that step.
Also, my day to day work is in an existing code base. Nearly every feature I build has existing patterns or reference code. LLMs do extremely well when you tell them "Build X feature. [some class] provides a similar implementation. Review that before starting." If I think something needs to be DRY'd up or refactored, I ask it to do that.
kamaal
Code that you can understand and fix later, is acceptable quality per my definition.
Either way, LLMs are actually high up the quality spectrum as they generate a very consistent style of code for everyone. Which gives it uniformity, that is good when other developers have to read and troubleshoot code.
leoh
Good employee, you get cookie and 1h extra pto
NineWillows
No, I get to spend 2 hours working with LLMs, and then spend the rest of the day doing whatever I please. Repeat.
bigstrat2003
I wholeheartedly agree. When the tools become actually worth using, I'll use them. Right now they suck, and they slow you down rather than speed you up. I'm hardly a world class developer and I can do far better than these things. Someone who is actually top notch will outclass them even more.
Chinjut
I understand not wanting to use LLMs that with no correctness guarantees that randomly hallucinate, but what's wrong with ordinary LSPs and autocompletion? Those seem like perfectly useful tools.
OsrsNeedsf2P
I had a professor who used `ed` to write his code. He said only bring able to see one line at a time forces you to think more about what you're doing.
Anyways, Cursor generates all my code now.
x1xx
If you are like me (same vim, python, no LLM, no autocompletion, no syntax highlighting noise), LSP will make you a better developer: it makes navigating the codebase MUCH easier, including stdlib and 3rd party dependencies.
As a result, you don't lose flow and end up reading considerably more code than you would have otherwise.
jgb1984
Actually, I'm kind of cheating because I use https://github.com/davidhalter/jedi-vim for that purpose: allows me to jump to definitions with <leader>d ;) Excellent plugin, and doesn't require an LSP.
incoming1211
Can pretty much guarantee with AI I'm a better software developer than you without. And I still love working on software used by millions of people every day, and take pride in what I do.
kebokyo
here's an archive of the original reddit post since it seemed to be instantly nuked: https://undelete.pullpush.io/r/cursor/comments/1jyy5am/psa_c...
rurp
It's funny seeing all of the comments trying to blame the users for this screwup by claiming they're using it wrong. It is reddit though, so I guess I shouldn't be surprised.
keeganpoppen
what is it about reddit that causes this behavior, when they otherwise are skeptical only of whatever the "official story" is at all costs? it is fascinating behavior.
klabb3
One reason is lazy one liners are allowed and have high cost/benefit for shitposters and attract many upvotes, so this gets voted highly setting the tone for flamewars in the thread.
It’s miles better on HN. Most bad responses are penalized. The culture is upvoting things that are contributing. I frequently upvote responses that disagree with me. Oftentimes I learn something from it.
AstroBen
think about the demographic that would join a cursor subreddit. Basically 90% superfans. Go against the majority opinion and you'll be nuked
dpkirchner
I think it's a combination of oppositional defiant disorder and insufficient moderation.
bytesandbits
wow they nuked it for damage control and only caused more damage
ddxv
Cursor is weird. They have a basically unused GitHub with a thousand unanswered Issues. It's so buggy in ways that VSCode isn't. I hate it. Also I use it everyday and pay for it.
That's when you know you've captured something, when people hate use your product.
Any real alternatives? I've tried continue and was unimpressed with the tab completion and typing experience (felt like laggy typing on a remote server).
adriand
VS Code with standard copilot for tab completion and Aider in a terminal window for all the heavier lifts, asking questions, architecting etc. And it’s cheap! I’ve been using it with OpenRouter (lets you easily switch models and providers) and my $10 of credits lasted weeks. Granted, I also use Claude a lot in the browser.
dtquad
The reason many prefer Cursor over VSCode + GitHub Copilot is because of how much faster Cursor is for tab completion. They use some smaller models that are latency optimized specifically to make the tab completion feel as fast as possible.
caelinsutch
If you don't mind leaving VSCode I'm a huge fan of Zed. Doesn't support some languages / stacks yet but their AI features are on-par with VSCode
presentation
That's the wrong IDE to compare it to though, Cursor's AI features are 10x better than VSCode's. I tried Zed last month and while the editing was great, the AI features were too half-baked so I ended up going back to Cursor. Hopefully it gets better fast!
dkersten
Agreed. My laptop has never used swap until I started using cursor… it’s a resource hog, I dislike using it, but it’s still the best AI coding aid and for the work I’m doing right now, the speed boost is more valuable than hand crafted code in enough cases that it’s worth it for me. But I don’t enjoy using the IDE itself, and I used vscode for a few years.
Personally, I will jump ship to Zed as soon as it’s agent mode is good enough (I used Zed as a dumb editor for about a year before I used cursor, and I love it)
permo-w
I find that if you turn off telemetry (i.e. turn on privacy) the resource hogging slows down a lot
dkersten
Hmm. I double checked and I have privacy mode enabled, so I don't think that's the root cause. I also removed all but the bare essential extensions (only the theme I'm using and the core language support extensions for typescript and python).
d357r0y3r
Cline is pretty solid and doesn't require you to use a completely unsustainable VSCode fork.
alok-g
I have heard Roo Code is a fork of Cline that is better. I have never used either so far.
SkyPuncher
I prefer Roo, but they're largely the same right now. They each have some features the other doesn't.
smaddox
I switched to Windsurf.ai when cursor broke for me. Seems about the same but less buggy. Haven't used it in the last couple weeks, though, so YMMV.
omneity
I found the Windsurf agent to be relatively less capable, but their inline tool (and the “Tab” they’re promoting so much) has been extremely underwhelming, compared to Cursor.
The only one in this class to be even worse in my experience is Github Copilot.
null
htrp
cant bother fixing their issues because they are too busy vibe coding new features
behnamoh
Cursor + Vim plugin never worked for me, so I switched back to Nvim and never looked back. Nvim already has: avante, codeCompanion, copilot, and many other tools + MCP + aider if you're into that.
bytesandbits
Cursor sucks. Not as a product. As a team. Their customer support is terrible.
I was offered in writing a refund by the team who cold reached out to me to ask me why I cancelled my sub one week after start. Then they ignored my 3+ emails in response asking them to refund, and other means of trying to communicate with them. Offering me a refund as a bait to gain me back, then when I accept it they ghost me. Wow. Very low.
The product is not terrible but the team responses are. And this, if you see how they handled it, is also a very poor response. First thing you notice if you open the link is that the Cursor team removed the reddit post! As if we were not going to see it or something? Who do they think they are? Censoring bad comments which are 100% legit.
I am giving it a go to competitors just out of sheer frustration with how they handle customers, and I do recommend everybody to explore other products before you settle on Cursor. I don't intend to ever re-subscribe and have recommended friends to do the same, most of which agree with my experience.
JohnKemeny
> Their customer support is terrible.
You just don't know how to prompt it correctly.
Crosseye_Jack
sounds like perfect grounds for a chargeback to me. Company offered a full refund via one of its Agents, company then refused to honour that offer, time to make your bank force them to refund you.
Just because you use AI for customer service doesn't mean you don't have to honour its offers to customers. Air Canada recently lost a case where its AI offered a discount to a customer but then refused to offer it "IRL"
https://www.forbes.com/sites/marisagarcia/2024/02/19/what-ai...
einsteinx2
Same exact thing happened to me. I tried out Vursor after hearing all the hype and canceled after a few weeks. Got an email asking if I wanted a refund and asking for any feedback. I replied with detailed feedback on why I canceled and accepted the refund offer, then never heard back from them.
samanator
Interesting. The same thing happened to me. Was offered a refund (graciously, as I had forgotten to cancel the subscription). And after thanking them and agreeing to the refund, was promptly ignored!
Very strange behavior honestly.
null
pzo
I had the same exact experience - after disappointment (couldn't use like 2/3 of my premium credits because every second request failed after they upgraded to 0.46) unsubscribed. They offered refund in email. I replied I wanted refund but no reply
gblargg
Apparently they use AI to read emails. So the future of email will be like phone support now, where you keep writing LIVE AGENT until you get a human responding.
PaulStatezny
This reminds me of how small of a team they are, and makes me wonder if they have a customer support team that's growing commensurately with the size of the user base.
andybak
I just cancelled - not because I thought the policy change was real - but simply because this article reminded me I hadn't used it much this month.
scarface_74
This is where Kagi’s subscription policy comes in handy. If you don’t use it for a month, you don’t pay for it that month. There is no need to cancel it and Kagi doesn’t have to pay user acquisition costs.
paxys
Slack does this as well. It's a genius idea from a business perspective. Normally IT admins have to go around asking users if they need the service (or more likely you have to request a license for yourself), regularly monitor usage, deactivate stale users etc., all to make sure the company isn't wasting money. Slack comes along and says - don't worry, just onboard every user at the company. If they don't log in and send at least N messages we won't bill them for that month.
ashu1461
They mention an user taking an action will be billed. I guess even sending a message or reacting with an emoji would count as taking an action ? Even logging in ?
tshaddox
That's a fun one. It could be interpreted as a generous implementation of a monthly subscription, or a hostile implementation of a metered plan.
elcritch
Wow, I wish more services did that.
permo-w
Kagi should take it a step further and just charge per search
scarface_74
History shows that metered plans are extremely unpopular whether it be cell phone service, video, music etc.
jay_kyburz
surely the first thing you do when you subscribe to Kagi is set your default browser search to Kagi.
mirekrusin
Really? Brilliant idea.
theturtletalks
Cursor is trapped in a cat and mouse game against "hacks" where users create new accounts and get unlimited use. The repo was even trending on Github (https://github.com/yeongpin/cursor-free-vip).
Sadly, Cursor will always be hampered by maintaining it's own VSCode fork. Others in this niche are expanding rapidly and I, myself, have started transitioning to using Roo and Cline.
permo-w
literally any service with a free trial--i.e. literally any service--has this "problem". it's an integral part of the equation in setting up free trials in the first place, and by no means a "trap". you're always going to have a % of users who do this, the business model relies on the users who forget and let the subscription cross over to the next month or simply feel its worth paying
theturtletalks
This is true but Cursor’s problems are a bit worse than a normal paywalled service.
Cursor allows users to get free credits without a credit card and this forced them to change their VSCode fork on how it handles identification so they can stop users from spawning new accounts.
Another is that normally, companies have a cost for each free user. For Cursor, this cost is so sporadic since it doesn’t charge per million context, they use credits. Free users get 50 credits but 1 credit could be 200k+ context each so it could be $40-50 per free user per month. And these users get 50 credits every month.
Lastly, the cursor vip free repo has trended on GitHub many times and users who do pay might stop and use this repo instead.
The Cursor vip free creator is well within his rights to do what they want and get “free” access. This unfortunately hurts paying customers since Cursor has to stop these “hacks.”
This is why Cursor should just move to a VSCode extension. I’ve used Augment and other VSCode extensions and the feature set is close to Cursor so it’s possible for them just to be an extension. The other would be to remove free accounts but allow users to bring their own keys. To use Composer/Agent, you can’t bring your own keys.
This will allow Cursor to stop maintaining a VSCode fork, helps them stop caring if users create new accounts (since all users are paying) and lets users bring their own keys if they don’t want to pay. Hell, if they charge a lifetime fee to bring our own keys for Agent, that would bring in revenue too. But as I see now, Roo and Cline’s agent features are catching up and Cursor won’t have a moat soon.
sumedh
> but 1 credit could be 200k+ context each
There is a thread on Cursor forums where the context is around 20K to 30K tokens.
GabrielHawk
> Cursor is trapped in a cat and mouse game against "hacks" where users create new accounts and get unlimited use
Actually, you don't even have to make a new account. You can delete your account and make it again reusing the same email.
I did this on accident once because I left the service and decided to come back, and was surprised to get a free tier again. I sent them an email letting them know that was a bug, but they never responded.
I paid for a month of access just to be cautious, even though I wasn't using it much. I don't understand why they don't fix this.
saintfire
The AI support triage agent must have deemed it unworthy.
john2x
> I don't understand why they don't fix this.
It makes number go up and to the right
_fat_santa
What I don't understand is why maintain a fork instead of just an extension? I would assume an extension like this would be pretty difficult to create but would it be more difficult than literally forking VSCode?
theturtletalks
Cursor was one of the first AI editors and there was no way to add those features to VSCode thru extensions at that time. Another issue was that VSCode made special exceptions and opened APIs for CoPilot, Microsoft’s product, but not for other players. Not sure if this has changed now.
Cursor took the best course of action at the time by forking but needs to come back into the fold. If VSCode is restricting access to APIs to CoPilot, forking it publicly and putting that in the Readme, “We forked VSCode since they give preferential treatment to CoPilot” would get a lot of community support.
sitkack
There will be a fork-a-month for these products until they have the same lockin as a textbox that you talk at, "make million dollar viral facebook marketplace post"
Aeolun
They can just drop any free usage right?
cma
Tradeoff with slowing down user acquisition
SkyPuncher
Yes and no.
In a corporate environment, compliance needs are far more important than some trivial cost.
hsbauauvhabzb
Embrace, extend, extinguish.
null
mgraczyk
What is the evidence that "dozens of users publicly canceled their subscriptions"?
A total of 4 users claimed that they did or would cancel their subscriptions in the comments, and 3/4 of them hedged by saying that they would cancel if this problem were real or happened to them. It looks like only 1 person claimed to have cancelled already.
Is there some other discussion you're looking at?
dang
Submitted title was "Cursor IDE support hallucinates lockout policy causes mass user cancellations" - I've de-massed it now.
Since the HN title rule is "Please use the original title, unless it is misleading or linkbait" and the OP title is arguably misleading, I kept the submitter's title. But if there's a more accurate or neutral way to say what happened, we can change it again.
hombre_fatal
Yeah, it's a reddit thread with 59 comments.
Yet if you went by the HN comments, you'd think it were the biggest item on primetime news.
People are really champing at the bit.
instagib
Currently, there are 59 comments after the post was deleted and the thread was locked.
It is worth mentioning that the comments that remain are valuable, as they highlight the captured market size and express concern about the impending deterioration of the situation.
AndyKelley
From the top Reddit post:
> Apologies about the confusion here.
If this was a sincere apology, they'd stop trying to make a chat bot do support.
zb3
This drama is a very good thing because 1) now companies might reconsider replacing customer support with AI and 2) the initial victim was an AI company.
It could be better though.. I wish this happened to a company providing "AI support solutions"..
dylan604
A company of 10 devs might not actually have a customer support at all. We all know, devs are not the greatest customer support people.
tshaddox
It's an AI company that is presumably drowning in money. What humans do work there have probably already had a good laugh and forgotten about this incident.
Earlier today Cursor, the magical AI-powered IDE started kicking users off when they logged in from multiple machines.
Like,you’d be working on your desktop, switch to your laptop, and all of a sudden you're forcibly logged out. No warning, no notification, just gone.
Naturally, people thought this was a new policy.
So they asked support.
And here’s where it gets batshit: Cursor has a support email, so users emailed them to find out. The support peson told everyone this was “expected behavior” under their new login policy.
One problem. There was no support team, it was an AI designed to 'mimic human responses'
That answer, totally made up by the bot, spread like wildfire.
Users assumed it was real (because why wouldn’t they? It's their own support system lol), and within hours the community was in revolt. Dozens of users publicly canceled their subscriptions, myself included. Multi-device workflows are table stakes for devs, and if you're going to pull something that disruptive, you'd at least expect a changelog entry or smth.
Nope.
And just as people started comparing notes and figuring out that the story didn’t quite add up… the main Reddit thread got locked. Then deleted. Like, no public resolution, no real response, just silence.
To be clear: this wasn’t an actual policy change, just a backend session bug, and a hallucinated excuse from a support bot that somehow did more damage than the bug itself.
But at that point, it didn’t matter. People were already gone.
Honestly one of the most surreal product screwups I’ve seen in a while. Not because they made a mistake, but because the AI support system invented a lie, and nobody caught it until the userbase imploded.