Skip to content(if available)orjump to list(if available)

Ask HN: How much better are AI IDEs vs. copy pasting into chat apps?

Ask HN: How much better are AI IDEs vs. copy pasting into chat apps?

137 comments

·May 8, 2025

I just wanted to hear peoples experiences with AI IDEs.

For context, I'm a heavy user of Gemini / ChatGPT for coding and Copilot. But I haven't used Cursor / Windsurf / etc..

Copy pasting into chat apps is a first world problem: it will do the work for you, but you have to give it all the context in the prompt, which for a larger project, gets tedious.

The issue with Copilot is that it's not as smart as the "thinking" chat apps.

This makes it clear why there's such a need for AI IDEs. I don't want to construct my context to a chat app. The context is already in my codebase, so the AI should pick up on it. But I also hear that it gets expensive because of the pay-per-use pricing, as opposed to effectively unlimited prompts for a thinking chat app if you pay the monthly subscription.

So I just wanted to get the lay of the land. How good are these IDEs on constructing your context to the LLMs? How much more expensive is it, and is it worth it for you?

SeanAnderson

I'm a staff software engineer doing a mix of front-end and back-end with emphasis on front-end.

I use both Cursor on Claude 3.7 and ChatGPT on 4o/o3. Cursor seems kind of "dumb" compared to 4o, but it's a good workhorse.

I let Cursor handle the basics - basically acting as a glorified multi-file autocomplete. I think through common problems with 4o, tough problems with o3, I copy all of Svelte's docs into 4o (https://svelte-llm.khromov.se/) to get good Svelte 5-focused feedback, I have 4o code-review what Cursor writes from time to time, I have 4o, sometimes o3, generate "precise" prompts that I'll give to Cursor when me talking off-the-cuff to Cursor doesn't get good results after a few attempts.

I don't consider myself an expert in these areas yet so I might be misusing Cursor, or not making enough use of its rules system, or something. I feel like I get good value for my ChatGPT subscription. I don't feel like I get good value for my Cursor subscription, but I also still feel like keeping it because $20 to type a lot less is still pretty nice. I would be upset if I only had a Cursor subscription and no access to ChatGPT. I am pretty hesitant to pay for AI à la carte. I feel much better about working within the limitations of a known subscription cost.

aeve890

>I let Cursor handle the basics - basically acting as a glorified multi-file autocomplete.

Same experience. I'm in no need to let the AI write large, real working code. But it's amazing for fast refactoring, function signatures, boilerplate, etc. The real boring parts. I only wish it wouldn't be so eager to try to dump a 50 lines of bullshit code every time. But for small changes, super cool.

babyent

I honestly believe they do this BS shit so you waste tokens and pay more.

Engagement, after all.

m13rar

same experience. Editor agents with basic level tasks and boilerplate. Problem solving and decomposition of it in the chat apps seems to be better. What I've noticed is the agent modes of these editors use the embeddings API of ChatGPT and how the LLM model maps the context of the codebase. Often what has happened is the LLM's in Agent mode ignore the default setup of the codebase example package managers and use the package managers which would be popular throughout their training data.

To summarise the Agent Mode editors don't try to fill in their context gaps and be aware of the environment they operate it unless explicitly specified by the prompter to first review the codebase, understand it's structure and then proceed with implementing features.

starstripe

Can you please elaborate on how you're copying Svelte's docs into 4o?

SeanAnderson

I just download the Svelte + SvelteKit (Recommended - LLM Distilled) file from the link mentioned in my original comment and drag-and-drop upload it into the chat context when I'm talking to 4o about coding. Nothing magical. It takes a second to read through the document, but then the LLM works significantly better. It stops suggesting invalid Svelte 4 syntax and prioritizes using Runes.

clarity_89

> Cursor seems kind of "dumb" compared to 4o

What do you mean by that? You can use 4o model in Cursor?

throwaw12

I think part of the reason is, ChatGPT has additional system prompt in chat, API doesn't have.

even if both companies use same model (OpenAI and Cursor), system prompt has a huge impact on results

caleblloyd

Cursor tries to save more money on context than ChatGPT. So when both are targeting GPT 4o, the ChatGPT App will incorporate more context history and memories than Cursor.

hoherd

Cursor lists a large number of models out of the box, and you can select which one you want it to use. https://docs.cursor.com/settings/models

Ey7NFZ3P0nzAe

I'm SO surprised to not see aider-chat everywhere mentionned.

I've been using it for about a year and it's incredible.

My take is that most people did not invest the little time necessary to get acustomed to its workflow.

My advices for aider are:

- familiarize yourself with the chat modes (architect, code, ask)

- familiarize yourself with the edit modes (diff, whole, etc) and know which to use for a given model. Indeed not all models handle all modes equally well.

- make the code one feature at a time, by small chunks if needed, by limiting the contxt to the relevant files.

- practice to learn how to best phrase stuff.

- write you coding preferences into aider convention files. Things like "always use type hints, beartype for type checking, python click for the cli, add many comments".

I'm mainly doing python and with proper comments and type hints it's really easy for models to write code that works with the rest of the repo even with only a few files in its context.

extralarge

Do you find that the time you spend splitting a feature up into small chunks and explaining each small chunk to the LLM takes longer than just writing the code yourself?

I've used aider but found just pounding out some code in neovim to be faster. Perhaps I'm not using aider correctly.

Ey7NFZ3P0nzAe

I do think that it can take maybe a little more time with aider but this time investment pays itself off as soon as there is some complexification or issue that LLMs can't fix.

See, if you have no idea of what kind of architecture was used for the passed 150 commits, it will take more time to debug or see if a given new featufe you have in mind is doable without major refactoring for example.

This will probably matter less as LLMs grow more powerful I admit

m13rar

Agree, I think aider chat forces the prompter to develop good habits in the prompting process, like providing the proper context, problem decomposition and utilizing the LLM capabilities properly. The Aider interface and additional setup required in the process makes it more productive over time.

Imanari

Same here, I'd even say you made the learning curve sound worse than it is, no offence :)

This is my workflow for 90% of tasks:

1) /ask: I want to do X. / How would you do X? / What options do I have for X?

2) (optional) /ask>: make these adaptations / consider this / follow up question.

3) "go ahead" (this defaults to /code mode)

For a model I like the style of Sonnet3.7 best.

One caveat: I add my files manually. Knowing what you want and where you want it is 90% of the game and this is where LLMs usually get it wrong.

SamDc73

Aider is to other AI tools what Vim/Neovim is to other IDEs. Super powerful, but with a learning curve most folks don’t want to climb.

trees101

https://github.com/hotovo/aider-desk is a gui, takes 5 mins to install, has MCP support (try context7). Definitely worth a look and is an "easy" way in to aider.

sepositus

This is exactly how I use agent mode in Cursor. I’ve never understood why I would use aider. Am I missing something else?

Ey7NFZ3P0nzAe

Possibly: I'm a neovim person and barely leave the terminal. So naturally aider fits perfectly into my setup. There are even neovim plugin for having aider directly in it.

That said, I don't get why people are fine switching their IDE to cursor. Lile, if you're in an IDE it's because it has features you want that others didn't. So either cursor does not have them or if it does then you have to configure the whole cursor IDE to work like your original IDE.

In total you've either lost features or been captured into a new walled garden. And lost some time in the process.

Aider's use of litellm makes it 5s change to switch to the latest models. And it's always up to date thanks to uvx.

No really I don't get why people are not using aider more.

Plus the dev is lovely, responsive and helpful.

0x6c6f6c

Cursor is a VS Code fork. Anyone coming from VS Code, one of the most popular editors in the ecosystem, will immediately be comfortable. The full marketplace of extensions is compatible. The switchover is effectively seamless besides a handful of keyboard shortcuts used for the new AI features.

fhd2

I settled on gptel, which is an LLM package for Emacs. It's kind of a mix of both approaches:

1. You have chats right there in the editor, easy to copy/paste and manage context/history without switching to a browser. You can also quickly add files or portions of files to the context or remove them again.

2. You can choose which model you want to use for what, granted you have an API key.

3. You can quickly highlight some code and ask for a change to it, which along with managed context is powerful.

I tried auto complete again and again but it doesn't work for me. At first I think "yeah, that's what I wanted to write", and then I have to look closer to realise it's not, and that completely breaks my flow and concentration. I can always write some pseudo code and proactively convert it to real code, I like to be in the driver seat.

Context management is really central to my workflow, I manage it like a hawk. Models tend to deteriorate as context content increases, in my experience, so I really try to keep it narrow.

For that reason, and because our clients didn't sign up for their code to be sent to Anthropic et al, I _mostly_ use models like I would use StackOverflow, not to generate non-trivial code I'd actually use.

But having the chats in my editor is really invaluable for me. Powerful text wrangling features make a difference in both speed and motivation.

I use it pretty heavily with pretty much only the high-end models and pay about $15 per month.

binary132

FWIW I have been using gptel with local inference by a deepcoder 14b q6 and llama.cpp and it’s not too bad. But it can’t handle big asks. 24GB of VRAM is not enough to make local inference smart enough to be really useful. It’s okayish at things like writing commit messages or reformatting / refactoring local regions of code. 32B quants are slow and not that much better.

faustlast

have you tried aider? Using it with magit to see the changes is a good workflow. the aider integrations to emacs are also worth it, you can easily add/remove files and send highlighted parts of your buffer.

fhd2

Not yet, but I'm intrigued! Good to know there's good Emacs integration.

gollum999

Do you use aidermacs or something else?

olalonde

I use Claude Code (cli tool) and it's on another level. Not even comparable to code autocomplete à la Copilot or copy/pasting into a LLM chat app. It knows about your whole code base, can use external tools, read documentation, run tests, etc. Even as an experienced developer, it's been a huge productivity boost. The main downside is that it can quickly get expensive.

rcarmo

You get the same experience inside VS Code with Copilot these days, as long as you select agent mode. Yesterday I even asked it to read documentation off a URL and it asked me permission to fetch the page, summarized it, and applied the API changes needed.

pmelendez

Not really. This is an example of something that Claude Code handles very easily and for Github Copilot, is not even an starter (because can't read SQLite files)

>Read the Sqlite file in this directory and create an UX that would visualize and manage the data. The application should be responsive (so it can be use on mobiles) and should have some graphs when appropiate.

rcarmo

Actually, it can. I've just gone and checked. Agent Mode executed sqlite on the terminal, got the schema, and issued the relevant changes.

alserio

Does it really know about the whole codebase? How big are we talking about? Last I tried, some time ago, llm didn't work very well "in the large". But I have never tried Claude Code.

jamil7

No it doesn’t, it tries to be smart about how it loads context that it thinks it needs to perform a task. In reality, in a large codebase I’ve found manual supplying the exact context to Aider with Anthropic’s models to work better.

fho

Same. It's usually very good at generating greenfield projects from scratch, but once you get going you have to manually provide all context to get good results.

botto

I have tried claude code on a few different size code bases and so far very impressed. The code it generates is usually in the same fashion and tone as the rest of the code base.

olalonde

I guess I should have said that it has access to all your code. But I think it tries to be smart about which files it reads depending on the task at hand.

SamDc73

Claude Code is by far the most expensive tool I used so far. I generally use it for fronted stuff since it's wonderful there !

cwegener

I don't use IDEs to begin with. I use a text editor, since I am a UNIX man. So, I don't even have a reason to look at the "AI" flavor of an IDE. The closest thing to an AI IDE that I have tried out were tools like AIDER and Jack's "goose" agent. Neither of those specialized tools has been satisfactory. They all performed worse than just the LLM IMO. I am sticking to crafting my own context that I supply to the LLM. Tools like Simon W's `llm` tool help A LOT to be more efficient at using LLMs in a daily setting.

k4rli

I'm mostly on the same page, but jumping into a new project with a not-so-familiar language has been made much less painful with Cursor. Regular anthropic chats spew nonsense most of the time and are disappointing, but Cursor really seems to be useful.

Of course it also gets stuff wrong and everything needs to be properly validated, but it's a nice tool to try out.

bilekas

I didn't know Unix users were not able to use IDEs. The more you know /s

Edit : Sarcastic

fmxsh

I don't know about Unix users, but perhaps, instead, it is they are able to not use IDEs.

bilekas

I like to laugh at myself a lot but this made me really giggle. Well phrased!

fho

Not true and if only then by choice.

VSCode/Cursor run natively under Linux.

bilekas

It obviously wasn't obvious enough but I was being sarcastic. My fault anyway, pre coffee.

rcarmo

I use VS Code on Fedora Silverblue.

maxcomperatore

agentic ides are better for vibe coding because every time prompted it has the context of the whole codebase and it can edit multiple files and even run console commands or write docs, its easier to vibe debug too, so its much better than copypasting file by file with something like google ai studio. though agentic ides usually come with an elevated price tag.

reidbarber

My current favorite (and free) workflow:

1. Drop project files into https://files2prompt.com to copy a prompt with all file contents

2. Paste into https://aistudio.google.com/ and set a low (or 0) temperature and low top_p

Since Gemini 2.5 Pro is free in AI Studio at the moment, and there's a 1M token limit, this works for most things I need. Cursor is better in some cases where I need a bunch of small edits across a bunch of different files.

glohbalrob

Just curious, why change the temp to 0

reidbarber

I've found that to be best for coding, especially when you're describing the exact behavior you want implemented. Plus the format it returns is very consistent (not the case in the various chat UIs).

alanfranz

You don’t want the model to stray or become too inventive when coding.

LanceJones

I've been using Windsurf for a few weeks. I'm a novice programmer trying to build a web app using React and NextJS.

The "context-in-the-codebase" thing for AI-based IDEsis overrated IMO. To extract the most from it, you have to remember to @mention the files.

If you don't remember to @mention specific files, the agent simply tries to perform searches (i.e., it has access to that tool) on the files and folders until it gets some results... and will usually keep broadening the search until it does.

It works well enough I suppose. But I still find myself beginning new chats (for example, per feature) because the model still loses its place, and with all the code/lint fixes it does, it starts to lose context.

Then you're right back having to @mention more files to ensure the model knows how to structure your front end, back end, etc.

(Please excuse any misnamed development terms.) :-)

dangus

I don't think you're quite right about this, as I've noticed (at least in Cursor) that appropriate context tends to be pulled in automatically.

What I do notice is that the AI systems seem to forget to pull in a lot of context as the project size grows. It's almost as if there's something of a limit to the amount of input data, either to manage costs or compute cycles or something like that.

When I start "vibecoding" it goes really well in the beginning but then when the project grows in complexity, asking it to do something relatively simple falls apart as the AI "forgets" to do that really simple thing in every applicable place (e.g., removing a feature that is no longer needed).

rcarmo

In VS Code, you automatically get the current file context and you can add all open files, specific files, etc. just before you hit send.

ghiculescu

I rolled out Claude Code and Cursor to my whole company. They’re really much better than copy paste. And very affordable compared to professional developer wages.

Wrote about it: https://ghiculescu.substack.com/p/nobody-codes-here-anymore

munksbeer

I'm relatively ignorant in this respect, but how do you "roll out Cursor"? Does that mean developers must now switch to using the Cursor IDE? That's quite an ask. I really don't think I'd like having to switch IDEs at this point.

ghiculescu

We didn’t mandate it. We just told everyone they can buy if they want to.

dangus

What's so difficult to understand? How do you get your other software at work?

The developers now have access to the license to the software, probably via their work email address.

I don't think the person you're replying to necessarily said that they are forcing their developers to use the tool.

Aurornis

> The issue with Copilot is that it's not as smart as the "thinking" chat apps.

Upgrade to a paid plan and you can use many of the same “thinking” models.

Honestly, your entire question is best answered by signing up for a free trial of one of these tools and using it. Not the free tier, a trial or a paid plan.

Copy and pasting into another app is extremely inefficient. You really need to try a tool that integrates your model of choice into the editor.

jeswin

I've finally settled on this:

- A custom vscode plugin to help me copy file contents from the tree view, along with file paths

- A chat based ide (LibreChat, but anything will do)

- An agent that syncs code back, once I'm happy with the conversation (https://github.com/codespin-ai/codebox)

Sometimes I add the agent right at the beginning. Sometimes I do that when I feel the code is ready to be written back. Another thing with the codebox agent is that it lets the agent run unit tests (even inside a docker compose environment, if you want integration tests); so in some cases it can be fire-and-forget.

rcarmo

Why use an external tool if you now have agent mode in VS Code?

barrenko

How "easy" is it is to scaffold a personal vscode plugin?

jeswin

Not that trivial to hand code, I just had this feature (copying) in another larger plugin I wrote.

However, I think it can be "vibe coded"these days.

barrenko

Thanks, overall I should make something like that, but generally optimistic that vscode will soon provide all that's necessary, so laziness prevails.

ValtteriL

I use copilot through VSCode. The copilot plugin lets me either ask questions or ask for edits to my code. I can choose which files to include as an additional context to the prompt. I can additionally choose which model to use. Some models are unlimited, others have some monthly quota. GPT-4o is unlimited and quite good with short tasks.

It has also shortcuts for instance fixing errors in the code by just selecting "fix with copilot". Its a bit hit or miss but for simple things work well.

It is very good and seamless. Cannot think of ever copy pasting multiple files of code to some web chats and back.

cwegener

I think the Coplit plugin is probably the closest thing to an AI IDE that actually works. As per my other comment, I don't actually use any IDE to begin with, but one of my coworkers who uses VSCode has been very fond of the latest iteration of the Copilot plugin.

From your description of how this plugin works, I see why it works so well. It is basically just focusing on the essentials (populating the appropriate context) and then does as best as it can to get out of the way ( <- this part is really important I think)

benfortuna

But you aren't just asking for edits to your code, right? Copilot will actively suggest code completion as you type. This is the real benefit over chat as it's like pair programming.

That, and the way it generates commit comments, I don't think I will write another commit message even again.

danieldk

The first paragraph implies that they are. I’m doing the same, works really well for writing new code or refactoring. Then polish it up by pointing out bad parts in the chat. At the end I step in and do some manual polish.

rcarmo

I use VS Code too and in agent mode I can say things like “move the database querias into a new module called models.py and convert the SQL into SQLModel classes”. It’s pretty awesome to do chores.