Skip to content(if available)orjump to list(if available)

Pitfalls of premature closure with LLM assisted coding

jbellis

I think this is correct, and I also think it holds for reviewing human-authored code: it's hard to do the job well without first having your own idea in your head of what the correct solution looks like [even if that idea is itself flawed].

danielbln

I put the examples he gave into Claude 4(Sonnet) purely asking to eval the code, it pointed out every single issue about the code snippets (N+1 Query, race condition, memory leak). The article doesn;t mention which model was used, or how exactly it was used, or in which environment/IDE it was used.

The rest of the advice in there is sound, but without more specifics I don't know how actionable the section "The spectrum of AI-appropriate tasks" really is.

metalrain

It's not about "model quality". Most models can improve code their output when asked, but problem is the lack of introspection by the user.

shayonj

My experience hasn't changed between models, given the core issue mentioned in the article. Primarily I have used Gemini and Claude 3.x and 4. Some GPT 4.1 here and there.

All via Cursor, some internal tools and Tines Workbench

suddenlybananas

I initially thought that layout of the sections was an odd and terrible poem.

shayonj

haha! I didn't see it that way originally. Shall take it as a compliment and rework that ToC UI a bit :D.

llmenth

[dead]