AI is blurring the line between PMs and engineers?
58 comments
·February 25, 2025ryandvm
Forget PMs and engineers, the discussion I want to see happening is just how much more easily an LLM could replace the average CEO...
tombert
Hey that's a topic I can get behind!
Most of the big CEOs appear to be CEO of multiple companies (Musk and Bezos come to mind). If you can be the CEO of five different companies, it really doesn't seem like it can be that hard of a job. ChatGPT might be able to do that just fine.
Aurornis
> Most of the big CEOs appear to be CEO of multiple companies (Musk and Bezos come to mind).
Most big CEOs are definitely not in charge of multiple companies. You hear about Musk and Bezos because they’re all over the news, but they’re all over the news because they’re not normal CEOs.
It’s well known that Musk’s relationship to his companies is primarily one of ownership and delegation. Bezos hasn’t been in charge of Amazon for a long time but people conveniently forget that at every opportunity.
> it really doesn't seem like it can be that hard of a job.
It’s funny how often I hear this about AI replacing jobs, with one exception: Everyone who repeats it is highly confident that AI can replace other jobs but once you get to the work they do they’ll stop and explain why it’s actually much harder than you think.
throwup238
> It’s funny how often I hear this about AI replacing jobs, with one exception: Everyone who repeats it is highly confident that AI can replace other jobs but once you get to the work they do they’ll stop and explain why it’s actually much harder than you think.
Call it the Turing-Lovelace Amnesia Effect.
amarcheschi
They just work harder and deserve to rule over humanity, nothing to be worried about
mlboss
At that level its more about persuasion and personality than the ability to do the job. High performing CEOs can persuade employees, customers, government, media, investors, suppliers.
LLMs still cannot persuade anybody right now but maybe soon.
apercu
And, at least for one of them, they also tweet all day while somehow being invaluable in providing leadership for like 5 companies. At the same time. I can't even successfully effectively juggle 4-5 _projects_ at a time.
mlboss
Its pretty simple management style....ask what employee did last week. If you don't like the answer then fire them.
DaiPlusPlus
Real-talk: using an LLM for CEO-work makes sense for when hard, rational, decisions need to be made (e.g. new divisions, layoffs, expansion, etc) - but CEOs also need to be a leader and inspire confidence in their subordinates - a non-human agent with non-human motivations simply can’t do that.
Aurornis
> using an LLM for CEO-work makes sense for when hard, rational, decisions need to be made (e.g. new divisions, layoffs, expansion, etc)
Hard disagree. Using an LLM for these decisions transforms it into a game of manipulating inputs.
If you thought it was bad when people were gaming metrics for performance reviews, imagine the nightmare of a company where everyone is trying to manipulate their work to appeal to the current preferences of the prompts HR is feeding to the LLM. Now imagine the HR people manipulating the prompts to get the answers they want, or leaking the prompts to their friends.
Without humans in the loop to recognize when incentives are being gamed it becomes a free for all.
listenallyall
I'd expect that AI making top-level decisions would be disastrous for employees, especially in tech. An AI is going to relentlessly push for efficiency, and human talent simply can't be reduced to a single input to measure against cost (salary, benefits, devtools, etc). AI will constantly suggest lower costs and thus fewer people, it has no perception of product quality or satisfaction, and it's always going to question the value of the employees' contributions to building software (or other stuff), especially if those employees are themselves relying on AI assistance.
mncharity
A theme of pre-LLM computer-human hybrid systems, was one expensive/scarce expert might be replaced by a less-expensive computer-facilitated redundant group of lesser experts.
So perhaps imagine something like Kaggle competitions, but for Harvard Business School case studies. Open to LLMs, humans, and collaboratives.
A first step might be to create a leadership/managerial LLM test set. I wonder if HBS case studies are in training sets. And whether we can generate good case studies yet. Perhaps use military leadership training material?
apercu
Just about every problem is a leadership problem.
6510
In my vision the automated CEO is able to talk to all of the customers, all of the employees and all the investors simultaneously 24/7. It doesn't have to be perfect, if it is able to do that at all it would dramatically change the game. What is also hilarious is that it has a unit of attention so it can accurately divide it over the audience. A human CEO has few tokens that are extremely hard to divide.
Aurornis
> “No, no,” she said. “Engineers aren’t allowed to edit the prompts. It’s only the PMs and domain experts who do prompt engineering. They do it in a custom UI, and then the prompts are committed to the codebase.”
This feels like a desperate power grab.
Why can’t engineers be involved with the prompts? Why aren’t they allowed to do things like automated A/B testing or to implement ideas from papers they’ve read?
Banning engineers from prompts altogether feels extremely arbitrary. If this was a healthy relationship they’d at least say the engineers and PMs work together on prompts.
Banning them is just politics.
slowmovintarget
"Job security for me but not for thee."
develoopest
Wishing the best to my PM, who struggles to write a ticket title.
diegof79
These company blog articles are usually for marketing. Humanloop develops software to assist with prompting during the development process, so the author’s conclusions reflect the company’s intentions more than an objective industry observation.
AI is transforming how we prototype and iterate, and products like v0 or Replit are scratching the surface. However, historically, low-code platforms lacked a good integration with complex development cycles. There were many attempts, but they either failed or shifted their focus: Microsoft Expression Blend had a brilliant concept of integrating early sketching and ideation with development, but the product ultimately died with Silverlight; Framer had an editor that allowed users to integrate React components with a design tool, but they repurposed their product into a CMS-oriented tool like Webflow; Builder.io is following a similar path. It seems that in today’s market, there is no clear fit for the RAD tools of the late 1990s. Maybe AI can change that and create the new equivalent to Visual Basic. The hardest part is the extra mile that goes from the prototype to something robust and complies with multiple quality attributes: scalability, performance, security, maintainability, and testability.
djohnston
I’m SWE on a product and generally agree with this, at least for user-facing prompts. Once you’re walking an LLM thru OOXML it’s better done directly by ENG.
For tools, not clear how this works since as you adjust parameters and whatnot you’re also presumably changing the code downstream when you “execute the call”.
But probably both sides of this will be done by LLMs directly in the future. I rarely write or tune prompts by hand now.
1970-01-01
A joke: Good PMs will do invisible work. Bad PMs are just invisible.
margalabargala
As anyone who has worked with a bad PM can tell you, they are anything but invisible.
corysama
I've been watching several "vibe coding" vids on the youtubes. https://x.com/karpathy/status/1886192184808149383
Some of them are painful. Some are impressive. The projects are small. Sometimes pulling in powerful off-the-shelf modules. They are getting better fast.
As a greybeard software architect, my current annoyance is that I'm spending all day talking to people when I want to get some practice voice prompting code :P
jfbfkdnxbdkdb
Whisper.CPP + llama.CPP = what you want??
ilrwbwrkhv
i think the plan of folks like these is keep peddling agi, replacement, line blurring and make the money through secondaries for as long as possible till it all collapses like crypto.
bufferoverflow
PMs (generally) can't tell if AI wrote bad code.
marcusestes
Absolutely agree with this observation. And I think it’s a great outcome.
The vibe coding Reddit (http://reddit.com/r/vibecoding) already contains the full spectrum of “first time trying to code” to “just rolled my own custom GPT to optimize this.”
tombert
These articles are a little annoying.
I love ChatGPT [1]. I use it all the time. I use it for coding, I use it to generate stuff like form letters, I use it for parsing out information from PDFs. Point is, I'm not a luddite with this stuff; I'm perfectly happy to play with and use new tech, including but not limited to AI.
Which makes me confident when I say this: Anyone who thinks that AI in its current state is "blurring the line between PMs and Engineers" doesn't know what they are talking about. ChatGPT is definitely very useful, but it's nowhere near a replacement for an engineer.
ChatGPT is really only useful if you already kind of know what you want. Like, if I asked it "I have a table with the columns name (a string), age (an integer), location (string), can you write me an upsert statement for Postgres for the values 'tom', 34, 'new york'?". This will likely give you exactly what you want, will give you the proper "ON CONFLICT" command, and it's cool and useful.
If I ask it "I want to put a value into a table. I also want to make sure that if there's a value in there, we don't just put the value in there, but instead we get the value, update it, and then put the new value back in", it's not as guaranteed to be correct. It might give you the upsert command, but it also might fetch the value from the database, check if it exists, and if it doesn't do an "insert" and if it does do an "update", which is likely incorrect because you risk race conditions.
My point is, the first example required knowing what an upsert is, and how to word it in a technical and precise way.
It certainly doesn't "blur the line" between PM and engineer for me. I have to pretty heavily modify and babysit its outputs, even when it is giving me useful stuff. You might be saying "well that's what a PM does!!", but not really; project managers aren't typically involved in the technical minutia of a project in my experience, they're not going to correct me for using the wrong kind of upsert.
These kinds of articles always seem to be operating on a theoretical "what if AIs could do this??" plane of existence.
[1] Deepseek is cool too, what I'm saying applies to that as well.
ETA:
Even if I wasn't a fan, this article definitely shouldn't have been flagged.
apercu
I feel like I agree with you - AI is great if you're a seasoned, experienced person within the domain you are asking help for. If you're not, there is no way to know if what you are getting as feedback is accurate or not.
AI is like a lot of "Startup Founders", great at presenting all sorts of things as facts, but if you really dig and drill, they don't really have any domain expertise.
razcle
Hey Tombert,
wrt did you read the article? I was quite specific about the ways I think LLMs are blurring the lines. I don't think its true for general engineering but I do think its true for applications being built with LLMs.
Also its still very early
nexus_six
It's not early. It has reached a plateau. Are there betting odds for "AI" (LLM) benchmarks somewhere? I will bet money
tombert
Reaching a plateau doesn't imply that it's not early. It's still entirely plausible that we come up with a newer better model in ten years that gives us true AGI or runs on cheaper hardware or just gives us a closer approximation to human reasoning.
rsynnott
> Also its still very early
“Jam tomorrow” will only get you so far.
tombert
I'm just saying that I don't think I agree with some of this; even if PMs are writing the prompts (and calling that "prompt engineering"), it's not equivalent because they don't know how to audit the code given to them.
A PM might generate that SQL thing I mentioned and just blindly cut and paste it. For any application with more than one user, that is a bug, it's incorrect, and it's not like this is some deep cut: upserts happen all over the place.
I didn't finish the entire article, I disagreed with the line, "Prompting Is Here To Stay and PMs—Not Engineers—Are Going to Do It", because I fundamentally do not think that is true unless AI models get considerably better.
It's possible they will, maybe OpenAI will crack AGI or maybe these models will just get a lot better at figuring out your intent or maybe there's another variable that I'm not thinking of that will fix it.
I hate the term "prompt engineer" because I don't think it's engineering, at least not really. I will agree that there's a skill to getting ChatGPT to give you what you want, I think I'm pretty good at it even, but I hesitate to call it engineering because it lacks a lot of "objectivity". I can come up with a "good" prompt that will 90% of the time give me a good answer, but 10% of the time give me utter rubbish, which doesn't really feel like engineering to me.
I saw the line: `As AI models become able to write complex applications end-to-end, the work of an engineer will increasingly start to resemble that of a product manager.`, and while I don't completely disagree, I also don't completely agree either. Even when I heavily abuse ChatGPT for code generation, it doesn't feel at all like I'm barking orders to a human. It might superficially resemble it but I'm not convinced that it's actually that similar.
I hope I'm not coming off as too much of a dick here, I apologize if I am, and obviously a blog post in which you wax philosophical about the implications of new technology is perfectly fine. I think I'm just a bit on edge with this stuff because you get morons like Zuckerberg claiming they'll be able to replace all their junior and mid level engineers with AI soon, and I think that's ridiculous unless they have access to considerably better models than I do.
arcsincosin
My read–which may be wrong–is that much of the article is discussing applications where the end user is interacting with an interface that queries an LLM using baked-in prompts (in one case, a marketing content generation tool). These prompts are being written by the PM. The PM is not writing prompts LLMs to generate code, the PM is writing prompts which are hidden behind a web form or button or something in an interface, hence the prompts being part of the codebase. The author argues that when a PM is editing these prompts they are delivering an artifact that looks more like an engineer's artifact than a PM's artifact, traditionally.
I'm fortunate to work for a company with great PMs at the moment, but in the past, it's usually been the case that PMs blur the line between PMs and engineers by making us do their job for them. That they can use AI to appear productive is not a promising development.