Skip to content(if available)orjump to list(if available)

AI Horseless Carriages

AI Horseless Carriages

329 comments

·April 23, 2025

joshstrange

I could not agree more with this. 90% of AI features feel tacked on and useless and that’s before you get to the price. Some of the services out here are wanting to charge 50% to 100% more for their sass just to enable “AI features”.

I’m actually having a really hard time thinking of an AI feature other than coding AI feature that I actually enjoy. Copilot/Aider/Claude Code are awesome but I’m struggling to think of another tool I use where LLMs have improved it. Auto completing a sentence for the next word in Gmail/iMessage is one example, but that existed before LLMs.

I have not once used the features in Gmail to rewrite my email to sound more professional or anything like that. If I need help writing an email, I’m going to do that using Claude or ChatGPT directly before I even open Gmail.

petekoomen

One of the interesting things I've noticed is that the best experiences I've had with AI are with simple applications that don't do much to get in the way of the model, e.g. chatgpt and cursor/windsurf.

I'm hopeful that as devs figure out how to build better apps with AI we'll have have more and more "cursor moments" in other areas in our lives

danielbln

I enjoy Claude as a general purpose "let's talk about this niche thing" chat bot, or for general ideation. Extracting structured data from videos (via Gemini) is quite useful as well, though to be fair it's not a super frequent use case for me.

That said, coding and engineering is by far the most common usecase I have for gen AI.

joshstrange

Oh, I'm sorry if it wasn't clear. I use Claude and ChatGPT to talk to about a ton of topics. I'm mostly referring to AI features being added to existing SaaS or software products. I regularly find that moving the conversation to ChatGPT or Claude is much better than trying to use anything that they may have built into their existing product.

teeray

> This demo uses AI to read emails instead of write them

LLMs are so good at summarizing that I should basically only ever read one email—from the AI:

You received 2 emails today that need your direct reply from X and Y. 1 is still outstanding from two days ago, _would you like to send an acknowledgment_? You received 6 emails from newsletters you didn’t sign up for but were enrolled after you bought something _do you want to unsubscribe from all of them_ (_make this a permanent rule_).

namaria

I have fed LLMs PDF files, asked about the content and gotten nonsense. I would be very hesitant to trust them to give me an accurate summary of my emails.

HdS84

One of our managers uses Ai to summarize everything. Too bad it missed important caveats for an offer. Well, we burned an all nighters to correct the offer, but he did not read twenty pages but one...

nradov

LLMs are terrible at summarizing technical emails where the details matter. But you might get away with it, at least for a while, in low performing organizations that tolerate preventable errors.

imp0cat

This. LLMs seem to be great for 90+% of stuff, but sometimes, they just spew weird stuff.

koolba

> LLMs are so good at summarizing that I should basically only ever read one email—from the AI

This could get really fun with some hidden text prompt injection. Just match the font and background color.

Maybe these tools should be doing the classic air gap approach of taking a picture of the rendered content and analyzing that.

joshstrange

What system are you using to do this? I do think that this would provide value for me. Currently, I barely read my emails, which I'm not exactly proud of, but it's just the reality. So something that summarized the important things every day would be nice.

sanderjd

I think the other application besides code copiloting that is already extremely useful is RAG-based information discovery a la Notion AI. This is already a giant improvement over "search google docs, and slack, and confluence, and jira, and ...".

Just integrated search over all the various systems at a company was an improvement that did not require LLMs, but I also really like the back and forth chat interface for this.

knightscoop

I wonder sometime if this is why there is such an enthusiasm gap over AI between tech people and the general public. It's not just that your average person can't program; it's that they don't even conceptually understand why programming could unlock.

nicolas_t

I like perplexity when I need a quick overview of a topic with references to relevant published studies. I often use it when researching what the current research says on parenting questions or education. It's not perfect but because the answers link to the relevant studies it's a good way to get a quick overview of research on a given topic

bigstrat2003

Honestly I don't even enjoy coding AI features. The only value I get out of AI is translation (which I take with a grain of salt because I don't know the other language and can't spot hallucinations, but it's the best tool I have), and shitposting (e.g. having chatGPT write funny stories about my friends and sending it to them for a laugh). I can't say there's an actual productive use case for me personally.

apwell23

garmin wants me to pay for some gen-ai workout messages on connect plus. Its the most absurd AI slop of all. Same with strava. I workout for mental relaxation and i just hate this AI stuff being crammed in there.

Atleast clippy was kind of cute.

nradov

Strava employees claim that casual users like the AI activity summaries. Supposedly users who don't know anything about exercise physiology didn't know how to interpret the various metrics and charts. I don't know if I believe that but it's at least plausible.

Personally I wish I could turn off the AI features, it's a waste of space.

rurp

Anytime someone from a company says that users like the super trendy thing they just made I take it with a sizeable grain of salt. Sometimes it's true, and maybe it is true for Strava, but I've seen enough cases where it isn't to discount such claims down to ~0.

danielbln

Strava's integration is just so lackluster. It literally turns four numbers from right above the slop message into free text. Thanks Strava, I'm a pro user for a decade, finally I can read "This was a hard workout" after my run. Such useful, much AI.

bigstrat2003

At this point, "we aren't adding any AI features" is a selling point for me. I've gotten real tired of AI slop and hype.

mNovak

Just want to say the interactive widgets being actually hooked up to an LLM was very fun.

To continue bashing on gmail/gemini, the worst offender in my opinion is the giant "Summarize this email" button, sitting on top of a one-liner email like "Got it, thanks". How much more can you possibly summarize that email?

petekoomen

Thank you! @LewisJEllis and I wrote a little framework for "vibe writing" that allows for writing in markdown and adding vibe-coded react components. It's a lot of fun to use!

ipaddr

Can we all quickly move to a point in time where vibe-code is not a word

namaria

I kinda appreciate the fact that vibe as a word is usually a good signal I have no interest in the adjacent content.

skrebbel

What would be better? AI-hack? Claude-bodge? I agree that it's a cringey term but cringey work deserves a cringey term right?

DesaiAshu

My websites have this too with MDX, it's awesome. Reminds me of the old Bret Victor interactive tutorials back around when YC Research was funding HCI experiments

skeptrune

MDX is awesome. Incredibly convenient tooling.

carterschonwald

Very nice example of an actually usefully interactive essay.

bambax

It is indeed a working demo, hitting

  https://llm.koomen.dev/v1/chat/completions
in the OpenAI API format, and it responds to any prompt without filtering. Free tokens, anyone?

More seriously, I think the reason companies don't want to expose the system prompt is because they want to keep some of the magic alive. Once most people understand that the universal interface to AI is text prompts, then all that will remain is the models themselves.

petekoomen

That's right. llm.koomen.dev is a cloudflare worker that forwards requests to openai. I was a little worried about getting DDOSed but so far that hasn't been an issue, and the tokens are ridiculously cheap.

amiantos

Blog author seems smart (despite questionable ideas about how much real world users would want to interact with any of his elaborate feature concepts), you hope he's actually just got a bunch of responses cached and you're getting a random one each time from that endpoint... and that freely sent content doesn't actually hit OpenAI's APIs.

bambax

I tested it with some prompts, it does answer properly. My guess is it just forwards the queries with a key with a cap, and when the cap is reached it will stop responding...

jihadjihad

It's like the memes where people in the future will just grunt and gesticulate at the computer instead.

ChaitanyaSai

Loved those! How are those created?

Xenoamorphous

I used that button in Outlook once and the summary was longer than the original email

oceanplexian

A lot of people assume that AI naturally produces this predictable style writing but as someone who has dabbled in training a number of fine tunes that's absolutely not the case.

You can improve things with prompting but can also fine tune them to be completely human. The fun part is it doesn't just apply to text, you can also do it with Image Gen like Boring Reality (https://civitai.com/models/310571/boring-reality) (Warning: there is a lot of NSFW content on Civit if you click around).

My pet theory is the BigCo's are walking a tightrope of model safety and are intentionally incorporating some uncanny valley into their products, since if people really knew that AI could "talk like Pete" they would get uneasy. The cognitive dissonance doesn't kick in when a bot talks like a drone from HR instead of a real person.

palsecam

> My pet theory is the BigCo's are walking a tightrope of model safety and are intentionally incorporating some uncanny valley into their products, since if people really knew that AI could "talk like Pete" they would get uneasy. The cognitive dissonance doesn't kick in when a bot talks like a drone from HR instead of a real person.

FTR, Bruce Schneier (famed cryptologist) is advocating for such an approach:

We have a simple proposal: all talking AIs and robots should use a ring modulator. In the mid-twentieth century, before it was easy to create actual robotic-sounding speech synthetically, ring modulators were used to make actors’ voices sound robotic. Over the last few decades, we have become accustomed to robotic voices, simply because text-to-speech systems were good enough to produce intelligible speech that was not human-like in its sound. Now we can use that same technology to make robotic speech that is indistinguishable from human sound robotic again.https://www.schneier.com/blog/archives/2025/02/ais-and-robot...

MichaelDickens

Reminds me of the robot voice from The Incredibles[1]. It had an obviously-robotic cadence where it would pause between every word. Text-to-speech at the time already knew how to make words flow into each other, but I thought the voice from The Incredibles sounded much nicer than the contemporaneous text-to-speech bots, while also still sounding robotic.

[1] https://www.youtube.com/watch?v=_dxV4BvyV2w

nyanpasu64

That doesn't sound like ring modulation in a musical sense (IIRC it has a modulator above 30 Hz, or inverts the signal instead of attenuating?), so much as crackling, cutting in and out, or an overdone tremolo effect. I checked in Audacity and the signal only gets cut out, not inverted.

momojo

Like adding the 'propane smell' to propane.

Semaphor

Interestingly, it's just kinda hiding the normal AI issues, but they are all still there. I think people know about those "normal" looking pictures, but your example has many AI issues, especially with hands and background

GuinansEyebrows

> but can also fine tune them to be completely human

what does this mean? that it will insert idiosyncratic modifications (typos, idioms etc)?

pmarreck

Loved the fact that the interactive demos were live.

You could even skip the custom system prompt entirely and just have it analyze a randomized but statistically-significant portion of the corpus of your outgoing emails and their style, and have it replicate that in drafts.

You wouldn't even need a UI for this! You could sell a service that you simply authenticated to your inbox and it could do all this from the backend.

It would likely end up being close enough to the mark that the uncanny valley might get skipped and you would mostly just be approving emails after reviewing them.

Similar to reviewing AI-generated code.

The question is, is this what we want? I've already caught myself asking ChatGPT to counterargue as me (but with less inflammatory wording) and it's done an excellent job which I've then (more or less) copy-pasted into social-media responses. That's just one step away from having them automatically appear, just waiting for my approval to post.

Is AI just turning everyone into a "work reviewer" instead of a "work doer"?

petekoomen

honestly you could try this yourself today. Grab a few emails, paste them into chatgpt, and ask it to write a system prompt that will write emails that mimic your style. Might be fun to see how it describes your style.

to address your larger point, I think AI-generated drafts written in my voice will be helpful for mundane, transaction emails, but not for important messages. Even simple questions like "what do you feel like doing for dinner tonight" could only be answered by me, and that's fine. If an AI can manage my inbox while I focus on the handful of messages that really need my time and attention that would be a huge win in my book.

crote

It all depends on how you use it, doesn't it?

A lot of work is inherently repetitive, or involves critical but burdensome details. I'm not going to manually write dozens of lines of code when I can do `bin/rails generate scaffold User name:string`, or manually convert decimal to binary when I can access a calculator within half a second. All the important labor is in writing the prompt, reviewing the output, and altering it as desired. The act of generating the boilerplate itself is busywork. Using a LLM instead of a fixed-functionality wizard doesn't change this.

The new thing is that the generator is essentially unbounded and silently degrades when you go beyond its limits. If you want to learn how to use AI, you have to learn when not to use it.

Using AI for social media is distinct from this. Arguing with random people on the internet has never been a good idea and has always been a massive waste of time. Automating it with AI just makes this more obvious. The only way to have a proper discussion is going to be face-to-face, I'm afraid.

emaro

About writing a counterargument for social media: I kinda get it, but what's the end game of this? People reading generated responses others (may have) approved? Do we want that? I think I don't.

mvieira38

It's what we want, though, isn't it? AI should make our lives easier, and it's much easier (and more productive) to review work already done than to do it yourself. Now, if that is a good development morally/spiritually for the future of mankind is another question... Some would argue industrialization was bad in that respect and I'm not even sure I fully disagree

ai_

No? Not everyone's dream is being a manager. I like writing code, it's fun! Telling someone else to go write code for me so that I can read it later? Not fun, avoid it if possible (sometimes it's unavoidable, we don't have unlimited time).

bluGill

What is the point? The effort to write the email is equal to the effort to ask the AI to write the email for you. Only when the AI turns your unprofessional style into something professional is any effort saved - but the "professional" sounding style is most of the time wrong and should get dumped into junk.

tlogan

At the end of the day, it comes down to one thing: knowing what you want. And AI can’t solve that for you.

We’ve experimented heavily with integrating AI into our UI, testing a variety of models and workflows. One consistent finding emerged: most users don’t actually know what they want to accomplish. They struggle to express their goals clearly, and AI doesn’t magically fill that gap—it often amplifies the ambiguity.

Sure, AI reduces the learning curve for new tools. But paradoxically, it can also short-circuit the path to true mastery. When AI handles everything, users stop thinking deeply about how or why they’re doing something. That might be fine for casual use, but it limits expertise and real problem-solving.

So … AI is great—but the current diarrhea of “let’s just add AI here” without thinking through how it actually helps might be a sign that a lot of engineers have outsourced their thinking to ChatGPT.

petekoomen

> They struggle to express their goals clearly, and AI doesn’t magically fill that gap—it often amplifies the ambiguity.

One surprising thing I've learned is that a fast feedback loop like this:

1. write a system prompt 2. watch the agent do the task, observe what it gets wrong 3. update the system prompt to improve the instructions

is remarkably useful in helping people write effective system prompts. Being able to watch the agent succeed or fail gives you realtime feedback about what is missing in your instructions in a way that anyone who has ever taught or managed professionally will instantly grok.

kristjank

I have also experienced this in the specific domain of well-learned idiots finding pseudo-explanations for why a technical choice should be taken, despite not knowing anything about the topic.

I have witnessed a colleague look up a component datasheet on ChatGPT and repeating whatever it told him (despite the points that it made weren't related to our use case). The knowledge monopoly in about 10 years when the old-guard programming crowd finally retires and/or unfortunately dies will be in the hands of people that will know what they don't know and be able to fill the gaps using appropriate information sources (including language models). The rest will probably resemble Idiocracy on a spectrum from frustrating to hilarious.

crote

I think a big problem is that the most useful AI agents essentially go unnoticed.

The email labeling assistant is a great example of this. Most mail services can already do most of this, so the best-case scenario is using AI to translate your human speech into a suggestion for whatever format the service's rules engine uses. Very helpful, not flashy: you set it up once and forget about it.

Being able to automatically interpret the "Reschedule" email and suggest a diff for an event in your calendar is extremely useful, as it'd reduce it to a single click - but it won't be flashy. Ideally you wouldn't even notice there's a LLM behind it, there's just a "confirm reschedule button" which magically appears next to the email when appropriate.

Automatically archiving sales offers? That's a spam filter. A really good one, mind you, but hardly something to put on the frontpage of today's newsletters.

It can all provide quite a bit of value, but it's simply not sexy enough! You can't add a flashy wizard staff & sparkles icon to it and charge $20 / month for that. In practice you might be getting a car, but it's going to look like a horseless carriage to the average user. They want Magic Wizard Stuff, not invest hours into learning prompt programming.

petekoomen

> Most mail services can already do most of this

I'll believe this when I stop spending so much time deleting email I don't want to read.

sanderjd

Yeah but I'm looking forward to the point where this is not longer about trying to be flashy and sexy, but just quietly using a new technology for useful things that it's good at. I think things are headed that direction pretty quickly now though! Which is great.

crote

Honestly? I think the AI bubble will need to burst first. Making the rescheduling of appointments and dozens of tasks like that slightly more convenient isn't a billion-dollar business.

I don't have a lot of doubt that it is technically doable, but it's not going to be economically viable when it has to pay back hundreds of billions of dollars of investments into training models and buying shiny hardware. The industry first needs to get rid of that burden, which means writing off the training costs and running inference on heavily-discounted supernumerary hardware.

sanderjd

Yeah this sounds right to me.

kristjank

I tread carefully with anyone that by default augments their (however utilitarian or conventionally bland) messages with language models passing them as their own. Prompting the agent to be as concise as you are, or as extensive, takes just as much time in the former case, and lacks the underlying specificity of your experience/knowledge in the latter.

If these were some magically private models that have insight into my past technical explanations or the specifics of my work, this would be a much easier bargain to accept, but usually, nothing that has been written in an email by Gemini could not have been conceived of by a secretary in the 1970s. It lacks control over the expression of your thoughts. It's impersonal, it separates you from expressing your thoughts clearly, and it separates your recipient from having a chance to understand you the person thinking instead of you the construct that generated a response based on your past data and a short prompt. And also, I don't trust some misandric f*ck not to sell my data before piping it into my dataset.

I guess what I'm trying to say is: when messaging personally, summarizing short messages is unnecessary, expanding on short messages generates little more than semantic noise, and everything in between those use cases is a spectrum deceived by the lack of specificity that agents usually present. Changing the underlying vague notions of context is not only a strangely contortionist way of making a square peg fit an umbrella-shaped hole, it pushes around the boundaries of information transfer in a way that is vaguely stylistic, but devoid of any meaning, removed fluff or added value.

petekoomen

Agreed! As i mentioned in the piece I don't think LLMs are very useful for original writing because instructing an agent to write anything from scratch inevitably takes more time than writing it yourself.

Most of the time I spend managing my inbox is not spent on original writing, however. It's spent on mundane tasks like filtering, prioritizing, scheduling back-and-forths, introductions etc. I think an agent could help me with a lot of that, and I dream of a world in which I can spend less time on email and finally be one of those "inbox zero" people.

Retric

The counter argument is some people are terrible at writing. Millions of people sit at the bottom of any given bell curve.

I’d never trust a summery from a current generation LLM for something as critical as my inbox. Some hypothetical drastically improved future AI, sure.

petekoomen

Smarter models aren't going to somehow magically understand what is important to you. If you took a random smart person you'd never met and asked them to summarize your inbox without any further instructions they would do a terrible job too.

You'd be surprised at how effective current-gen LLMs are at summarizing text when you explain how to do it in a thoughtful system prompt.

derektank

For the case of writing emails, I tend to agree though I think creative writing is an exception. Pairing with an LLM really helps overcome the blank page / writer's block problem because it's often easier to identify what you don't want and then revise all the flaws you see.

jonplackett

Why can’t the LLM just learn your writing style from your previous emails to that person?

Or a your more general style for new people.

It seems like Google at least should have a TONNE of context to use for this.

Like in his example emails about being asked to meet - it should be checking the calendar for you and putting in if you can / can’t or suggesting an alt time you’re free.

If it can’t actually send emails without permission there’s less harm with giving an LLM more info to work with - and it doesn’t need to get it perfect. You can always edit.

If it deals with the 80% of replies that don’t matter much then you have 5X more time to spend on the 20% that do matter.

unoti

> Why can’t the LLM just learn your writing style from your previous emails to that person?

It totally could. For one thing you could fine tune the model, but I don't think I'd recommend that. For this specific use case, imagine an addition to the prompt that says """To help you with additional context and writing style, here snippets of recent emails Pete wrote to {recipient}: --- {recent_email_snippets} """

samrolken

They are saving this for some future release I would guess. A “personalization”-focused update wave/marketing blitz/privacy Overton window shift.

jon_richards

Writing an email with AI and having the recipient summarize it with AI is basically all the fun of jpeg compression, but more bandwidth instead of less.

https://m.youtube.com/watch?v=jmaUIyvy8E8

skeptrune

>As I mentioned above, however, a better System Prompt still won't save me much time on writing emails from scratch.

>The thing that LLMs are great at is reading text and transforming it, and that's what I'd like to use an agent for.

Interestingly, the OP agrees with you here and noted in the post that the LLMs are better at transforming data than creating it.

kristjank

I reread those paragraphs. I find the transformative effect of the email missing from the whole discussion. The end result of the inbox examples is to change some internal information in the mind of the recipient. Agent working within the context of the email has very little to contribute because it does not know the OP's schedule, dinner plans, whether he has time for the walk and talk or if he broke his ankle last week... I'd be personally afraid to have something rummaging in my social interface that can send (and let's be honest, idiots will CtrlA+autoreply their whole inboxes) invites, timetables, love messages etc. in my name. It has too many lemmas that need to be fulfilled before it can be assumed competent, and none of those are very well demonstrated. It's cold fusion technology. Feasible, should be nice if it worked, but it would really be a disappointment if someone were to use it in its current state.

jimbokun

A lot of people would love to have a 1970s secretary capable of responding to many mundane requests without any guidance.

bluGill

I have a large part of that though. The computer (outlook today) just schedules meetings rooms for me ensuring there are not multiple different meetings in it at the same time. I can schedule my own flights.

When I first started working the company rolled out the first version of meeting scheduling (it wasn't outlook), and all the other engineers loved it - finally they could figure out how to schedule our own meetings instead of having the secretary do it. Apparently the old system was some mainframe based things other programmers couldn't figure out (I never worked with it so I can't comment on how it was). Likewise scheduling a plane ticket involved calling travel agents and spending a lot of time on hold.

If you are a senior executive you still have a secretary. However by the 1970s the secretary for most of us would be department secretary that handled 20-40 people not just our needs, and thus wasn't in tune with all those details. However most of us don't have any needs that are not better handled by a computer today.

kristjank

I would too, but I would have to trust AI at least as much as a 1970s secretary not to mess up basic facts about myself or needlessly embellish/summarize my conversations with known correspondents. Comparing agents and past office cliches was not to imply agents do it and it's stupid; I'm implying agents claim to do it, but don't.

AlienRobot

So AI is SaaS (Secretary as a Service)

AndrewHart

Aside from saving time, I'm bad at writing. Especially emails. I often open ChatGPT, paste in the whole email chain, write out the bullets of the points I want to make and ask it to draft a response which frames it well.

ori_b

I'd prefer to get the bullet points. There's no need to waste time reading autogenerated filler.

Swizec

> write out the bullets of the points I want to make

Just send those bullet points. Everyone will thank you

worik

My boss does that I am sure

One of their dreadful behaviors, among many

My advice is to stop doing this for the sake of your colleagues

foxglacier

There's a whole lot of people who struggle to write professionally or when there's any sort of conflict (even telling your boss you won't come to work). It can be crippling trying to find the right wording and certainly take far longer than writing a prompt. AI is incredible for these people. They were never going to express their true feelings anyway and were just struggling to write "properly" or in a way that doesn't lead to misunderstandings. If you can just smash out good emails without a second thought, you wouldn't need it.

calf

AI for writing or research is useful like a dice roll. Terence Tao famously showed how talking to an LLM gave him an idea/approach to a proof that he hadn't immediately thought of (but probably he would have considered it eventually). The other day I wrote an unusal, four-word neologism that I'm pretty sure no one has ever seen, and the AI immediately drew the correct connection to more standard terminology and arguments used, so I did not even have to expand/explain and write it out myself.

I don't know but I am considering the possibility that even for everyday tasks, this kind of exploratory shortcut can be a simple convenience. Furthermore, it is precisely the lack of context that enables LLMs to make these non-human, non-specific connective leaps, their weakness also being their strength. In this sense, they bode as a new kind of discursive common-ground--if human conversants are saying things that an LLM can easily catch then LLMs could even serve as the lowest-common-denominator for laying out arguments, disagreements, talking past each other, etc. But that's in principle, and in practice that is too idealistic, as long as these are built and owned as capitalist IPs.

benterris

I really don't get why people would want AI to write their messages for them. If I can write a concise prompt with all the required information, why not save everyone time and just send that instead ? And especially for messages to my close ones, I feel like the actual words I choose are meaningful and the process of writing them is an expression of our living interaction, and I certainly would not like to know the messages from my wife were written by an AI. On the other end of the spectrum, of course sometimes I need to be more formal, but these are usually cases where the precise wording matters, and typing the message is not the time-consuming part.

IshKebab

> If I can write a concise prompt with all the required information, why not save everyone time and just send that instead ?

This point is made multiple times in the article (which is very good; I recommend reading it!):

> The email I'd have written is actually shorter than the original prompt, which means I spent more time asking Gemini for help than I would have if I'd just written the draft myself. Remarkably, the Gmail team has shipped a product that perfectly captures the experience of managing an underperforming employee.

> As I mentioned above, however, a better System Prompt still won't save me much time on writing emails from scratch. The reason, of course, is that I prefer my emails to be as short as possible, which means any email written in my voice will be roughly the same length as the User Prompt that describes it. I've had a similar experience every time I've tried to use an LLM to write something. Surprisingly, generative AI models are not actually that useful for generating text.

stronglikedan

People like my dad, who can't read, write, or spell to save his life, but was a very, very successful CPA, would love to use this. It would have replaced at least one of his office staff I bet. Too bad he's getting up there in age, and this newfangled stuff is difficult for him to grok. But good thing he's retired now and will probably never need it.

tarboreus

What a missed oppurtunity to fire that extra person. Maybe the AI could also figure out how to do taxes and then everyone in the office could be out a job.

istjohn

Well, you know this employment crisis all started when the wheel was invented and put all the porters out of work. Then tech came for lamplighters, ice cutters, knocker-uppers, switchboard operators, telegraph operators, human computers, video store clerks, bowling alley pinsetters, elevator operators, film developers, lamp lighters, coopers, wheelwrights, candle makers, weavers, plowmen, farriers, street sweepers. It's a wonder anyone still has a job, really.

DrillShopper

Let's just put an AI in charge of the IRS and have it send us an actual bill which is apparently something that just too complicated for the current and past IRS to do./s

Edit: added /s because it wasn't apparent this was sarcastic

ARandumGuy

Shorter emails are better 99% of the time. No one's going to read a long email, so you should keep your email to just the most important points. Expanding out these points to a longer email is just a waste of time for everyone involved.

My email inbox is already filled with a bunch of automated emails that provide me no info and waste my time. The last thing I want is an AI tool that makes it easier to generate even more crap.

mitthrowaway2

Definitely. Also, another thing that wastes time is when requests don't provide the necessary context for people to understand what's being asked for and why, causing them to spend hours on the wrong thing. Or when the nuance is left out of a nuanced good idea causing it to get misinterpreted and pattern-matched to a similar-sounding-but-different bad idea, causes endless back-and-forth misunderstandings and escalation.

Emails sent company-wide need to be especially short, because so many person-hours are spent reading them. Also, they need to provide the most background context to be understood, because most of those readers won't already share the common ground to understand a compressed message, increasing the risk of miscommunication.

This is why messages need to be extremely brief, but also not.

nosianu

There was an HN topic less than a month ago or so where somebody wrote a blog post speculating that you end up with some people using AI to write lengthy emails from short prompts adhering to perfect polite form, while the other people use AI to summarize those blown-up emails back into the essence of the message. Side effect, since the two transformations are imperfect meaning will be lost or altered.

dang

Can anybody find the thread? That sounds worth linking to!

philipkglass

It was more than a month ago, but perhaps this one:

https://news.ycombinator.com/item?id=42712143

How is AI in email a good thing?!

There's a cartoon going around where in the first frame, one character points to their screen and says to another: "AI turns this single bullet point list into a long email I can pretend I wrote".

And in the other frame, there are two different characters, one of them presumably the receiver of the email sent in the first frame, who says to their colleague: "AI makes a single bullet point out of this long email I can pretend I read".

The cartoon itself is the one posted above by PyWoody.

pizzathyme

If that's the case, you can easily only write messages to your wife yourself.

But for the 99 other messages, especially things that mundanely convey information like "My daughter has the flu and I won't be in today", "Yes 2pm at Shake Shack sounds good", it will be much faster to read over drafts that are correct and then click send.

The only reason this wouldn't be faster is if the drafts are bad. And that is the point of the article: the models are good enough now that AI drafts don't need to be bad. We are just used to AI drafts being bad due to poor design.

allturtles

I don't understand. Why do you need an AI for messages like "My daughter has the flu and I won't be in today" or "Yes 2pm at Shake Shack sounds good"? You just literally send that.

Do you really run these things through an AI to burden your reader with pointless additional text?

djhn

100% agree. Email like you’re a CEO. Saves your time, saves other people’s time and signals high social status. What’s not to like?

_factor

They are automatically drafted when the email comes in, and you can accept or modify them.

It’s like you’re asking why you would want a password manager when you can just type the characters yourself. It saves time if done correctly.

wat10000

Being so direct is considered rude in many contexts.

bigstrat2003

> But for the 99 other messages, especially things that mundanely convey information like "My daughter has the flu and I won't be in today", "Yes 2pm at Shake Shack sounds good", it will be much faster to read over drafts that are correct and then click send.

It takes me all of 5 seconds to type messages like that (I timed myself typing it). Where exactly is the savings from AI? I don't care, at all, if a 5s process can be turned into a 2s process (which I doubt it even can).

ARandumGuy

How would an AI know if "2pm at Shake Shake" works for me? I still need to read the original email and make a decision. The actual writing out the response takes me basically no time whatsoever.

bluGill

An AI could read the email and check my calendar and then propose 2pm. Bonus if the AI works with his AI to figure out that 2pm works for both of us. A lot of time is wasted with people going back and forth trying to figure out when they can meet. That is also a hard problem even before you note the privacy concerns.

sanderjd

Totally agree, for myself.

However, I do know people who are not native speakers, or who didn't do an advanced degree that required a lot of writing, and they report loving the ability to have it clean up their writing in professional settings.

This is fairly niche, and already had products targeting it, but it is at least one useful thing.

bluGill

Cleaning up writing is very different from writing it. Lawyers will not have themselves as a client. I can write a novel or I can edit someone else's novel - but I am not nearly as good at editing my own novels as I would be editing someone else's. (I don't write novels, but I could. As for editing - you should get a better editor than me, but I'd be better than you doing it to your own writing)

fragmede

When it's a simple data transfer, like "2 pm at shake shack sounds good", it's less useful. it's when we're doing messy human shit with deep feelings evoking strong emotions that it shines. when you get to the point where you're trading shitty emails to someone that you, at one point, loved, but are now just getting all up in there and writing some horrible shit. Writing that horrible shit helps you feel better, and you really want to send it, but you know it's not gonna be good, but you just send it anyway. OR - you tell ChatGPT the situation, and have it edit that email before you send it and have it take out the shittiness, and you can have a productive useful conversation instead.

the important point of communicating is to get the other person to understand you. if my own words fall flat for whatever reason, if there are better words to use, I'd prefer to use those instead.

"fuck you, pay me" isn't professional communication with a client. a differently worded message might be more effective (or not). spending an hour agonizing over what to say is easier spent when you have someone help you write it

lawn

There are people who do this but on forums; they rely on AI to write their replies.

And I have to wonder, why? What's the point?

kkoncevicius

For me posts like these go in the right direction but stop mid-way.

Sure, at first you will want an AI agent to draft emails that you review and approve before sending. But later you will get bored of approving AI drafts and want another agent to review them automatically. And then - you are no longer replying to your own emails.

Or to take another example where I've seen people excited about video-generation and thinking they will be using that for creating their own movies and video games. But if AI is advanced enough - why would someone go see a movie that you generated instead of generating a movie for himself. Just go with "AI - create an hour-long action movie that is set in ancient japan, has a love triangle between the main characters, contains some light horror elements, and a few unexpected twists in the story". And then watch that yourself.

Seems like many, if not all, AI applications, when taken to the limit, reduce the need of interaction between humans to 0.

petekoomen

Do you want an LLM writing and sending important messages for you? I don't, and I don't know anyone who does. I want to reduce time I spend managing my inbox, archiving stuff I don't need to read, endless scheduling back-and-forths, etc. etc.

gameman144

> Sure, at first you will want an AI agent to draft emails that you review and approve before sending. But later you will get bored of approving AI drafts and want another agent to review them automatically.

This doesn't seem to me like an obvious next step. I would definitely want my reviewing step to be as simple as possible, but removing yourself from the loop entirely is a qualitatively different thing.

As an analogue, I like to cook dinner but I am only an okay cook -- I like my recipes to be as simple as possible, and I'm fine with using premade spice mixes and such. Now the simplest recipe is zero steps: I order food from a restaurant, but I don't enjoy that as much because it is (similar to having AI approve and send your emails without you) a qualitatively different experience.

null

[deleted]

hiatus

> I order food from a restaurant, but I don't enjoy that as much because it is (similar to having AI approve and send your emails without you) a qualitatively different experience.

What do you like less about it? Is it the smells of cooking, the family checking on the food as it cooks, the joy of realizing your own handiwork?

gameman144

For me, I think it's the act of control and creation -- I can put the things I like together and try new thing and experiment with techniques or ingredients, whereas ordering from a restaurant I'll only be seeing the end results from someone else's experimentation or experience.

I don't dislike restaurants, to be clear -- I love a dinner out. It just scratches a different itch than cooking a meal at home.

bambax

The cooking analogy is good. I too love to cook, and what I make is often not as good as what I could order, but that's not the point. The point is to cook.

saalweachter

So here's where this all feels a bit "build me a better horse" to me.

You're telling an AI agent to communicate specific information on your behalf to specific people. "Tell my boss I can't come in today", "Talk to comcast about the double billing".

That's not abstracted away enough.

"My daughter's sick, rearrange my schedule." Let the agent handle rebooking appointments and figuring out who to notify and how. Let their agent figure out how to convey that information to them. "Comcast double-billed me." Resolve the situation. Communicate with Comcast, get it fixed, if they don't get it fixed, communicate with the bank or the lawyer.

If we're going to have AI agents, they should be AI agents, not AI chatbots playing a game of telephone over email with other people and AI chatbots.

aaronbaugher

Exactly. To be a useful assistant, it has to be more proactive than they're currently able to be.

Someone posted here about an AI assistant he wrote that sounded really cool. But when I looked at it, he had written a bunch of scripts that fetched things like his daily calendar appointments and the weather forecast, fed them to an AI to be worded in a particular way, and then emailed the results to him. So his scripts were doing all the work except wording the messages differently. That's a neat toy, but it's not really an assistant.

An assistant could be told, "Here's a calendar. Track my appointments, enter new ones I tell you about, and remind me of upcoming ones." I can script all that, but then I don't need the AI. I'm trying to figure out how to leverage AI to do something actually new in that area, and not having much luck yet.

a4isms

Short reply:

I agree, it only goes half-way.

Elaboration:

I like the "horseless carriage" metaphor for the transitionary or hybrid periods between the extinction of one way of doing things and the full embrace of the new way of doing things. I use a similar metaphor: "Faster horses," which is exactly what this essay shows: You're still reading and writing emails, but the selling feature isn't "less email," it's "Get through your email faster."

Rewinding to the 90s, Desktop Publishing was a massive market that completely disrupted the way newspapers, magazines, and just about every other kind of paper was produced. I used to write software for managing classified ads in that era.

Of course, Desktop Publishing was horseless carriages/faster horses. Getting rid of paper was the revolution, in the form of email over letters, memos, and facsimiles. And this thing we call the web.

Same thing here. The better interface is a more capable faster horse. But it isn't an automobile.

programd

> You're still reading and writing emails, but the selling feature isn't "less email," it's "Get through your email faster."

The next logical step is not using email (the old horse and carriage) at all.

You tell your AI what you want to communicate with whom. Your AI connects to their AI and their AI writes/speaks a summary in the format they prefer. Both AIs can take action on the contents. You skip the Gmail/Outlook middleman entirely at the cost of putting an AI model in the middle. Ideally the AI model is running locally not in the cloud, but we all know how that will turn out in practice.

Contact me if you want to invest some tens of millions in this idea! :)

mNovak

Taking this a step farther; both AIs also deeply understand and advocate for their respective 'owner', so rather than simply exchanging a formatted message, they're evaluating the purpose and potential fit of the relationship writ large (for review by the 'owner' of course..). Sort of a preliminary discussion between executive assistants or sales reps -- all non-binding, but skipping ahead to the heart of the communication, not just a single message.

echelon

> > Seems like many, if not all, AI applications, when taken to the limit, reduce the need of interaction between humans to 0.

> Same thing here. The better interface is a more capable faster horse. But it isn't an automobile.

I'm over here in "diffusion / generative video" corner scratching my head at all the LLM people making weird things that don't quite have use cases.

We're making movies. Already the AI does things that used to cost too much or take too much time. We can make one minute videos of scale, scope, and consistency in just a few hours. We're in pretty much the sweet spot of the application of this tech. This essay doesn't even apply to us. In fact, it feels otherworldly alien to our experience.

Some stuff we've been making with gen AI to show you that I'm not bullshitting:

- https://www.youtube.com/watch?v=Tii9uF0nAx4

- https://www.youtube.com/watch?v=7x7IZkHiGD8

- https://www.youtube.com/watch?v=_FkKf7sECk4

Diffusion world is magical and the AI over here feels like we've been catapulted 100 years into the future. It's literally earth shattering and none of the industry will remain the same. We're going to have mocap and lipsync, where anybody can act as a fantasy warrior, a space alien, Arnold Schwarzenegger. Literally whatever you can dream up. It's as if improv theater became real and super high definition.

But maybe the reason for the stark contrast with LLMs in B2B applications is that we're taking the outputs and integrating them into things we'd be doing ordinarily. The outputs are extremely suitable as a drop-in to what we already do. I hope there's something from what we do that can be learned from the LLM side, but perhaps the problems we have are just so wholly different that the office domain needs entirely reinvented tools.

Naively, I'd imagine an AI powerpoint generator or an AI "design doc with figures" generator would be so much more useful than an email draft tool. And those are incremental adds that save a tremendous amount of time.

But anyway, sorry about the "horseless carriages". It feels like we're on a rocket ship on our end and I don't understand the public "AI fatigue" because every week something new or revolutionary happens. Hope the LLM side gets something soon to mimic what we've got going. I don't see the advancements to the visual arts stopping anytime soon. We're really only just getting started.

namaria

You make some very strong claims and presented material. I hope I am not out of line if I give you my sincere opinion. I am not doing this to be mean, to put you down or to be snarky. But the argument you're making warrants this response, in my opinion.

The examples you gave as "magical", "100 years into the future", "literally earth shattering" are very transparently low effort. The writing is pedestrian, the timing is amateurish and the jokes just don't land. The inflating tea cup with magically floating plate and the cardboard teabag are... bad. These are bad man. At best recycled material. I am sorry but as examples of why using automatically generated art they are making the opposite argument from what you think you're making.

I categorically do not want more of this. I want to see crafted content where talent shines through. Not low effort, automatically generated stuff like the videos in these links.

hiatus

> Seems like many, if not all, AI applications, when taken to the limit, reduce the need of interaction between humans to 0.

This seems to be the case for most technology. Technology increasingly mediates human interactions until it becomes the middleman between humans. We have let our desire for instant gratification drive the wedge of technology between human interactions. We don't want to make small talk about the weather, we want our cup of coffee a few moments after we input our order (we don't want to relay our orders via voice because those can be lost in translation!). We don't want to talk to a cab driver we want a car to pick us up and drop us off and we want to mindlessly scroll in the backseat rather than acknowledge the other human a foot away from us.

Strilanc

Related short story: the whispering earring http://web.archive.org/web/20121008025245/http://squid314.li...

kkoncevicius

Great suggestion, thank you. It's appropriately short and more fitting than I anticipated. Specially the part about brain atrophy.

braza

> AI applications, when taken to the limit, reduce the need of interaction between humans to 0. > But if AI is advanced enough - why would someone go see a movie that you generated instead of generating a movie for himself.

I would be the first to pay if we have a GenAI that does that.

For a long time I had a issue with a thing that I found out that was normal for other people that is the concept of dreaming.

For years I did not know what was about, or how looks like during the night have dreams about anything due to a light CWS and I really would love to have something in that regard that I could visualise some kind of hyper personalized move that I could watch in some virtual reality setting to help me to know how looks like to dream, even in some kind of awake mode.

scrozier

Are you saying this is what you'd like to happen? That you would like to remove the element of human creation?

bluGill

I'm not sure? Are humans - at least sometimes - more creative?

Many sci-fi novels feature non-humans, but their cultures are all either very shallow (all orcs are violent - there is no variation at all in what any orc wants), or they are just humans with a different name and some slight body variation. (even the intelligent birds are just humans that fly). Can AI do better, or will it be even worse because AI won't even explore what orcs love for violent means for the rest of their cultures and nations.

The one movie set in Japan might be good, but I want some other settings once in a while. Will AI do that?

achierius

Why is "creativity" the end-all be-all? It's easy to get high-entropy white noise -- what we care about is how grounded these things are in our own experience and life, commonalities between what we see in the film and what we live day-to-day.

Terr_

> To illustrate this point, here's a simple demo of an AI email assistant that, if Gmail had shipped it, would actually save me a lot of time:

Glancing over this, I can't help thinking: "Almost none of this really requires all the work of inventing, training, and executing LLMs." There are much easier ways to match recipients or do broad topic-categories.

> You can think of the System Prompt as a function, the User Prompt as its input, and the model's response as its output:

IMO it's better to think of them as sequential paragraphs in a document, where the whole document is fed into an algorithm that tries to predict what else might follow them in a longer document.

So they're both inputs, they're just inputs which conflict with one-another, leading to a weirder final result.

> when an LLM agent is acting on my behalf I should be allowed to teach it how to do that by editing the System Prompt.

I agree that fixed prompts are terrible for making tools, since they're usually optimized for "makes a document that looks like a conversation that won't get us sued."

However even control over the system prompt won't save you from training data, which is not so easily secured or improved. For example, your final product could very well be discriminating against senders based on the ethnicity of their names or language dialects.

Karrot_Kream

The reason so many of these AI features are "horseless carriage" like is because of the way they were incentivized internally. AI is "hot" and just by adding a useless AI feature, most established companies are seeing high usage growth for their "AI enhanced" projects. So internally there's a race to shove AI in as quickly as possible and juice growth numbers by cashing in on the hype. It's unclear to me whether these businesses will build more durable, well-thought projects using AI after the fact and make actually sticky product offerings.

(This is based on my knowledge the internal workings of a few well known tech companies.)

petekoomen

That sounds about right to me. Massive opportunity for startups to reimagine how software should work in just about every domain.

HeyLaughingBoy

Sounds a lot like blockchain 10 years ago!

sanderjd

Totally. I think the comparison between the two is actually very interesting and illustrative.

In my view there is significantly more there there with generative AI. But there is a huge amount of nonsense hype in both cases. So it has been fascinating to witness people in one case flailing around to find the meat on the bones while almost entirely coming up blank, while in the other case progressing on these parallel tracks where some people are mostly just responding to the hype while others are (more quietly) doing actual useful things.

To be clear, there was a period where I thought I saw a glimmer of people being on the "actual useful things" track in the blockchain world as well, and I think there have been lots of people working on that in totally good faith, but to me it just seems to be almost entirely a bust and likely to remain that way.

Karrot_Kream

This happens whenever something hits the peak of the Gartner Hype Cycle. The same thing happened in the social network era (one could even say that the beloved Google Plus was just this for Google), the same thing happened in the mobile app era (Twitter was all about sending messages using SMS lol), and of course it happened during Blockchain as well. The question is whether durable product offerings emerge or whether these products are the throwaway me-too horseless carriages of the AI era.

Meta is a behemoth. Google Plus, a footnote. The goal is to be Meta here and not Google Plus.

kfajdsl

One of my friends vibe coded their way to a custom web email client that does essentially what the article is talking about, but with automatic context retrieval and and more sales oriented with some pseudo-CRM functionality. Massive productivity boost for him. It took him about a day to build the initial version.

It baffles me how badly massive companies like Microsoft, Google, Apple etc are integrating AI into their products. I was excited about Gemini in Google sheets until I played around with it and realized it was barely usable (it specifically can’t do pivot tables for some reason? that was the first thing I tried it with lol).

sanderjd

It's much easier to build targeted new things than to change the course of a big existing thing with a lot of inertia.

This is a very fortunate truism for the kinds of builders and entrepreneurs who frequent this site! :)