AI is impressive because we've failed at personal computing
131 comments
·August 8, 2025headcanon
roadside_picnic
> Semantic web was a standard suggested by Google
Was this written by AI? I find it hard to believe anyone who was interested in Semantic Web would have not known it's origin (or at least that it's origin was not Google).
The concept of a Semantic web was proposed by Tim Berners-Lee (who hopefully everyone recognizes as the father of HTTP, WWW, HTML) in 1999 [0]. Google, to my knowledge, had no direct development or even involvement in the early Semweb standards such as RDF [1] and OWL [2]. I worked with some of the people involved in the latter (not closely though), and at the time Google was still quite small.
0. https://archive.org/details/isbn_9780062515872/mode/2up
headcanon
That was a human-generated hallucination, my apologies. I always associated semantic web with something Google was pushing to assist with web crawling, and my first exposure to it was during the Web 2.0 era (early 2010s) as HTML5 was seeing adoption, and I always associated it with Google trying to enhance the web as the application platform of the future.
W3C of course deserves credit for their hard work on this standard.
My main point was that regardless of the semantic "standard", nothing prevented us from putting everything in a generic div, so complaining that everyone's just "not on board" isn't a useful lament.
tjr
Some ideas going back even further than that, like 1994:
rco8786
The phrase “if everyone just” is an automatic trigger for me. Everyone is never going to just. A different solution to whatever the problem is will be necessary.
fouc
"Semantic web was a standard suggested by Google", sorry that's false. They only contributed a bit towards it.
Tim Berners-Lee coined it in 1999 and further expanded on the concept in a 2001 Scientific American article by Berners-Lee, Hendler, and Lassila.
discostrings
I think you're confusing XHTML and semantic web on the "break web pages" part.
justincormack
The semantic web came out of work on Prolog and formal systems for AI which just didnt work well... LLMs and vector databases give us new tools that are pretty usable.
ajkjk
eh I feel this but it's a lot simpler than that. Not "if everyone built everything correctly" but "if everyone's work was even slightly better than complete garbage". I do not see many examples of companies building things that are not complete embarrassing shit. I worked at some companies and the things we built was complete embarrassing shit. The reasons are obvious: nobody cares internally to do it, and nobody externally has any standards, and the money still flows if you do a bad job so why do better?
What happens in practice is that the culture exterminates the drive for improvement: not only are things bad, but people look at you if you're crazy if you think things should be better. Like in 2025 people defend C, people defend Javascript, people write software without types, people write scripts in shell languages; debugging sometimes involves looking at actual bytes with your eyes; UIs are written in non-cross-platform ways; the same stupid software gets written over and over at a million companies, sending a large file to another person is still actually pretty hard; leaving comments on it is functionally impossible ... these are software problems, everything is shit, everything can be improved on, nothing should be hard anymore but everything still is; we are still missing a million abstractions that are necessary to make the world simple. Good lord, yesterday I spent two hours trying to resize a PDF. We truly live in the stone age; the only progress we've made is that there are now ads on every rock.
I really wish it was a a much more ruthlessly competitive landscape. One in which if your software is bad, slow, hard to debug, hard to extend, not open source, not modernized, not built on the right abstractions, hard to migrate on or off of, not receptive to feedback, covered in ads... you'd be brutally murdered by the competition in short order. Not like today where you can just lie on your marketing materials and nobody can really do anything because the competition is just as weak. People would do a much better job if they had to to survive.
laserlight
> the money still flows if you do a bad job so why do better?
I'll raise. The money flows because you do a bad job. Doing a good job is costly and takes time. The money cannot invest that much time and resources. Investing time and resources builds an ordinary business. The money is in for the casino effect, for the bangs. Pull the arm and see if it sticks. If yes, good. Keep pulling the arm. If not, continue with another slot machine.
Quarrelsome
I would argue that the money is short termism though. It just assumes short term returns are the correct choice because it lacks the technical understanding of the long term benefits of a good job.
In my experience many acquisitions set the peak of a given software product. The money then makes the argument that its "good enough" and flogs the horse until its dead and a younger more agile team of developers eventually build a new product that makes it irrelevant. The only explanation I have for why so many applications fail to adapt is because of a cultural issue between the software and the money, that always gets resolved by the money winning.
For example I would suggest that the vast majority of desktop apps, especially those made by SMEs, originally in MFC or something fail to make the transition to online services that they need today because of this conflict. The company ends up dying and the money never works out what it did wrong because its much harder to appreciate those long term effects than the short term ones that gave them more money at the start.
Quarrelsome
we have to accept that the vast majority of people don't think like us. They don't think its complete garbage because they can't look hard enough to appreciate that fidelity.
While it might be better if everyone thought like us and wanted things to be _fundamentally_ good, most people don't, and most people money >> less people money and the difference in scale is vast. We could try to build a little fief where we get to set the rules but network effects are against us. If anything our best shot is to try to reverse the centralisation of the internet because that's a big cause of enshittification.
BobaFloutist
Wait, which is the correct programming language to defend? C and Javascript are on pretty opposite sides of most spectra....
oivey
He seems to have mistaken his personal opinions on which languages and language features are good for some objective truth.
Ironically, that’s part of why we can’t have nice things. People who aren’t open to other viewpoints and refuse to compromise when possible impede progress.
jeltz
Languages which are not hard to use and dangerous. I work in C professionally right now but would much rather work in either Zig or Rust.
xigoi
They are on the same side of the “number of serious design flaws due to historical incidents” spectrum.
ajkjk
the new ones we should have made by now that fix their problems...
Zig and Rust and TS are a start
tempfile
Programming languages are not on a spectrum. I guess they dislike the syntax of JS and the low-level no-batteries of C.
dmurray
Presumably the spectrum that would have Rust and Typescript at the other end.
MarkusQ
Fresh plums right off the tree taste significantly better than the ones you can get in the produce isle, which are in turn better than canned, which are themselves still better than re-hydrated prunes.
In scaling out computation to the masses, we went from locally grown plums that took a lot of work and were only available to a small number of people that had a plum tree or knew someone that had one, to building near magical prune-cornucopia devices that everyone could carry around in their pockets, giving them an effectively unlimited supply of prunes.
LLMs re-hydrate these for us, making them significantly more palatable; if you're used to gnawing dried fruit, they seem amazing.
But there's still a lot of work to be done.
random3
Perhaps, but we still failed and not at personal computing, nor just semantic web, but computing and programming in general. The failure is between the original intent (computing was originally more or less AI) along with theory and actual result with every software project turning into unsustainable infinite regress. Things likely broke around ALGOL.
Also LLMs are failing too, for different reasons, but IMO unlikely AI in general will— it will correct a 60 years or so failure in industrial computer science.
heresie-dabord
> we still failed [at] semantic web
The most reprehensible knowledge-search-and-communication failure of all.
We gave people monetisation of drek instead. Then asked them to use it for education. Then trained bot systems on it. Then said that even those answers must be made to confirm to transient propagandists.
sudohalt
The analogy doesn't make any sense, computers today are better by any conceivable metric than computers before.
null
askl
Or slop with some plum aroma added. Seems like a good analogy.
Fomite
Extruded synthetic plum substrate
echelon
I really don't like this analogy, and I really don't like the premise of this article.
Writing software is only so scalable. It doesn't matter all of the shortcuts we take, like Electron and JavaScript. There are only so many engineers with so much time, and there are abundantly many problems to solve.
A better analogy would be to look at what's happening to AI images and video. Those have 10,000x'd the fundamental cost savings, time savings, and personnel requirements. It's an industrialization moment. As a filmmaker who has made several photons-on-glass films, this is a game changer and lifts the entire creative industry to a level where individuals can outdo Pixar.
That is the lens with which to look at what AI will do to software. We're going from hand-carving stone wheels to the Model T.
This is all just getting started. We've barely used what the models of today offer us.
MarkusQ
Totally agree with the core of your position. But the two perspectives are complementary, and perhaps even more deeply linked.
Initial attempts to alleviate any shortage are likely to result in a decrease in quality; initial attempts to improve quality are likely to reduce variability and thus variety. My point (and my reading of the article) is that we're at the stage where we've figured out how to turn out enough Hostess Twinkies that "let them eat cake!" is a viable option, and starvation is being driven back.
This is definite progress, but also highlights previous failures and future work.
armitron
This is a massive cope. AI image/video slop is still slop. Yes it's getting better, but it's still better .. slop. Unless radical new breakthroughs are made, the current LLM paradigm will not outdo Pixar or any other apex of human creativity. It'll always be instantly recognizable, as slop.
And if we allow it to take over society, we'll end up with a society that's also slop. Netflixification/marvelization only much much worse..
echelon
> This is a massive cope. AI image/video slop is still slop.
Slop content is slop content, AI or not. You don't need an AI to make slop, it just makes it more accessible.
You are biased by media narratives and slop content you're encountering on social media. I work in the industry and professionals are using AI models in ways you aren't even aware of. I guarantee you can't identify all AI content.
> And if we allow it to take over society, we'll end up with a society that's also slop. Netflixification/marvelization only much much worse..
Auteurs and artists aren't going anywhere. These tools enable the 50,000 annual film students to sustainably find autonomy, where previously there wasn't any.
bkummel
Yeah dream on. I’m an engineer and know what structured data is. And yet I miserably fail to store my private files in a way that I can find them back without relying on search tools. So how on earth are we ever going to organize all the world’s data and knowledge? Thank god we found this sub-optimal “band aid” called LLMs!
yeyeyeyeyeyeyee
Librarians have succeeded in precisely this for a long time now.
ctoth
Precisely this. This article might seem reasonable to anybody who has never tried to organize something as simple as a local music collection.
pjm331
Made me think about John Wilkins' "philosophical language" which I first heard about in Neal Stephenson's book Quicksilver
https://en.wikipedia.org/wiki/An_Essay_Towards_a_Real_Charac...
I'm sure there have been countless similar attempts at categorizing knowledge
one of the more successful ones being the dewey decimal system
I have my doubts about whether the thing the OP alleges we have "failed" at is even possible at all
whartung
Well, this runs straight into one of the massive, concrete pillars of computing: naming things.
Because that’s what a lot of this falls into.
Overwhelming amount of stuff with no names. No categories, no nothing.
With extended file attributes we could hang all sorts of meta bits off of arbitrary files. But that’s very fragile.
So we ask the systems to make up names for data based on their content, which turns out to not necessarily work as well as we might like.
jandrese
No names is not the biggest problem. You just have to come up with a name. The problem is when things have multiple names, or even worse when people disagree on what names are appropriate for something. The world rarely allows you to neatly categorize large datasets. There are always outliers.
For example, you have a set of balls and you want to sort them by color. Where does orange stop and red begin? What about striped balls or ones with logos printed on them? What if it is a hypercolor ball that changes based on heat? It gets messy very fast.
Kapura
I have often thought about how computers are significantly faster than they were in the early 2000s, but they are significantly harder to use. Using Linux for the first time in college was a revelation, because it gave me the tools to tell the computer "rename all of the files in this directory, keeping only the important parts of the name."
But instead of iterating on better interfaces to effectively utilize the N thousands of operations per second a computer is capable of, the powers that be behind the industry have decided to invest billions of dollars in GPUs to get a program that seems like it understands language, but is incapable of counting the number of B's in "blueberry."
svachalek
Prompt: "Spell blueberry and count the letter b".
They're not claiming AGI yet, so human intelligence is required to operate an LLM optimally. It's well known that LLMs process tokens rather than characters s, so without space for "reasoning" there's no representation of the letter b in the prompt. Telling it to spell or think about it gives it room to spell it out, and from there it can "see" the letters and it's trivial to count.
xandrius
Is counting the number of B's vital? Also, I'm pretty sure you can get an LLM to parse text the way you want it, it just doesn't see your text as you do, so that simple operation is not straightforward. Similarly, are you worthless because you seem like you understand language but are incapable of counting the number of octects in "blueberry"?
dreamcompiler
Let's say I hire a plumber because of his plumbing expertise and he bills me $35 and I pay him with a $50 bill and he gives me back $10 in change. He insists he's right about this.
I am now completely justified in worrying about whether the pipes he just installed were actually made of butter.
xandrius
Really? Is that easy? Happens quite often to really believe something and be wrong. Maybe you both are right and the $5 bill is on the floor?
copularent
As shown by the GPT-5 reaction, a majority of people just have nothing better to ask the models than how many times does the letter "s" appear in "stupid".
jandrese
I think this is a completely valid thing to do when you have Sam Altman going on the daily shows and describing it as a genius in your pocket and how it's smarter than any human alive. Deflating hype bubbles is an important service.
tekno45
if i have to talk to it a specific way, why not just use programming. The specific way we talk to computers effectively...
jazzypants
I mean, I think that anyone who understands UTF-8 will know that there are nine octets in blueberry when it is written on a web page. If you wanted to be tricky, you could have thrown a Β in there or something.
xandrius
> anyone who understands UTF-8
So not too many?
squigz
> Similarly, are you worthless because you seem like you understand language but are incapable of counting the number of octects in "blueberry"?
Well, I would say that if GP advertised themselves as being able to do so, and confidently gave an incorrect answer, their function as someone who is able to serve their advertised purpose is practically useless.
xandrius
So ChatGPT was advertised as a letter counter?
Also, no matter what hype or marketing says: GPT is a statistical word bag with a mostly invisible middleman to give it a bias.
A car is a private transportation vehicle but companies still try to sell it as a lifestyle choice. It's still a car.
SV_BubbleTime
IDK, I think there is something adorable about taking a system that over trillions of iterations always performs the same operation with the same result, reliability unmatched in all of the universe…
And making it more of “IDK what it answered the way it did, but it might be right!!”
codr7
Humans like games :)
grumbel
Cyc[0] tried for 40 years to make handwritten semantic rules a thing, but still don't have much to show for. Humans just aren't good or fast at writing down the rather fuzzy semantics of the real world into a computer readable format.
With RDF specially, there was also the issue of "WTF is this good for?". Semantic Web sounds lofty in theory, but was there ever even a clear plan on how the UI would look like? How would I explore all that semantic data if it ever came into existence? How would it deal with link rot?
And much like with RSS, I think a big failure of RDF is that it's some weird thing outside the Web, instead of just some addition HTML tags to enrich existing documents. If there is a failure, it's that. Even today, a lot of basic semantic tags are missing from HTML, we finally got <date> in 2011, but we still have nothing for names, cities, units, books[1], movie, gps coordinates and a lot of other stuff we use daily.
Another big failure is that HTML has become a read-only format, the idea that one uses HTML as a source format to publish documents seems to have been completely abandoned. HTML is just a UI language for Web apps nowadays, Markdown, LaTeX or whatever is what one uses to write content.
0. https://en.wikipedia.org/wiki/Cyc
1. <a href="urn:isbn:..."> exists, but browsers don't support it natively
seanwilson
> What animal is featured on a flag of a country where the first small British colony was established in the same year that Sweden's King Gustav IV Adolf declared war on France? ... My point is that if all knowledge were stored in a structured way with rich semantic linking, then very primitive natural language processing algorithms could parse question like the example at the beginning of the article, and could find the answer using orders of magnitude fewer computational resources.
So as well as people writing posts in English, they would need to provide semantic markup for all the information like dates, flags, animals, people, and countries? It's difficult enough to get people to use basic HTML tags and accessible markup properly, so what was the plan for how this would scale?
TZubiri
So wikipedia and wikidata?
This actually happened already and it's part of why llms are so smart, I haven't tested this but I venture a guess that without wikipedia and wikidata and wikipedia clones and stolen articles, LLMs would be quite a lot dumber. You can only get so far with reddit articles and embedded knowledge of basic info on higher order articles.
My guess is when fine tuning and modifying weights, the lowest hanging fruit is to overweigh wikipedia sources and reduce the weight of sources like reddit.
foundart
Expresses a longing for the semantic web.
> Remember Semantic Web? The web was supposed to evolve into semantically structured, linked, machine-readable data that would enable amazing opportunities. That never happened.
I think the lesson to be learned is in answering the question "Why didn't the semantic web happen?"
jandrese
The semantic web was theoretically great for data scientists and metadata scrapers, but offered close to zero value for ordinary humans, both on the publishing side an the consumption side. Also, nobody did the hard work of defining all of the categories and protocols in a way that was actually usable.
The whole concept was too high minded and they never got the implementation details down. Even if they did it would have been horrendously complex and close to impossible to manage. Asking every single publisher to neatly categories their data into this necessarily enormous scheme would have resulted in countless errors all over the web that would have seriously undercut the utility of the project anyway. Ultimately the semantic web doesn't scale very well. It failed for the same reason command economies fail: It's too overwhelming for the people in control to manage and drowns in its own bureaucracy.
cognivore
"Why didn't the semantic web happen?"
I have literally been doing we development since their was a web, and the companies I developed for are openly hostile to the idea of putting their valuable, or perceived valuable, information online in a format that could be easily scraped. Information doesn't want to be free, it wants to be paid for. Unless the information shared pulls visitors to the site it doesn't need to be public.
gvurrdon
I've encountered a similar issue in academia - PI's don't want to make their data available to be scraped (or, at least not easily) because the amount of grant funding is limited, and a rival who has scraped one's data could get the grant money instead by using that scraped data to bolster their application.
thewebguyd
> Information doesn't want to be free, it wants to be paid for. Unless the information shared pulls visitors to the site it doesn't need to be public.
That's a cultural and societal problem, not a technology problem. The motivations (profit) are wrong, and don't lead to true innovations, only to financialization.
So long as people need to pay to eat, then information will also want to continue to be paid for, and our motivations will continue to be misaligned with true innovations, especially if said innovations would make life easier but wouldn't result in profit.
llbbdd
You need profit or you need post-scarcity or nothing works at all
HankB99
I was thinking of that in terms of siloed web sites but your description of walling off information is broader and more appropriate.
neilalexander
> "Why didn't the semantic web happen?"
Advertising.
throwaway346434
To a degree re ads on pages, but why didn't big business end up publishing all of their products in JSON-LD or similar? A lot did, to get aggregated, but not all.
Workaccount2
>"Why didn't the semantic web happen?"
Because web content is generated by humans, not engineers.
abeppu
But also because companies that produce web content wanted it to be seen by humans who would look at ads, not consumed by bots and synthesized with other info into a product owned by some other firm.
dreamcompiler
And yet today most websites are being scraped by LLM bots which don't look at ads and which synthesize with other info into a product owned by some other firm.
Mikhail_Edoshin
Because you cannot build it with merchants. This is a job for monks.
seydor
because semantic web was more limited than language
AlienRobot
Semantic web never existed. There was Google and Google had an API to get breadcrumbs to show on search results. And that's what people called "semantic web." A few years later they gave up and made everything look like a breadcrumb anyway. And that sums up the whole semantic web experience.
dawnofdusk
Basically the same as decrying why we should have to learn foreign languages instead of everyone speaking Esperanto.
open_
A pretty reductionist and a poor take.
"Standing on the shoulders of giants, it is clear that the giants failed to reach the heights we have reached."
atrettel
I find traditional web search and LLM search to be complementary technologies, and this is a good example of that. Both have their uses and if you get the information you need using one or the other, we are all happy.
I think the example query here actually shows a problem with the query languages used in web search rather than an intrinsic inability of web search. It contains what amounts to a natural language subquery starting with "in the same year". In other words, to evaluate this query properly, we need to first evaluate this subquery and then use that information to evaluate the overall query. Google Search and almost all other traditional web search engines use intentionally oversimplified query languages that disallow nested queries let alone subqueries, so this example really is just revealing a problem with the query language rather than a problem with web search overall. With a better query language, we might get better results.
taeric
Leading in with what feels like a p-hacked question that almost certainly isn't typical kind of hurts the point. Reminds me of seeing clearly rehearsed demos for things that ignore the reason it worked is because it was rehearsed. A lot.
Am I wrong in that this was a completely organic question?
jkaptur
"... if all knowledge were stored in a structured way with rich semantic linking..." this sounds a lot like Google's "Knowledge Graph". https://developers.google.com/knowledge-graph. (Disclosure: I work at Google.)
If you ask an LLM where you can find a structured database of knowledge with structured semantic links, they'll point you to this and other knowledge graphs. TIL about Diffbot!
In my experience, it's a lot more fun to imagine the perfect database like this than it is to work with the actual ones people have built.
I feel like I see this attitude a lot amongst devs: "If everyone just built it correctly, we wouldn't need these bandaids"
To me, it feels similar to "If everyone just cooperated perfectly and helped each other out, we wouldn't need laws/money/government/religion/etc."
Yes, you're probably right, but no that won't happen the way you want to, because we are part of a complex system, and everyone has their very different incentives.
Semantic web was a standard suggested by Google, but unless every browser got on board to break web pages that didn't conform to that standard, then people aren't going to fully follow it. Instead, browsers (correctly in my view) decided to be as flexible as possible to render pages in a best-effort way, because everyone had a slightly different way to build web pages.
I feel like people get too stuck on the "correct" way to do things, but the reality of computers, as is the reality of everything, is that there are lots of different ways to do things, and we need to have systems that are comfortable with handling that.