Obituary for Cyc
302 comments
·April 8, 2025ChuckMcM
musicale
> there remains no evidence of its general intelligence
This seems like a high bar to reach.
We all know that symbolic AI didn't scale as well as LLMs trained on huge amounts of data. However, as you note, it also tried to address many things that LLMs still don't do well.
ChuckMcM
This is exactly correct, LLMs did scale with huge data, symbolic AI did not. So why? One of the things I periodically ask people working on LLMs is "what does a 'parameter' represent? The simplistic answer is 'it's a weight in a neural net node' but that doesn't much closer. Consider something like a bloom filter where a '0' bit represents the nth bit of all hashes of strings this filter has not seen. I would be interested in reading a paper that does a good job of explaining what a parameter ends up representing in an LLM model.[1]
I suspect that McCarthy was on to something with the context thing. Organic intelligence certainly fails in creative ways without context it would not be disqualifying to have AI fail in similarly spectacular ways.
[1] I made a bit of progress on this considering it to be the permeability for progress such that the higher the weight the easier it was to 'pass thorough' this particular neuron but the cyclic nature of the graph makes a purely topological explanation pretty obtuse :-).
thesz
> LLMs did scale with huge data, symbolic AI did not.
Symbolic AI have not had a privilege to be applied or "trained" with huge data. 30 millions assertions is not a big number.
joe_the_user
LLMs did scale with huge data, symbolic AI did not. So why? [1]
Neural networks, not LLMs in particular, were just about the simplest thing that could scale - they scaled and everything else has been fine-tuning. Symbolic AI basically begins with existing mathematical models of reality and of human reason and indeed didn't scale.
The problem imo is: The standard way mathematical modeling works[2] is you have a triple of <data, model-of-data, math-formalism>. The math formalism characterizes what the data could be, how data diverges from reality etc. The trouble is that the math formalism really doesn't scale even if a given model scales[3]. So even if you were to start plugging numbers into some other math model and get a reality-approximation like an LLM, it would be a black box like an LLM because the meta-information would be just as opaque.
Consider the way Judea Pearl rejected confidence intervals and claimed probabilities were needed as the building blocks for approximate reasoning systems. But a look at human beings, animals or LLMs shows that things that "deal with reality" don't have and couldn't access to "real" probabilities.
I'd just offer that I believe that for a model to scale, the vast majority of it's parameters would have to be mathematically meaningless to us. And that's for the above reasons.
[1]. Really key point, imo [2]. That innclude symbolic and probabilistic model "at the end of the day" [3]. Contrast the simplicity of plugging data into a regression model versus the multitudes of approaches explaining regression and loss/error functions etc.
krackers
>I would be interested in reading a paper that does a good job of explaining what a parameter ends up representing in an LLM model.
https://distill.pub/2020/circuits/ https://transformer-circuits.pub/2025/attribution-graphs/bio...
YeGoblynQueenne
>> This is exactly correct, LLMs did scale with huge data, symbolic AI did not. So why?
Like the rock salad you're mixing up two disparate contexts here. Symbolic AI like SAT solvers and planners is not trying to learn from data and there's no context in which it has to "scale with huge data".
Instead, what modern SAT solvers and planners do is even harder than "scaling with data" - which, after all, today means having imba hardware and using it well. SAT solving and planning can't do that: SAT is NP-complete and planning is PSPACE-complete so it really doesn't matter how much you "scale" your hardware, those are not problems you can solve by scaling, ever.
And yet, today both SAT and planning are solved problems. NP complete? Nowadays, that's a piece of cake. There are dedicated solvers for all the classical sub-categories of SAT and modern planners can solve planning problems that require sequences of thousands of actions. Hell, modern planners can even play Atari games from pixels alone, and do very well indeed [1].
So how did symbolic AI manage those feats? Not with bigger computers but precisely with the approach that the article above seems to think has failed to produce any results: heuristic search. In SAT solving, the dominant approach is an algorithm called "Conflict Driven Clause Learning", that is designed to exploit the special structure of SAT problems. In Planning and Scheduling, heuristic search was always used, but work really took off in the '90s when people realised that they could automatically estimate a heuristic cost function from the structure of a planning problem.
There are parallel and similar approaches everywhere you look at, in classical AI problems, like verification, theorem proving, etc, and that work has even produced a few Turing awards [2]. But do you hear about that work at all, when you hear about AI research? No, because it works, and so it's not AI.
But it works, it runs on normal hardware, it doesn't need "scale" and it doesn't need data. You're measuring the wrong thing with the wrong stick.
____________
[1] Planning with Pixels in (Almost) Real Time: https://arxiv.org/pdf/1801.03354 Competitive results with humans and RL. Bet you didn't know that.
[2] E.g. Pnueli for temporal logic in verification, or Clarke, Emerson and Sifakis, for model checking.
otabdeveloper4
Well, we haven't tried symbolic AI with huge amounts of data. It's a hard problem.
(And ironically this problem is much easier now that we have LLMs to help us clean and massage textual data.)
adastra22
Such as what? What can GOFAI do well that LLMs still cannot?
sgt101
I think logical reasoning - so reasoning about logical problems, especially those with transitive relations like two way implication. A way round this is to get them to write prolog relations and then reason over them... with prolog. This isn't a fail - it's what things like prolog do, and not what things like nns do. If I was asked to solve these problems I would write prolog too.
I think quite a lot of planning.
I think scheduling - I tried something recently and GPT4 wrote python code which worked for very naive cases but then failed at any scale.
Basically though - trusted reasoning. Where you need a precise and correct answer LLM's aren't any good. They fail in the limit. But where you need a generally decent answer they are amazing. You just can't rely on it.
Whereas GOFAI you can, because if you couldn't the community thew it out and said it was impossible!
musicale
"Tried to address" is not the same as "can do well."
I was responding to PP, but some other (maybe obvious?) examples are logical reasoning and explainability.
As PP suggests, some of the classical symbolic ideas may be applicable or complementary to current approaches.
YeGoblynQueenne
SAT solving, verification and model checking, automated theorem proving, planning and scheculing, knowledge representation and reasoning. Those are fields of AI research where LLMs have nothing to offer.
mepian
Theorem proving.
fancyfredbot
What a fantastic article. One of the best on HN this year.
I first heard about Cyc as a child in a TV documentary about AI. The example involved a man shaving with an electric razor. Cyc concluded that while shaving the man was not human since the razor was an electrical device and electrical devices were not human. It really caught my imagination as a child and made me want to study AI. The magic diminished a bit once I learned more about how Cyc worked using prolog style relations and I ended up studying CS instead of AI but I still credit Cyc with sparking that initial interest in me.
Lenart seems like a strange man but we need obsessives and true believers like him to push through the various "winters". Who knows if knowledge graphs like Cyc will become relevant again in future as we seek to eliminate hallucinations from statistical learning.
codr7
I remember the very same example (re)-sparking my interest in AI back in the days.
My gut feeling says there is something in this approach that's needed to make GenAI work reliably. The brain has an associative feature, sure; but it's not very useful without filters sorting signal from nonsense, making sense of the content.
Have they been able to get Cyc to generate its own content in meaningful ways? I would expect such a system to eventually be able to derive a lot of details by itself, needing less and less spoon feeding.
cubefox
> Cyc concluded that while shaving the man was not human since the razor was an electrical device and electrical devices were not human.
I can't parse this sentence? Is there supposed to be a comma before and after "while shaving"?
fancyfredbot
Yes. Sorry. I was actually just googling this and realised this same anecdote is in cited the intro to the Deep Learning book by Goodfellow et al. Their write-up is hopefully clearer:
"For example, Cyc failed to understand a story about a person named Fred shaving in the morning (Linde, 1992). Its inference engine detected an inconsistency in the story: it knew that people do not have electrical parts, but because Fred was holding an electric razor, it believed the entity “FredWhileShaving” contained electrical parts. It therefore asked whether Fred was still a person while he was shaving"
https://www.deeplearningbook.org/contents/intro.html
The (Linde, 1992) citation is they give is the 4th episode of a TV series - presumably the one I saw as a kid!
https://en.m.wikipedia.org/wiki/The_Machine_That_Changed_the...
And of course it's on YouTube:
https://youtube.com/clip/UgkxRcsHT-s1iZ-VRWFRXA-qg4kjTYe-a6j...
cubefox
To be fair, the object "FredWhileShaving" is a state (of Fred), or simply an event, and as such it arguably is not a person in the first place. As persons are presumably neither states nor events.
pea
Deleuze would argue something similar
null
mac3n
along this line, “Gravity has no friends.”
pmarreck
> seems like a strange man
seemed.
Unfortunately, he passed away a couple of years ago (which I wish I had known before now!)
musicale
You can still read his papers and textbooks. I thought he was an engaging speaker as well, and several of his lectures and interviews are available on youtube, as well as tributes and retrospectives.
Rochus
A very interesting and worthwhile article (better than the comments here would suggest). However, I find it a bit of a pity that the author places so much emphasis on the assumption that the project has failed. The approach has not simply failed because the search for a solution has been going on for forty years. It took even more than forty years and costed at least as much before neural networks became really useful, and nobody would claim that the project failed because of that. And today's LLM are not really that intelligent yet. Maybe Cycorp's knowledge base will be made generally accessible at some point, so that it can be used to train LLMs. Perhaps then a greater benefit of this data will become apparent.
zozbot234
> Maybe Cycorp's knowledge base will be made generally accessible at some point, so that it can be used to train LLMs.
More likely, it will be made increasingly irrelevant as open alternatives to it are developed instead. The Wikipedia folks are working on some sort of openly developed interlingua that can be edited by humans, in order to populate Wikipedias in underrepresented languages with basic encyclopedic text. (Details very much TBD, but see https://en.wikipedia.org/wiki/Abstract_Wikipedia and https://meta.wikimedia.org/wiki/Abstract_Wikipedia ) This will probably be roughly as powerful as the system OP posits at some point in the article, that can generate text in both English and Japanese but only if fed with the right "common sense" to begin with. It's not clear exactly how useful logical inference on such statements might turn out to be, but the potential will definitely exist for something like that too, if it's found to be genuinely worthwhile in some way.
Rochus
> made increasingly irrelevant as open alternatives to it are developed instead
Certainly interesting what these projects are going for, but it's unlikely an "open alternative", given that the degree of formalization and rigor achieved by Cyc's higher-order logic specification is likely not achievable by statistical learning, and a symbolic approach is barely achievable in a shorter time than Cyc.
yowzadave
It would be very surprising if the results from this approach were superior to simply machine-translating the entries from another language—because e.g. English already has so much content and contributor activity, and LLMs are already very good at translating. I can’t imagine you’d get more than a fraction of people’s interest in authoring entries in this abstract language.
yorwba
LLMs are good at translating between languages that have significant amounts of written content on the internet. There are few languages in this category that do not already have correspondingly large Wikipedias.
There are plenty of languages with millions of speakers that are only rarely used in writing, often because some other language is enforced in education. If you try to use an LLM to translate into such a language, you'll just get garbage.
It's very easy for a hand-crafted template to beat an LLM if the LLM can't do the job at all.
yellowapple
https://www.wikidata.org/wiki/Wikidata:Main_Page, for those curious about the interlingua in question.
zozbot234
Strictly speaking, Wikidata is an existing project which only provides a rather restrictive model for its assertions; they are not fully compositional, thus are quite far from being able to express general encyclopedic text, especially in a way that can be 'seamlessly' translated to natural language. It does provide a likely foundation for these further planned developments, though.
smoyer
> on the assumption the project has failed.
My daughter's PhD thesis was largely negative results. Even if the project had failed, we could learn from it if it wasn't so secretive. It could be much more open without being OSS!
Rochus
It is at least as important to know which approaches do not work, but this gets significantly less press, which is not that attractive for scientists in the age of "publish or perish".
YeGoblynQueenne
>> My daughter's PhD thesis was largely negative results.
Well, give us a link man! :)
smoyer
You're forcing the proud dad function: https://pubmed.ncbi.nlm.nih.gov/36995257/
Paracompact
> Maybe Cycorp's knowledge base will be made generally accessible at some point
I would sooner hold my breath waiting for OpenAI to open up than Cycorp :)
> It took even more than forty years and costed at least as much before neural networks became really useful
The correct class of comparison to make with "neural networks" would be "symbolic AI" writ large. Symbolic AIs have been working quite well in some areas. Just not at all in terms of common sense reasoning, or anything approaching AGI.
If you want to keep "Cyc" in the comparison, then I would argue there is no comparison possible. Without exaggeration, there has never been a single project in AI as expensive as Cyc before 2020. Only with GPT-2 did the cost start to exceed the million USD mark. (Without exact figures, AlphaGo and Deep Blue probably also cost millions of dollars, but they unambiguously worked.)
It's also just not true that it took 40 years. Consider e.g. LeNet-5, which was up and running in 1998, and was used in ATMs to read real cheques. The main cost was 10 years of engineering stamina by LeCun's research group at Bell Labs. The finished version could be trained "for about 20 epoches over MNIST. It took 2 to 3 days of CPU time on a Silicon Graphics Origin 2000 server, using a single 200 MHz R10000 processor."
(1998 might technically be 40 years out from e.g. the inception of the perceptron in the 1950s, but if that is supposed to be our reference point for neural networks, then Cyc's reference point should be the inception of logical AIs in the same decade. And really, what use was Cyc in industry in 1998?)
Animats
Cyc is going great, according to the web site. "The Next Generation of Enterprise AI"[1]
Lenat himself died in 2023. Despite this, he is listed as the only member of the "leadership team".[2]
[1] https://cyc.com/
curiousObject
A sad epitaph
I’m very glad he tried, even if he only proved that it was impossible
vitiral
Maybe Cyc was a success and Lenat lives on as it's consciousness?
jibal
Dead is dead. And even if Cyc had consciousness--which it doesn't--it certainly wouldn't have his consciousness.
Cthulhu_
It's an interesting thought experiment / philosophy / sci fi story premise though; if he spent all those years encoding his own thought processes and decision making into a program, would he have been able to create a convincing facsimile of himself and his "consciousness"? A turing test with a chatbot based on decades of self-reflection.
zitterbewegung
You can run a version of CYC that was released online as opencyc https://github.com/asanchez75/opencyc . This is when a version of the system was posted on source forge and the GitHub has the dataset and the KB and inference engine. Note it has been written in an old version of Java.
drob518
Many years ago, I went on a quest to read as much about AM and EURISKO as I could fine. Unfortunately, Lenat kept all of his work very private (secret, even), and there was very little to read other than high level papers and articles. I always felt that was a shame, as it prevented anyone else from building on anything Lenat built. I think we’re poorer for it.
mark_l_watson
The article is a very good review of Symbolic AI, in general, not just Cyc.
I have spent a lot of time with OpenCyc in the past, but haven’t touched it in ten years.
I believe tif there is a productive future for symbolic AI that it will involve using LLMs to construct knowledge graphs, symbolic relatikns, etc. from unstructured data.
YeGoblynQueenne
>> The article is a very good review of Symbolic AI, in general, not just Cyc.
As a "review of symbolic AI" I found it uninformed and superficial and felt that it rehashed the same old points about how symbolic AI "failed", which are disputed by the facts; specifically the fact that major symbolic AI fields like SAT solving, automated theorem proving and planning and scheduling are still going strong and have produced real-world results, so much so that e.g. SAT solving, Planning, program verification, and automated theorem proving aren't even considered "AI" anymore because they now actually work, and work very well indeed.
mark_l_watson
Technically, you are probably correct. I did find it a good walk ‘down memory lane.’ I have been working in the field since 1982 and the article made me nostalgic.
trhway
With LLM itself using those graphs to validate its answers, etc. And at some point it will be internalized into architecture as a graph attention layer.
vannevar
I would argue that Lenat was at least directionally correct in understanding that sheer volume of data (in Cyc's case, rules and facts) was the key in eventually achieving useful intelligence. I have to confess that I once criticized the Cyc project for creating an ever-larger pile of sh*t and expecting a pony to emerge, but that's sort of what has happened with LLMs.
cmrdporcupine
I suspect at some point the pendulum will again swing back the other way and symbolic approaches will have some kind of breakthrough and become trendy again. And, I bet it will likely have something to do with accelerating these systems with hardware, much like GPUs have done for neural networks, in order to crunch really large quantities of facts
luma
The Bitter Lesson has a few things to say about this.
wzdd
The Bitter Lesson says "general methods that leverage computation are ultimately the most effective". That doesn't seem to rule out symbolic approaches. It does rule out anything which relies on having humans in the loop, because terabytes of data plus a dumb learning process works better than megabytes of data plus expert instruction.
(I know your message wasn't claiming that The Bitter Lesson was explicitly a counterpoint, I just thought it was interesting.)
null
bcoates
Imho, this is wrong. Even independent of access to vast amounts of compute, symbolic methods seem to consistently underperform statistical/numerical ones across a wide variety of domains. I can't help but think that there's more to it than just brute force.
kevin_thibedeau
Real AGI will need a way to reason about factual knowledge. An ontology is a useful framework for establishing facts without inferring them from messy human language.
null
null
IshKebab
These guys are trying to combine symbolic reasoning with LLMs somehow: https://www.symbolica.ai/
specialgoodness
check out Imandra's platform for neurosymbolic AI - https://www.imandra.ai/
whiplash451
Or maybe program synthesis combined by LLMs might be the way?
cmrdporcupine
It does seem like the Cyc people hit the wall with simply collecting facts. Having to have a human in the loop.
The problem I think is if you have LLMs figuring out the propositions, the whole system is just as prone to garbage-in-garbage-out as LLMs are.
jibal
But
a) The pile of LLM training data is vastly larger. b) The data is actual human utterances in situ--these are ponies, not pony shit. c) LLMs have no intelligence ... they channel the intelligence of a vast number of humans by pattern matching their utterances to a query. This has indeed proved useful because of how extremely well the statistical apparatus works, but the fact that LLMs have no cognitive states puts great limits on what this technology can achieve.
With Cyc, OTOH, it's not even clear what you can get out of it. The thing may well be useful if combined with LLMs, but it's under lock and key.
The big conclusions about symbolic AI that the author reaches based on this one system and approach are unwarranted. As he himself notes, "Even Ernest Davis and Gary Marcus, highly sympathetic to the symbolic approach to AI, found little evidence for the success of Cyc, not because Cyc had provably failed, but simply because there was too little evidence in any direction, success or failure."
YeGoblynQueenne
>> they channel the intelligence of a vast number of humans by pattern matching their utterances to a query.
Just a little problem with that: to understand the utterances of a vast number of humans you need to channel it to something that can understand the utterances of humans in the first place. Just channeling it around from statistic to statistic doesn't do the trick.
jibal
Um, the "something" is the person reading the LLM's output. I'm afraid that you have completely missed the context and point of the discussion, which was not about LLMs understanding things--they understand nothing ("LLMs have no cognitive states"). But again, "because of how extremely well the statistical apparatus works", their outputs are useful to intelligent consumers who do have cognitive states--us.
TechDebtDevin
The dataset for lots of LLMs is literally called "The Pile" lmao
chubot
That’s hilarious, but at least Llama was trained on libgen, an archive of most books and publications by humanity, no? Except for the ones which were not digitized I guess
So there is probably a big pile of Reddit comments, twitter messages, and libgen and arxiv PDFs I imagine
So there is some shit, but also painstakingly encoded knowledge (ie writing), and yeah it is miraculous that LLMs are right as often as they are
cratermoon
libgen is far from an archive of "most" books and publications, not even close.
The most recent numbers from libgen itself are 2.4 million non-fiction books and 80 million science journal articles. The Atlantic's database published in 2025 has 7.5 million books.[0] The publishing industry estimates that many books are published each year. As of 2010, Google counted over 129 million books[1]. At best an LLM like Llama will have have 20% of all books in its training set.
0. https://www.theatlantic.com/technology/archive/2025/03/libge...
1. https://booksearch.blogspot.com/2010/08/books-of-world-stand...
UltraSane
On libgen.mx they claim to have 33,569,200 books and 84,844,242 articles
ChadNauseam
It's a miracle, but it's all thanks to the post-training. When you think of it, for so-called "next token predictors", LLMs talk in a way that almost no one actually talks, with perfect spelling and use of punctuation. The post-training somehow is able to get them to predict something along the lines of what a reasonably intelligent assistant with perfect grammar would say. LLMs are probably smarter than is exposed through their chat interface, since it's unlikely the post-training process is able to get them to impersonate the smartest character they'd be capable of impersonating.
chubot
I dunno I actually think say Claude AI SOUNDS smarter than it is, right now
It has a phenomenal recall. I just asked it about "SmartOS", something I knew about, vaguely, in ~2012, and it gave me a pretty darn good answer. On that particular subject, I think it probably gave a better answer than anyone I could e-mail, call, or text right now
It was significantly more informative than wikipedia - https://en.wikipedia.org/wiki/SmartOS
But I still find it easy to stump it and get it to hallucinate, which makes it seem dumb
It is like a person with good manners, and a lot of memory, and which is quite good at comparisons (although you have to verify, which is usually fine)
But I would not say it is "smart" at coming up with new ideas or anything
I do think a key point is that a "text calculator" is doing a lot of work ... i.e. summarization and comparison are extremely useful things. They can accelerate thinking
baq
https://ai-2027.com/ postulates that a good enough LLM will rewrite itself using rules and facts... sci-fi, but so is chatting with a matrix multiplication.
josephg
I doubt it. The human mind is a probabilistic computer, at every level. There’s no set definition for what a chair is. It’s fuzzy. Some things are obviously in the category, and some are at the periphery of it. (Eg is a stool a chair? Is a log next to a campfire a chair? How about a tree stump in the woods? Etc). This kind of fuzzy reasoning is the rule, not the exception when it comes to human intuition.
There’s no way to use “rules and facts” to express concepts like “chair” or “grass”, or “face” or “justice” or really anything. Any project trying to use deterministic symbolic logic to represent the world fundamentally misunderstands cognition.
yellowapple
> There’s no set definition for what a chair is.
Sure there is: a chair is anything upon which I can comfortably sit without breaking it.
veqq
So you're just ignoring all the probabilistic, fuzzy etc. Prologs etc. which do precisely that? https://github.com/lab-v2/pyreason
woodruffw
> Any project trying to use deterministic symbolic logic to represent the world fundamentally misunderstands cognition.
The counterposition to this is no more convincing: cognition is fuzzy, but it's not really clear at all that it's probabilistic: I don't look at a stump and ascertain its chairness with a confidence of 85%, for example. The actual meta-cognition of "can I sit on this thing" is more like "it looks sittable, and I can try to sit on it, but if it feels unstable then I shouldn't sit on it." In other words, a defeasible inference.
(There's an entire branch of symbolic logic that models fuzziness without probability: non-monotonic logic[1]. I don't think these get us to AGI either.)
jgalt212
> The human mind is a probabilistic computer, at every level.
Fair enough, but an airplane's wing is not very similar to a bird's wing.
photonthug
> There’s no way to use “rules and facts” to express concepts like “chair” or “grass”, or “face” or “justice” or really anything. Any project trying to use deterministic symbolic logic to represent the world fundamentally misunderstands cognition.
Are you sure? In terms of theoretical foundations for AGI, AIXI is probabilistic but godel-machines are proof based and I think they'd meet criteria for deterministic / symbolic. Non-monotonic and temporal logics also exist, where chairness exists as a concept that might be revoked if 2 or more legs are missing. If you really want to get technical then by allowing logics with continuous time and changing discrete truth values, then you can probably manufacture a fuzzy logic where time isn't considered but truth/certainty values are continuous. Your ideas about logic might be too simple, it's more than just Aristotle
cyberax
> This kind of fuzzy reasoning is the rule, not the exception when it comes to human intuition.
That is indeed true. But we do have classic fuzzy logic, and it can be used to answer these questions. E.g. a "stool" maybe a "chair", but "automobile" is definitely not.
Maybe the symbolic logic approach could work if it's connected with ML? Maybe we can use a neural network to plot a path in the sea of assertions? Cyc really seems like something that can benefit the world if it's made open under some reasonable conditions.
nickpsecurity
"The human mind is a probabilistic computer, at every level."
We don't know that. It's mostly probabilistic. That innate behavior exists suggests some parts might be deterministic.
cess11
Words are used due to the absence of things. They fill an immediate experiential void and stand in for something else, because you want or need another person to evoke some fantasy to fill this absence and make understanding possible.
If you have a mind and it is a computer, then it is because of nurture, because the brain is nothing like a computer, and computers simulating language are nothing like brains.
mountainriver
How will the rules and facts be connected? By some discrete relationship? This stuff only works for math, and is the basis for the bitter lesson.
Intelligence is compression, and this is the opposite of that
null
null
cubefox
A similar failure of GOFAI was ABBYY's (a Russian company which was for a long time market leader in OCR software) monumental (multi decade long) attempt of creating advanced translation software entirely based on complex formal grammar parsing.
The story behind it is really interesting. This article was written by someone who worked at ABBYY:
https://sysblok.ru/blog/gorkij-urok-abbyy-kak-lingvisty-proi...
The piece is in Russian but can (ironically) be read in good English by using e.g. the Google Translate feature inside Chrome. Which is of course entirely based on machine learning.
The story is essentially similar to Cyc: symbolic AI/logical AI/GOFAI can produce initially impressive results (ABBYY was way better than early forms of Google Translate), but symbolic approaches doesn't scale well. Big Data + machine learning wins out eventually. The piece above mentions a 2009 piece from Google which put forward this thesis. "The Unreasonable Effectiveness of Data":
https://static.googleusercontent.com/media/research.google.c...
Note that 2009 was significantly before the existence of large language models, transformers, or even AlexNet.
pvg
A big Cyc thread about a year ago https://news.ycombinator.com/item?id=40069298
hitekker
A former employee of Cyc did an insightful AMA on HN back in 2019: https://news.ycombinator.com/item?id=21783828
pvitz
> But the longer I worked there the more I felt like the plan was basically:
> 1. Manually add more and more common-sense knowledge and extend the inference engine
> 2. ???
> 3. AGI!
In retrospect, this reasoning doesn't seem to be so wrong.
Cthulhu_
I mean if I were to oversimplify and over-abstract AGI into a long list of if / elses, that's how I'd go about it. It's just that there's A Lot to consider.
woodruffw
I don't have strong opinions about Cyc in either direction, but this obit feels pretty mean-spirited to me: it imputes moral failure on Lenat and others for working on symbolic AI for 40 years, without acknowledging that Lenat's basic belief ("priming the pump") is shared by ML too -- the main difference is that ML needed the pump to be primed with an overwhelming amount of compute, which grew faster than Cyc's ability to ingest facts and relations.
(This isn't to imply that I think symbolic AI "works"; only that more perspective on underlying beliefs is due.)
wpietri
I don't think there's anything wrong with exploring a field for decades. There are many scientists who have a mix of successes and failures. But this guy spend his whole life and many years of other people's lives trying one single thing that never really worked. You could call that being a single-minded visionary, but I don't think it's unreasonable for others to think it either kooky or a giant waste.
A useful comparison to me here is all the alchemical efforts to turn lead into gold. Can modern physicists do that? Not economically, but sure. [1] If alchemists had just persisted, would they have gotten there too? No, it was a giant waste, and pretty loony to a modern eye. And I'd say both alchemists and a number of AI proponents both are so wrapped up in pursuing specific outcomes (gold, AGI) that they indulge in a lot of magical thinking.
[1] https://www.scientificamerican.com/article/fact-or-fiction-l...
null
zozbot234
I don't think this is correct. The capability for drawing useful logical inferences from something like a Cyc knowledge base is far more compute limited than just doing ML on any given amount of data. We're talking exponential (or worse!) vs. pure linear scaling. This is the real-world, practical reason why the Cyc folks eventually found no value at all in their most general inference engine, and ended up exclusively relying on their custom-authored, more constrained inference generators instead.
woodruffw
Again, I'm not saying Cyc's approach is correct. I'm saying that the underlying hope that made Lenat plow through the AI winter is the same one that made ML researchers plow through it. It's just that the ML researchers reached the end of the tunnel first (for some senses of first).
YeGoblynQueenne
Symbolic AI works. So much that it's "not AI" anymore.
I had the funny thought that this is exactly what a sentient AI would write "stop looking here, there is nothing to see, move along." :-)
I (like vannevar apparently) didn't feel Cyc was going anywhere useful, there were ideas there, but not coherent enough to form a credible basis for even a hypothesis of how a system could be constructed that would embody them.
I was pretty impressed by McCarthy's blocks world demo, later he and a student formalized some of the rules for creating 'context'[1] for AI to operate within, I continue to think that will be crucial to solving some of the mess that LLMs create.
For example, the early failures of LLMs suggesting that you could make salad crunchy by adding rocks was a classic context failure, data from the context of 'humor' and data from the context of 'recipes' intertwined. Because existing models have no context during training, there is nothing in the model that 'tunes' the output based on context. And you get rocks in your salad.
[1] https://citeseerx.ist.psu.edu/document?repid=rep1&type=pdf&d...