The Fractured Entangled Representation Hypothesis
13 comments
·May 20, 2025null
scarmig
Did you investigate other search processes besides SGD? I'm thinking of those often termed "biologically plausible" (e.g. forward-forward, FA). Are their internal representations closer to the fractured or unified representations?
null
timewizard
> Much of the excitement in modern AI is driven by the observation that scaling up existing systems leads to better performance.
Scaling up almost always leads to better performance. If you're only getting linear gains though then there is absolutely nothing to be excited about. You are in a dead end.
akarshkumar0101
Tweet: https://x.com/kenneth0stanley/status/1924650124829196370 Arxiv: https://arxiv.org/abs/2505.11581
pvg
Sounds like you're one of the co-authors? Probably worth mentioning if the case so people know they can discuss the work with one of the work-doers.
akarshkumar0101
I mentioned that in the original post, but I don't see that text here anymore (thats why I added links via comment)... I am new to hackernews
messe
I believe they just mean that you should edit the comment where you added the links to mention that you are the author, to add that additional context.
macintux
I believe this could (or should) have been a Show HN, which would have allowed you to include explanatory text. See the top of this page for the rules.
https://news.ycombinator.com/show
Welcome to the site. There are a lot of features which are less obvious, which you’ll discover over time.
null
I can't really take this paper seriously because they are severely lacking references to the linear representation hypothesis. Much work on neural network interpretability lately has shown individual neurons are polysemantic, and therefore practically useless for explainability. My hypothesis is fitting linear probes (or a sparse autoencoder) would reveal linearly semantic attributes.
It is unfortunate because they briefly mention Neel Nanda's Othello experiments, but not the wide array of experiments like the NeurIPS Oral "Linear Representation Hypothesis in Language Models" or even golden gate Claude.