10 Comments
User's avatar
Stephen Tucker's avatar

Grateful for this read, which introduced me to the convergent representation hypothesis. I'm a neuroscience instrumentation engineer, not a computer scientist, and I've been following closely the developments in compressive sensing because I think these ideas may be important to brain recording. Anyway, I'm not sure I grasp your final point, or how these ideas relate--you suggest that we can hope to decode the few samples of Linear A we have by leveraging an otherwise complete corpus of language embeddings? At some point, the limited amount of Linear A we have still makes this a very hard inversion problem. (Luckily we can continue to record the whales...)

Expand full comment
Dr. Paul Wilhelm's avatar

Great article, it gave me inspiration for exploring the embedding space in a new way!

Expand full comment
suman suhag's avatar

There are several advantages to using word embeddings instead of character embeddings when training a deep neural network. First, word embeddings provide a higher level of abstraction than character embeddings. This allows the network to learn the relationships between words, rather than the individual characters that make up those words. This can lead to improved performance on tasks such as language modeling and machine translation.

Second, word embeddings are typically much smaller than character embeddings. This is because each word is represented by a single vector, rather than a vector for each character in the word. This can make training faster and more efficient.

Third, word embeddings are already available for many languages, which can save time when training a new model.

There are also some disadvantages to using word embeddings. One is that they can be less accurate than character embeddings, especially for rare words. Another is that they can be less effective for tasks that require understanding of the syntactic structure of a sentence, such as parsing.

Expand full comment
noreson's avatar

Same datesets will generate same model

Expand full comment
schizoid intellectual's avatar

This is a very bold claim. If I take this idea to, what I think, is its logical conclusion, then are you suggesting that representations in neural networks have the converge to representations inside our brains? In other words, are you claiming that the platonic representations are completely agnostic of the substance that is generating those representations?

Expand full comment
DotProduct's avatar

Thank you. So some thoughts: first multiple different organisations are ploughing bazillions each into creating more or less the same thing- somewhat redundant and inefficient duplicative effort n’est pas?; second, this is a platonic ai 2025- platonic ai 1925 was v different (racist, misogynist etc) and so will platonic ai 2125 because we are using time stamped culturally curated data; finally; whale song, bee dances and ant smells will be super cool if we haven’t bumped them off by 2050…

Expand full comment
Kavya's avatar

Fascinating read! Two questions that came up while reading:

(1) What if these models converge to the same but "wrong" abstraction in a domain? "Wrong" as in it doesn't reflect the actual truth of the world.

(2) How much does training data overlap contribute to true convergence?

Imo, the first question matters more because humans can create value if they find places where models' convergent representations veer away from reality.

Expand full comment
Jon Simon's avatar

1. Humans do this too, they're called optical illusions and cognitive biases

Expand full comment
Nithya Kota's avatar

Very nicely written, thank you very much for this blog, on a side note I am a bit skeptical on decoding Linear A , whale language, doesn't PRH rely on the fact that all the models are seeing the same world, where as the world as seen by Linear A or whale language are quite different from our own

Expand full comment
Praveen Selvaraj's avatar

Isn't it the case that the world isn't different but rather that it's different animals with different brains seeing a subset of the world ? There's much more overlap of brains and what's being seen in the case of the Ancient Greeks, ofc.

But even for animals with less overlap, it's possible that they have a representation of the world not too dissimilar from our own.

Expand full comment