Discussion about this post

User's avatar
Stephen Tucker's avatar

Grateful for this read, which introduced me to the convergent representation hypothesis. I'm a neuroscience instrumentation engineer, not a computer scientist, and I've been following closely the developments in compressive sensing because I think these ideas may be important to brain recording. Anyway, I'm not sure I grasp your final point, or how these ideas relate--you suggest that we can hope to decode the few samples of Linear A we have by leveraging an otherwise complete corpus of language embeddings? At some point, the limited amount of Linear A we have still makes this a very hard inversion problem. (Luckily we can continue to record the whales...)

Expand full comment
suman suhag's avatar

There are several advantages to using word embeddings instead of character embeddings when training a deep neural network. First, word embeddings provide a higher level of abstraction than character embeddings. This allows the network to learn the relationships between words, rather than the individual characters that make up those words. This can lead to improved performance on tasks such as language modeling and machine translation.

Second, word embeddings are typically much smaller than character embeddings. This is because each word is represented by a single vector, rather than a vector for each character in the word. This can make training faster and more efficient.

Third, word embeddings are already available for many languages, which can save time when training a new model.

There are also some disadvantages to using word embeddings. One is that they can be less accurate than character embeddings, especially for rare words. Another is that they can be less effective for tasks that require understanding of the syntactic structure of a sentence, such as parsing.

Expand full comment
8 more comments...

No posts