The Platonic Representation Hypothesis: Towards Universal Embedding Inversion and Whale Communication

2025-07-18
The Platonic Representation Hypothesis: Towards Universal Embedding Inversion and Whale Communication

Researchers have discovered that large language models converge towards a shared underlying representation space as they grow larger, a phenomenon termed the 'Platonic Representation Hypothesis'. This suggests that different models learn the same features, regardless of architecture. The paper uses the 'Mussolini or Bread' game as an analogy to explain this shared representation, and further supports it with compression theory and model generalization. Critically, based on this hypothesis, researchers developed vec2vec, a method for unsupervised conversion between embedding spaces of different models, achieving high-accuracy text embedding inversion. Future applications could involve decoding ancient texts (like Linear A) or translating whale speech, opening new possibilities for cross-lingual understanding and AI advancement.

Read more

Scaling RL: Next-Token Prediction on the Web

2025-07-13
Scaling RL: Next-Token Prediction on the Web

The author argues that reinforcement learning (RL) is the next frontier for training AI models. Current approaches of scaling many environments simultaneously are messy. Instead, the author proposes training models to reason by using RL for next-token prediction on web-scale data. This leverages the vast amount of readily available web data, moving beyond the limitations of current RL training datasets focused on math and code problems. By unifying RL with next-token prediction, the approach promises to create significantly more powerful reasoning models.

Read more
AI

AI's Bottleneck: Data, Not Algorithms?

2025-06-30
AI's Bottleneck: Data, Not Algorithms?

AI has seen incredible progress, but the pace seems to be slowing. This article argues that past major AI breakthroughs (DNNs, Transformers, RLHF, reasoning models) stemmed not from novel algorithms, but from unlocking new data sources (ImageNet, web text, human feedback, verifiers). The author suggests future breakthroughs will likely come not from algorithmic innovation, but from effectively utilizing new data sources like video and robotic sensors, as existing datasets may be approaching their knowledge limits.

Read more