Semantics at an Angle: When Cosine Similarity Works Until It Doesn't
By: Kisung You
Potential Business Impact:
Makes computer learning understand words better.
Cosine similarity has become a standard metric for comparing embeddings in modern machine learning. Its scale-invariance and alignment with model training objectives have contributed to its widespread adoption. However, recent studies have revealed important limitations, particularly when embedding norms carry meaningful semantic information. This informal article offers a reflective and selective examination of the evolution, strengths, and limitations of cosine similarity. We highlight why it performs well in many settings, where it tends to break down, and how emerging alternatives are beginning to address its blind spots. We hope to offer a mix of conceptual clarity and practical perspective, especially for quantitative scientists who think about embeddings not just as vectors, but as geometric and philosophical objects.
Similar Papers
How Small Transformation Expose the Weakness of Semantic Similarity Measures
Computation and Language
Finds computer code that means the same thing.
Testing the assumptions about the geometry of sentence embedding spaces: the cosine measure need not apply
Computation and Language
Computers understand sentences better than we thought.
On the Similarities of Embeddings in Contrastive Learning
Machine Learning (CS)
Improves AI learning from less data.