From Words to Worlds: Rethinking Embeddings and Ranking in Retrieval
Last Updated on October 18, 2025 by Editorial Team
Author(s): Hira Ahmad
Originally published on Towards AI.
To choose the right model for semantic search, consider the trade-offs between a bi-encoder’s speed, a cross-encoder’s precision, and ColBERT’s balance of both.
Words alone are insufficient to capture communication; the full message is a shared experience, linked with the intent of the speaker, the specific situation, and unspoken bonds between people. Embedding models help computers understand the complexity of human information by turning it into numbers. They take text, convert it into vectors, and let us search for meaning by geometry. But converting isn’t enough we must rank relevance. And how we do that and what trade-offs we accept shapes what our system can remember, can notice, can retrieve.

This article discusses the intricacies of embedding models used in semantic search, focusing on the importance of ranking approaches such as bi-encoders, cross-encoders, and ColBERT. It highlights the strengths and weaknesses of each model regarding speed, precision, and memory efficiency. Additionally, it provides insights into practical strategies for document retrieval and emphasizes the need for careful consideration when choosing a model that balances computational costs with the required accuracy in context-based querying.
Read the full blog for free on Medium.
Join thousands of data leaders on the AI newsletter. Join over 80,000 subscribers and keep up to date with the latest developments in AI. From research to projects and ideas. If you are building an AI startup, an AI-related product, or a service, we invite you to consider becoming a sponsor.
Published via Towards AI
Towards AI Academy
We Build Enterprise-Grade AI. We'll Teach You to Master It Too.
15 engineers. 100,000+ students. Towards AI Academy teaches what actually survives production.
Start free — no commitment:
→ 6-Day Agentic AI Engineering Email Guide — one practical lesson per day
→ Agents Architecture Cheatsheet — 3 years of architecture decisions in 6 pages
Our courses:
→ AI Engineering Certification — 90+ lessons from project selection to deployed product. The most comprehensive practical LLM course out there.
→ Agent Engineering Course — Hands on with production agent architectures, memory, routing, and eval frameworks — built from real enterprise engagements.
→ AI for Work — Understand, evaluate, and apply AI for complex work tasks.
Note: Article content contains the views of the contributing authors and not Towards AI.