
Exploring ColBERT with RAGatouille
Date : 2024-01-27
Description
This summary was drafted with mixtral-8x7b-instruct-v0.1.Q5_K_M.gguf
Simon Willison delves into the workings of ColBERT, a retrieval model designed for scalable BERT-based search over extensive text collections. He explains how ColBERT differs from regular embedding models and how it provides more information than traditional embedding search by showing which words in the document are most relevant. Willison then proceeds to use RAGatouille, a library that makes working with ColBERT easier, to create an index of his blog's content. He also demonstrates querying the index and implementing a basic question-answering mechanism using LLM. The article further explores re-ranking queries without building an index first.
Read article here
Recently on :
Artificial Intelligence
Information Processing | Computing
PITTI - 2026-03-05
Scaling Trust : a Missing Piece in Multi-Agent Worlds
Humanity’s ability to build complex civilizations relies on an "invisible infrastructure" - the shared culture, institutions, a...
PITTI - 2026-01-14
Cultural, Ideological and Political Bias in LLMs
Transcription of a talk given during the work sessions organized by Technoréalisme on December 9, 2025, in Paris. The talk pres...
WEB - 2025-11-13
Measuring political bias in Claude
Anthropic gives insights into their evaluation methods to measure political bias in models.
WEB - 2025-10-09
Defining and evaluating political bias in LLMs
OpenAI created a political bias evaluation that mirrors real-world usage to stress-test their models’ ability to remain objecti...
WEB - 2025-07-23
Preventing Woke AI In Federal Government
Citing concerns that ideological agendas like Diversity, Equity, and Inclusion (DEI) are compromising accuracy, this executive ...