How To Do RAG Without Vector Databases
Last Updated on March 13, 2024 by Editorial Team
Author(s): ___
Originally published on Towards AI.
Knowledge Graphs Are All You Need π
When it comes to bestowing Large Language Models (LLMs) with long-term memory, the prevalent approach often involves a Retrieval Augmented Generation (RAG) solution, with vector databases acting as the storage mechanism for the long-term memory. This begs the question: Can we achieve the same results without vector databases?
Enter RecallM: An Adaptable Memory Mechanism with Temporal Understanding for Large Language Models by Brandon Kynoch, Hugo Latapie, and Dwane van der Sluis. This paper proposes the use of an automatically constructed knowledge graph as the backbone of long-term memory for LLMs.
This blog post is a deep dive into the mechanics of RecallM, focusing on how it updates its knowledge graph and performs inference, underpinned by a series of illustrative examples.
Weβll start by exploring how the knowledge graph updates work, walking through two specific examples to clarify the process. Following that, weβll examine the inference mechanism of RecallM with another example, showcasing how it draws on the knowledge graph to generate responses. Our discussion will also cover examples of temporal reasoning, demonstrating RecallMβs proficiency in understanding and applying time-based knowledge. Lastly, weβll touch upon the limitations of this approach, providing a balanced view of its capabilities and areas for further development.
Letβs say we… Read the full blog for free on Medium.
Join thousands of data leaders on the AI newsletter. Join over 80,000 subscribers and keep up to date with the latest developments in AI. From research to projects and ideas. If you are building an AI startup, an AI-related product, or a service, we invite you to consider becoming aΒ sponsor.
Published via Towards AI