Master LLMs with our FREE course in collaboration with Activeloop & Intel Disruptor Initiative. Join now!

Publication

How To Do RAG Without Vector Databases
Data Science   Latest   Machine Learning

How To Do RAG Without Vector Databases

Last Updated on March 13, 2024 by Editorial Team

Author(s): ___

Originally published on Towards AI.

Knowledge Graphs Are All You Need 😀

When it comes to bestowing Large Language Models (LLMs) with long-term memory, the prevalent approach often involves a Retrieval Augmented Generation (RAG) solution, with vector databases acting as the storage mechanism for the long-term memory. This begs the question: Can we achieve the same results without vector databases?

Enter RecallM: An Adaptable Memory Mechanism with Temporal Understanding for Large Language Models by Brandon Kynoch, Hugo Latapie, and Dwane van der Sluis. This paper proposes the use of an automatically constructed knowledge graph as the backbone of long-term memory for LLMs.

This blog post is a deep dive into the mechanics of RecallM, focusing on how it updates its knowledge graph and performs inference, underpinned by a series of illustrative examples.

We’ll start by exploring how the knowledge graph updates work, walking through two specific examples to clarify the process. Following that, we’ll examine the inference mechanism of RecallM with another example, showcasing how it draws on the knowledge graph to generate responses. Our discussion will also cover examples of temporal reasoning, demonstrating RecallM’s proficiency in understanding and applying time-based knowledge. Lastly, we’ll touch upon the limitations of this approach, providing a balanced view of its capabilities and areas for further development.

Let’s say we… Read the full blog for free on Medium.

Join thousands of data leaders on the AI newsletter. Join over 80,000 subscribers and keep up to date with the latest developments in AI. From research to projects and ideas. If you are building an AI startup, an AI-related product, or a service, we invite you to consider becoming a sponsor.

Published via Towards AI

Feedback ↓