Name: Towards AI Legal Name: Towards AI, Inc. Description: Towards AI is the world's leading artificial intelligence (AI) and technology publication. Read by thought-leaders and decision-makers around the world. Phone Number: +1-650-246-9381 Email: pub@towardsai.net
228 Park Avenue South New York, NY 10003 United States
Website: Publisher: https://towardsai.net/#publisher Diversity Policy: https://towardsai.net/about Ethics Policy: https://towardsai.net/about Masthead: https://towardsai.net/about
Name: Towards AI Legal Name: Towards AI, Inc. Description: Towards AI is the world's leading artificial intelligence (AI) and technology publication. Founders: Roberto Iriondo, , Job Title: Co-founder and Advisor Works for: Towards AI, Inc. Follow Roberto: X, LinkedIn, GitHub, Google Scholar, Towards AI Profile, Medium, ML@CMU, FreeCodeCamp, Crunchbase, Bloomberg, Roberto Iriondo, Generative AI Lab, Generative AI Lab Denis Piffaretti, Job Title: Co-founder Works for: Towards AI, Inc. Louie Peters, Job Title: Co-founder Works for: Towards AI, Inc. Louis-François Bouchard, Job Title: Co-founder Works for: Towards AI, Inc. Cover:
Towards AI Cover
Logo:
Towards AI Logo
Areas Served: Worldwide Alternate Name: Towards AI, Inc. Alternate Name: Towards AI Co. Alternate Name: towards ai Alternate Name: towardsai Alternate Name: towards.ai Alternate Name: tai Alternate Name: toward ai Alternate Name: toward.ai Alternate Name: Towards AI, Inc. Alternate Name: towardsai.net Alternate Name: pub.towardsai.net
5 stars – based on 497 reviews

Frequently Used, Contextual References

TODO: Remember to copy unique IDs whenever it needs used. i.e., URL: 304b2e42315e

Resources

Take our 85+ lesson From Beginner to Advanced LLM Developer Certification: From choosing a project to deploying a working product this is the most comprehensive and practical LLM course out there!

Publication

RAG Explained: A Comprehensive Guide to Mastering Retrieval-Augmented Generation
Artificial Intelligence   Latest   Machine Learning

RAG Explained: A Comprehensive Guide to Mastering Retrieval-Augmented Generation

Last Updated on February 17, 2025 by Editorial Team

Author(s): Ajit Kumar Singh

Originally published on Towards AI.

Image by Author

Hi Everyone 👋

Recently, I worked on a use case involving product matching, which required implementing RAG modeling. At the time, I had limited knowledge of RAG and no hands-on experience, so I started researching. However, I found that most tutorials focused on specific aspects, making it difficult for beginners to connect the dots.

The main goal of this article is to provide a detailed, beginner-friendly guide to RAG modeling. Whether you’re completely new to the concept or already have some practical experience, this tutorial will help you build a clear and comprehensive understanding of RAG, from its fundamentals to implementation. 🚀

By the end of this article, you will have explored the following key topics 📚:

Introduction to RAG

What is RAG?

RAG, which stands for Retrieval-Augmented Generation, was initially introduced by researchers at Facebook in their paper. It is an AI framework that merges the advantages of traditional information retrieval systems (such as search engines and databases) with the capabilities of generative models, like Large Language Models (LLMs). Think of RAG as a hybrid model that utilizes both parametric and non-parametric memory. Parametric memory refers to the weights of a pre-trained transformer, while non-parametric memory consists of a dense vector index derived from your external database.

RAG operates through three key components:

  • Retrieval
  • Generation
  • Augmentation

In the following sections, I will delve into each of these components in greater detail.

Why does it matter in AI and NLP?

While large language models (LLMs) have made impressive strides, they still encounter notable challenges, especially when it comes to domain-specific or knowledge-intensive tasks. A major issue is the generation of “hallucinations,” where the model produces inaccurate or fabricated information, especially when faced with queries outside its training data or those requiring up-to-date knowledge.

To address these limitations, Retrieval-Augmented Generation (RAG) enhances LLMs by incorporating external knowledge. It does this by retrieving relevant document segments from an external knowledge base through semantic similarity calculations. By referencing this external information, RAG effectively mitigates the problem of generating factually incorrect content.

In what ways does RAG differ from traditional language models?

Here, I have outlined the key differences between traditional LLMs and RAG models.

Image by Author

The Importance and Necessity of RAG

Limitations of standard language models (LLMs/SLMs)

Large pre-trained language models have demonstrated the ability to store vast amounts of factual knowledge within their parameters, achieving state-of-the-art results when fine-tuned for various downstream NLP tasks. However, their ability to retrieve, update, and precisely manipulate knowledge remains a significant limitation. Since these models rely solely on their pre-trained weights, they struggle with knowledge-intensive tasks where real-time access to external information is crucial.

Moreover, pre-trained LLMs face several key challenges:

  1. Stale or Outdated Information — Once trained, they cannot dynamically update their knowledge without expensive retraining. This is a major drawback for tasks requiring the latest facts, such as news summarization, financial analysis, or legal research.
  2. Hallucination & Misinformation — They often generate incorrect or fabricated responses, as they lack the ability to fact-check against an external knowledge source.
  3. Lack of Interpretability — Traditional LLMs do not provide citations or verifiable sources, making it difficult to assess the credibility of their responses.
  4. Computational & Storage Costs — Storing knowledge within model parameters requires scaling up to massive architectures, increasing inference costs and latency.

Due to these limitations, knowledge-intensive NLP applications such as question answering, legal AI, and enterprise search demand a more efficient and factually grounded approach — this is where Retrieval-Augmented Generation (RAG) comes into play.

How RAG improves accuracy and relevance

Here’s how RAG enhances accuracy and relevance:

  1. Access to Up-to-Date Knowledge: Unlike purely parametric models that rely on static knowledge stored in their weights, RAG retrieves relevant documents from an external corpus (e.g., Wikipedia). This allows it to generate responses based on the latest and most accurate information
  2. Reduced Hallucination: Parametric-only models sometimes generate incorrect or fabricated information (“hallucinations”). By conditioning generation on retrieved factual documents, RAG reduces the likelihood of generating misleading or factually incorrect responses
  3. Better Handling of Knowledge-Intensive Tasks: RAG has been shown to outperform both extractive models (which select spans from retrieved texts) and purely generative models (which rely only on internal knowledge). It achieves state-of-the-art results in open-domain QA tasks such as Natural Questions and TriviaQA​.
  4. Flexible and Adaptive Learning: RAG allows for easy updating of its knowledge base by replacing or updating its retrieval corpus. This is unlike traditional fine-tuning, which requires retraining the entire model​.

Understanding the RAG Architecture and How It Functions

RAG Architecture (image by author)

Let’s dive deeper into each component in detail!

Pre-Processing Pipeline

Data Extraction — Converting documents into a structured format by extracting text, images, and other assets.

  • Basic Extraction: Outputs flat text without structure.
  • Structure-Preserving Extraction: Retains sections, sub-sections, and paragraph formatting.
  • Table, Image, and Asset Extraction: Captures additional document elements.

Chunking — Splitting documents into smaller, meaningful segments to fit within context windows.

  • Fixed-Length Chunking: Divides text based on predefined size.
  • Structure-Aware Chunking: Maintains document hierarchy and logical flow.
  • Recursive Chunking: Iteratively breaks down content while preserving coherence.
  • Semantic/Topic-Based Chunking: Segments text based on meaning and topic relevance.
  • Summarization-Based Chunking: Generates concise summaries at document and sub-document levels.
Document Chunking (image by author)

For the preprocessing steps, we have several key tools and options available to accomplish them.

Frameworks for data processing (image by author)

Embedding Models

Chunks are encoded into vector representations using an embedding model and stored in a vector database. This step is crucial for enabling efficient similarity searches in the subsequent retrieval phase.

Wide range of models with variations in:

  • Multilingual support
  • Context window (512 to 8k tokens)
  • Domain (finance, biomedical, etc)

Various embedding models are trained on diverse datasets and exhibit varying effectiveness across different domains. To determine which embedding best suits your specific task, it’s essential to evaluate the available options.

Below is a list of prominent embedding providers:

Available Embeddings (image by author)

Vector Database

It is a specialized system designed to store and index high-dimensional vector embeddings, which are numerical representations of data (such as text, images, etc.). It enables efficient similarity searches — often via approximate nearest neighbor (ANN) algorithms — making them ideal for tasks that require rapid retrieval of relevant items from large datasets. In RAG, the vector database serves as the non-parametric memory, holding dense embeddings of documents (e.g., passages from Wikipedia).

Indexing in RAG Pipeline (image by author)

There are numerous providers available for hosting your vector embeddings. The choice of provider depends on factors such as your specific use case and budget.

Vector Database Providers (image by author)

Retrieval

When you ask a question to the retriever, it uses similarity search to scan through a vast knowledge base of vector embeddings. It then pulls out the most relevant vectors to help answer that query.

Retrieval in RAG Model (image by author)

There are a few different techniques it can use to know what’s relevant:

  • Indexing process — organizes the data into your vector database in a way that makes it easily searchable. This allows the RAG to access relevant information when responding to a query.
  • Query vectorization — Once you have vectorized your knowledge base you can do the same to the user query. When the model sees a new query, it uses the same preprocessing and embedding techniques. This ensures that the query vector is compatible with the document vectors in the index.
  • Semantic Search — When the system needs to find the most relevant documents or passages to answer a query, it utilizes vector similarity techniques.

Generator

In a RAG-LLM framework, the generator is typically a large transformer model — examples include GPT-3.5, GPT-4, Llama2, Falcon, PaLM, and BERT. This generator accepts both the input query and the retrieved documents, which are merged into a single concatenated input. By incorporating the extra context and information from the retrieved documents, the generator can produce a more informed and accurate response, thereby reducing the likelihood of hallucinations.

Generator in RAG Model (image by author)

There are numerous providers of large language models (LLMs) that can serve as generators in your RAG pipeline.

LLMs Providers (image by author)

A Comparison of RAG, Fine-Tuning, and Prompt Engineering

When optimizing large language models (LLMs), three primary methods are often compared: Retrieval-Augmented Generation (RAG), Fine-Tuning (FT), and Prompt Engineering. They differ along two key dimensions: external knowledge requirements and the degree of model adaptation needed.

Prompt Engineering

Simple Prompting (image by author)
  • Approach: Leverages the model’s inherent capabilities by crafting precise prompts without significant external input or model adjustments.
  • Use Case: Ideal for tasks where the model already has sufficient internal knowledge and only minor guidance is needed.
  • Trade-offs: The outputs are limited to what the model has learned during its initial training, making it less effective for highly specialized or up-to-date information.

Retrieval-Augmented Generation (RAG)

RAG Archive (image by author)
  • Approach: Augments the LLM with an external “textbook” of knowledge by retrieving relevant information in real time.
  • Strengths: Excels in dynamic environments by enabling real-time knowledge updates and improving interpretability.
  • Trade-offs: Comes with higher latency and potential ethical concerns regarding the source and handling of retrieved data.

Fine-Tuning (FT)

Fine Tuning LLMs (image by author)
  • Approach: Involves retraining the model so it internalizes specific knowledge, styles, or formats over time.
  • Strengths: Allows deep customization, potentially reducing hallucinations and replicating specific content structures.
  • Trade-offs: It is a more static solution — updates require retraining — and it demands substantial computational resources. Moreover, LLMs may struggle to learn entirely new factual information through unsupervised fine-tuning.

RAG vs Long Context LLMs

RAG Models came into picture because of limitations of early LLMs like

  • Context Window Limitations: Early language models, like ChatGPT, had small context windows (a few thousand tokens).
  • Retrieval for More Information: To process more data than LLMs could fit in their memory, RAG was developed.
  • Real-Time & Accurate Retrieval: RAG allowed LLMs to fetch external documents, ensuring up-to-date and relevant answers beyond their training data.

Modern LC LLMs (e.g., Gemini-1.5, GPT-4o) now support 128k to 1M+ tokens, reducing the need to “fetch” relevant snippets. Instead of retrieving, LC stores all information in memory and processes it holistically. LC LLM Handles complex queries, multi-hop reasoning, and implicit information retrieval better than RAG.

Long context vs RAG (image by author)

Real-World Applications of RAG in AI and NLP

There is a wide variety of applications for RAG across several use cases. Here are some key areas where RAG is gaining significant popularity:

  • Conversational AI and Chatbots: RAG enables chatbots to access up-to-date information, providing users with accurate and timely responses.
  • Question Answering Systems: In domains like legal, healthcare, and research, RAG-powered systems retrieve pertinent data from extensive databases, delivering precise answers to complex queries.
  • Code Generation and Programming Assistants: By retrieving relevant code snippets and documentation, RAG assists in generating accurate code and offering programming support.
  • Personalized Recommendations and Search Augmentation: RAG enhances search engines by retrieving and generating content tailored to user preferences, improving the relevance of search results

The Challenges and Limitations of RAG

The key limitations of the RAG model are as follows:

Balancing Retrieval and Generation:

Striking the right balance is critical. If the retrieval component supplies data that is too broad or irrelevant, it can negatively impact the quality of the generated output. Conversely, overly specific or niche information might constrain the generator, reducing its ability to produce creative or flexible responses. This balance requires continual tuning and adjustment.

Latency issues in real-time applications:

Integrating a retrieval step adds extra latency to the overall processing time. Querying large external knowledge bases and then processing the combined input increases computational complexity, which can affect real-time applications.

Dependence on External Knowledge Quality:

The performance of a RAG model is highly dependent on the quality, relevance, and timeliness of its external data sources. If the retrieved documents contain outdated or inaccurate information, the generated responses may also be flawed.

Best Practices for Working with RAG

Before you rush your RAG system into production — or worse, settle for subpar results — pause and refine your approach. Here are some seasoned best practices to elevate your RAG performance and get it production-ready:

Garbage In Garbage Out:

Remember, garbage in equals garbage out. The richer and cleaner your source data, the more reliable your output. Ensure your data pipeline preserves critical information (like spreadsheet headers) while stripping away any extraneous markup that could confuse your LLM.

Master Your Data Splitting:

Not all datasets are created equal. Experiment with various text chunk sizes to maintain the right context for your RAG-enabled inference. Build multiple vector stores using different splitting strategies and determine which configuration best complements your architecture.

Fine-Tune Your System Prompt:

If your LLM isn’t fully leveraging the provided context, it’s time to recalibrate your system prompt. Clearly outline your expectations on how the model should process and utilize the information — it’s a small tweak that can yield big improvements.

Filter with Precision:

Customize your vector store results by filtering based on metadata. For instance, if you need procedural content, restrict results to documents tagged as ‘how-to’ by filtering on the appropriate metadata field. This targeted approach helps ensure that only the most relevant content is returned.

Experiment with Embeddings:

Different embedding models bring unique strengths to the table. Try out several options — and consider fine-tuning your own — to see which best captures your domain-specific nuances. Check out the MTEB leaderboard for the latest in high-performing open source embeddings. Utilizing your cleaned, processed knowledge base to fine-tune an embedding model can take your query results to the next level.

By applying these best practices, you not only improve your RAG system’s output but also set the stage for a robust, production-grade solution.

How to Evaluate RAG Models

Evaluating Retrieval-Augmented Generation (RAG) models involves assessing both the retrieval component (which handles information retrieval) and the generation component (which generates responses based on retrieved information).

Evaluation Metrics

  • Retrieval Component: Measure retrieval quality using metrics like Precision at K, Recall at K, MRR and NDCG.
  • Generation Component: Evaluate the fluency, relevance, and factuality of generated responses using metrics like BLEU, ROUGE, and METEOR.
  • End-to-End: Human evaluation for relevance, coherence, factuality, and fluency; assess task-specific performance.
  • Efficiency: Monitor latency, throughput, and resource usage.
  • Real-world Testing: Use A/B testing and user feedback for real-world performance insights.

A Step-by-Step Guide to Implementing RAG

Up until now, we’ve covered the theoretical aspects of RAG and its various details. Now, it’s time to roll up our sleeves and dive into the practical implementation. For the code and additional details, feel free to check out my github link.

Problem Statement:
The challenge is to build a Retrieval-Augmented Generation (RAG) system that processes PDFs locally, offering a solution for analyzing technical, legal, and academic documents while prioritizing privacy, cost efficiency, and customizability.

Proposed Solution:
The proposed solution combines LangChain, DeepSeek-R1, Ollama, and Streamlit to create a local RAG system. This system ingests PDFs, retrieves relevant information, and generates precise answers without relying on cloud services, ensuring data privacy and cost-effectiveness. It utilizes DeepSeek-R1’s powerful reasoning capabilities and LangChain’s modular framework to process, retrieve, and generate accurate responses.

Test Stack Used:

  • LangChain: AI framework for managing RAG workflows, including document loaders and vector stores.
  • DeepSeek-R1: A reasoning LLM (7B model) for problem-solving and technical tasks, deployed locally via Ollama.
  • Ollama: CLI tool for managing local deployment of AI models like DeepSeek-R1.
  • ChromaDB: Vector database for storing and retrieving document embeddings based on similarity.
  • Streamlit: User interface framework for building interactive web apps, used to allow users to interact with the RAG system.

GitHub Repository: Building-RAG-System-with-Deepseek-R1

Conclusion

RAG has been a game-changer in enhancing LLM reliability by reducing hallucinations and integrating real-time, factual knowledge into AI responses. It remains indispensable in high-stakes domains like finance, law, and healthcare, where accuracy and verifiable sources matter.

However, RAG is not without challenges. Its effectiveness depends on

  • retrieval quality
  • embedding accuracy
  • search efficiency

These factors can introduce complexity. The rise of long-context LLMs has also raised questions about RAG’s necessity, as these models now handle hundreds of thousands of tokens in a single prompt.

Despite criticism, RAG is here to stay for several reasons:

Scalability for Large-Scale Data — Long-context models still struggle with scaling efficiently, making RAG a cost-effective alternative.

Superior Performance in Enterprise Use Cases — Optimized RAG pipelines outperform long-context models when dealing with dynamic, multi-source enterprise data.

Ensuring a Reliable Source of Truth — RAG’s ability to retrieve real-time, up-to-date, and verifiable knowledge is unmatched.

Powering AI Agents with Multi-Source Intelligence — RAG enables agentic AI systems that dynamically pull insights from multiple databases, APIs, and knowledge graphs.

I hope you enjoyed the read and gained valuable insights to help speed up your own RAG pipeline.

Cheers! 🍻

References

  1. Retrieval-Augmented Generation for Large Language Models: A Survey
  2. Beyond RAG: Knowledge-Engineered Generation for LLMs
  3. RAG is here to stay: four reasons why large context windows can’t replace it
  4. Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks
  5. What is Retrieval-Augmented Generation (RAG)
  6. Best Open Source Vector Databases: A Comprehensive Guide
  7. What is RAG: Understanding Retrieval-Augmented Generation
  8. Prompt Engineering vs Fine-tuning vs RAG
  9. Fine-tuning vs. RAG: Understanding the Difference
  10. RAG vs. Long-context LLMs
  11. Beginner’s Guide to RAG with Prof. Tom Yeh
  12. Retrieval augmented generation: Keeping LLMs relevant and current

Join thousands of data leaders on the AI newsletter. Join over 80,000 subscribers and keep up to date with the latest developments in AI. From research to projects and ideas. If you are building an AI startup, an AI-related product, or a service, we invite you to consider becoming a sponsor.

Published via Towards AI

Feedback ↓

Sign Up for the Course
`; } else { console.error('Element with id="subscribe" not found within the page with class "home".'); } } }); // Remove duplicate text from articles /* Backup: 09/11/24 function removeDuplicateText() { const elements = document.querySelectorAll('h1, h2, h3, h4, h5, strong'); // Select the desired elements const seenTexts = new Set(); // A set to keep track of seen texts const tagCounters = {}; // Object to track instances of each tag elements.forEach(el => { const tagName = el.tagName.toLowerCase(); // Get the tag name (e.g., 'h1', 'h2', etc.) // Initialize a counter for each tag if not already done if (!tagCounters[tagName]) { tagCounters[tagName] = 0; } // Only process the first 10 elements of each tag type if (tagCounters[tagName] >= 2) { return; // Skip if the number of elements exceeds 10 } const text = el.textContent.trim(); // Get the text content const words = text.split(/\s+/); // Split the text into words if (words.length >= 4) { // Ensure at least 4 words const significantPart = words.slice(0, 5).join(' '); // Get first 5 words for matching // Check if the text (not the tag) has been seen before if (seenTexts.has(significantPart)) { // console.log('Duplicate found, removing:', el); // Log duplicate el.remove(); // Remove duplicate element } else { seenTexts.add(significantPart); // Add the text to the set } } tagCounters[tagName]++; // Increment the counter for this tag }); } removeDuplicateText(); */ // Remove duplicate text from articles function removeDuplicateText() { const elements = document.querySelectorAll('h1, h2, h3, h4, h5, strong'); // Select the desired elements const seenTexts = new Set(); // A set to keep track of seen texts const tagCounters = {}; // Object to track instances of each tag // List of classes to be excluded const excludedClasses = ['medium-author', 'post-widget-title']; elements.forEach(el => { // Skip elements with any of the excluded classes if (excludedClasses.some(cls => el.classList.contains(cls))) { return; // Skip this element if it has any of the excluded classes } const tagName = el.tagName.toLowerCase(); // Get the tag name (e.g., 'h1', 'h2', etc.) // Initialize a counter for each tag if not already done if (!tagCounters[tagName]) { tagCounters[tagName] = 0; } // Only process the first 10 elements of each tag type if (tagCounters[tagName] >= 10) { return; // Skip if the number of elements exceeds 10 } const text = el.textContent.trim(); // Get the text content const words = text.split(/\s+/); // Split the text into words if (words.length >= 4) { // Ensure at least 4 words const significantPart = words.slice(0, 5).join(' '); // Get first 5 words for matching // Check if the text (not the tag) has been seen before if (seenTexts.has(significantPart)) { // console.log('Duplicate found, removing:', el); // Log duplicate el.remove(); // Remove duplicate element } else { seenTexts.add(significantPart); // Add the text to the set } } tagCounters[tagName]++; // Increment the counter for this tag }); } removeDuplicateText(); //Remove unnecessary text in blog excerpts document.querySelectorAll('.blog p').forEach(function(paragraph) { // Replace the unwanted text pattern for each paragraph paragraph.innerHTML = paragraph.innerHTML .replace(/Author\(s\): [\w\s]+ Originally published on Towards AI\.?/g, '') // Removes 'Author(s): XYZ Originally published on Towards AI' .replace(/This member-only story is on us\. Upgrade to access all of Medium\./g, ''); // Removes 'This member-only story...' }); //Load ionic icons and cache them if ('localStorage' in window && window['localStorage'] !== null) { const cssLink = 'https://code.ionicframework.com/ionicons/2.0.1/css/ionicons.min.css'; const storedCss = localStorage.getItem('ionicons'); if (storedCss) { loadCSS(storedCss); } else { fetch(cssLink).then(response => response.text()).then(css => { localStorage.setItem('ionicons', css); loadCSS(css); }); } } function loadCSS(css) { const style = document.createElement('style'); style.innerHTML = css; document.head.appendChild(style); } //Remove elements from imported content automatically function removeStrongFromHeadings() { const elements = document.querySelectorAll('h1, h2, h3, h4, h5, h6, span'); elements.forEach(el => { const strongTags = el.querySelectorAll('strong'); strongTags.forEach(strongTag => { while (strongTag.firstChild) { strongTag.parentNode.insertBefore(strongTag.firstChild, strongTag); } strongTag.remove(); }); }); } removeStrongFromHeadings(); "use strict"; window.onload = () => { /* //This is an object for each category of subjects and in that there are kewords and link to the keywods let keywordsAndLinks = { //you can add more categories and define their keywords and add a link ds: { keywords: [ //you can add more keywords here they are detected and replaced with achor tag automatically 'data science', 'Data science', 'Data Science', 'data Science', 'DATA SCIENCE', ], //we will replace the linktext with the keyword later on in the code //you can easily change links for each category here //(include class="ml-link" and linktext) link: 'linktext', }, ml: { keywords: [ //Add more keywords 'machine learning', 'Machine learning', 'Machine Learning', 'machine Learning', 'MACHINE LEARNING', ], //Change your article link (include class="ml-link" and linktext) link: 'linktext', }, ai: { keywords: [ 'artificial intelligence', 'Artificial intelligence', 'Artificial Intelligence', 'artificial Intelligence', 'ARTIFICIAL INTELLIGENCE', ], //Change your article link (include class="ml-link" and linktext) link: 'linktext', }, nl: { keywords: [ 'NLP', 'nlp', 'natural language processing', 'Natural Language Processing', 'NATURAL LANGUAGE PROCESSING', ], //Change your article link (include class="ml-link" and linktext) link: 'linktext', }, des: { keywords: [ 'data engineering services', 'Data Engineering Services', 'DATA ENGINEERING SERVICES', ], //Change your article link (include class="ml-link" and linktext) link: 'linktext', }, td: { keywords: [ 'training data', 'Training Data', 'training Data', 'TRAINING DATA', ], //Change your article link (include class="ml-link" and linktext) link: 'linktext', }, ias: { keywords: [ 'image annotation services', 'Image annotation services', 'image Annotation services', 'image annotation Services', 'Image Annotation Services', 'IMAGE ANNOTATION SERVICES', ], //Change your article link (include class="ml-link" and linktext) link: 'linktext', }, l: { keywords: [ 'labeling', 'labelling', ], //Change your article link (include class="ml-link" and linktext) link: 'linktext', }, pbp: { keywords: [ 'previous blog posts', 'previous blog post', 'latest', ], //Change your article link (include class="ml-link" and linktext) link: 'linktext', }, mlc: { keywords: [ 'machine learning course', 'machine learning class', ], //Change your article link (include class="ml-link" and linktext) link: 'linktext', }, }; //Articles to skip let articleIdsToSkip = ['post-2651', 'post-3414', 'post-3540']; //keyword with its related achortag is recieved here along with article id function searchAndReplace(keyword, anchorTag, articleId) { //selects the h3 h4 and p tags that are inside of the article let content = document.querySelector(`#${articleId} .entry-content`); //replaces the "linktext" in achor tag with the keyword that will be searched and replaced let newLink = anchorTag.replace('linktext', keyword); //regular expression to search keyword var re = new RegExp('(' + keyword + ')', 'g'); //this replaces the keywords in h3 h4 and p tags content with achor tag content.innerHTML = content.innerHTML.replace(re, newLink); } function articleFilter(keyword, anchorTag) { //gets all the articles var articles = document.querySelectorAll('article'); //if its zero or less then there are no articles if (articles.length > 0) { for (let x = 0; x < articles.length; x++) { //articles to skip is an array in which there are ids of articles which should not get effected //if the current article's id is also in that array then do not call search and replace with its data if (!articleIdsToSkip.includes(articles[x].id)) { //search and replace is called on articles which should get effected searchAndReplace(keyword, anchorTag, articles[x].id, key); } else { console.log( `Cannot replace the keywords in article with id ${articles[x].id}` ); } } } else { console.log('No articles found.'); } } let key; //not part of script, added for (key in keywordsAndLinks) { //key is the object in keywords and links object i.e ds, ml, ai for (let i = 0; i < keywordsAndLinks[key].keywords.length; i++) { //keywordsAndLinks[key].keywords is the array of keywords for key (ds, ml, ai) //keywordsAndLinks[key].keywords[i] is the keyword and keywordsAndLinks[key].link is the link //keyword and link is sent to searchreplace where it is then replaced using regular expression and replace function articleFilter( keywordsAndLinks[key].keywords[i], keywordsAndLinks[key].link ); } } function cleanLinks() { // (making smal functions is for DRY) this function gets the links and only keeps the first 2 and from the rest removes the anchor tag and replaces it with its text function removeLinks(links) { if (links.length > 1) { for (let i = 2; i < links.length; i++) { links[i].outerHTML = links[i].textContent; } } } //arrays which will contain all the achor tags found with the class (ds-link, ml-link, ailink) in each article inserted using search and replace let dslinks; let mllinks; let ailinks; let nllinks; let deslinks; let tdlinks; let iaslinks; let llinks; let pbplinks; let mlclinks; const content = document.querySelectorAll('article'); //all articles content.forEach((c) => { //to skip the articles with specific ids if (!articleIdsToSkip.includes(c.id)) { //getting all the anchor tags in each article one by one dslinks = document.querySelectorAll(`#${c.id} .entry-content a.ds-link`); mllinks = document.querySelectorAll(`#${c.id} .entry-content a.ml-link`); ailinks = document.querySelectorAll(`#${c.id} .entry-content a.ai-link`); nllinks = document.querySelectorAll(`#${c.id} .entry-content a.ntrl-link`); deslinks = document.querySelectorAll(`#${c.id} .entry-content a.des-link`); tdlinks = document.querySelectorAll(`#${c.id} .entry-content a.td-link`); iaslinks = document.querySelectorAll(`#${c.id} .entry-content a.ias-link`); mlclinks = document.querySelectorAll(`#${c.id} .entry-content a.mlc-link`); llinks = document.querySelectorAll(`#${c.id} .entry-content a.l-link`); pbplinks = document.querySelectorAll(`#${c.id} .entry-content a.pbp-link`); //sending the anchor tags list of each article one by one to remove extra anchor tags removeLinks(dslinks); removeLinks(mllinks); removeLinks(ailinks); removeLinks(nllinks); removeLinks(deslinks); removeLinks(tdlinks); removeLinks(iaslinks); removeLinks(mlclinks); removeLinks(llinks); removeLinks(pbplinks); } }); } //To remove extra achor tags of each category (ds, ml, ai) and only have 2 of each category per article cleanLinks(); */ //Recommended Articles var ctaLinks = [ /* ' ' + '

Subscribe to our AI newsletter!

' + */ '

Take our 85+ lesson From Beginner to Advanced LLM Developer Certification: From choosing a project to deploying a working product this is the most comprehensive and practical LLM course out there!

'+ '

Towards AI has published Building LLMs for Production—our 470+ page guide to mastering LLMs with practical projects and expert insights!

' + '
' + '' + '' + '

Note: Content contains the views of the contributing authors and not Towards AI.
Disclosure: This website may contain sponsored content and affiliate links.

' + 'Discover Your Dream AI Career at Towards AI Jobs' + '

Towards AI has built a jobs board tailored specifically to Machine Learning and Data Science Jobs and Skills. Our software searches for live AI jobs each hour, labels and categorises them and makes them easily searchable. Explore over 10,000 live jobs today with Towards AI Jobs!

' + '
' + '

🔥 Recommended Articles 🔥

' + 'Why Become an LLM Developer? Launching Towards AI’s New One-Stop Conversion Course'+ 'Testing Launchpad.sh: A Container-based GPU Cloud for Inference and Fine-tuning'+ 'The Top 13 AI-Powered CRM Platforms
' + 'Top 11 AI Call Center Software for 2024
' + 'Learn Prompting 101—Prompt Engineering Course
' + 'Explore Leading Cloud Providers for GPU-Powered LLM Training
' + 'Best AI Communities for Artificial Intelligence Enthusiasts
' + 'Best Workstations for Deep Learning
' + 'Best Laptops for Deep Learning
' + 'Best Machine Learning Books
' + 'Machine Learning Algorithms
' + 'Neural Networks Tutorial
' + 'Best Public Datasets for Machine Learning
' + 'Neural Network Types
' + 'NLP Tutorial
' + 'Best Data Science Books
' + 'Monte Carlo Simulation Tutorial
' + 'Recommender System Tutorial
' + 'Linear Algebra for Deep Learning Tutorial
' + 'Google Colab Introduction
' + 'Decision Trees in Machine Learning
' + 'Principal Component Analysis (PCA) Tutorial
' + 'Linear Regression from Zero to Hero
'+ '

', /* + '

Join thousands of data leaders on the AI newsletter. It’s free, we don’t spam, and we never share your email address. Keep up to date with the latest work in AI. From research to projects and ideas. If you are building an AI startup, an AI-related product, or a service, we invite you to consider becoming a sponsor.

',*/ ]; var replaceText = { '': '', '': '', '
': '
' + ctaLinks + '
', }; Object.keys(replaceText).forEach((txtorig) => { //txtorig is the key in replacetext object const txtnew = replaceText[txtorig]; //txtnew is the value of the key in replacetext object let entryFooter = document.querySelector('article .entry-footer'); if (document.querySelectorAll('.single-post').length > 0) { //console.log('Article found.'); const text = entryFooter.innerHTML; entryFooter.innerHTML = text.replace(txtorig, txtnew); } else { // console.log('Article not found.'); //removing comment 09/04/24 } }); var css = document.createElement('style'); css.type = 'text/css'; css.innerHTML = '.post-tags { display:none !important } .article-cta a { font-size: 18px; }'; document.body.appendChild(css); //Extra //This function adds some accessibility needs to the site. function addAlly() { // In this function JQuery is replaced with vanilla javascript functions const imgCont = document.querySelector('.uw-imgcont'); imgCont.setAttribute('aria-label', 'AI news, latest developments'); imgCont.title = 'AI news, latest developments'; imgCont.rel = 'noopener'; document.querySelector('.page-mobile-menu-logo a').title = 'Towards AI Home'; document.querySelector('a.social-link').rel = 'noopener'; document.querySelector('a.uw-text').rel = 'noopener'; document.querySelector('a.uw-w-branding').rel = 'noopener'; document.querySelector('.blog h2.heading').innerHTML = 'Publication'; const popupSearch = document.querySelector$('a.btn-open-popup-search'); popupSearch.setAttribute('role', 'button'); popupSearch.title = 'Search'; const searchClose = document.querySelector('a.popup-search-close'); searchClose.setAttribute('role', 'button'); searchClose.title = 'Close search page'; // document // .querySelector('a.btn-open-popup-search') // .setAttribute( // 'href', // 'https://medium.com/towards-artificial-intelligence/search' // ); } // Add external attributes to 302 sticky and editorial links function extLink() { // Sticky 302 links, this fuction opens the link we send to Medium on a new tab and adds a "noopener" rel to them var stickyLinks = document.querySelectorAll('.grid-item.sticky a'); for (var i = 0; i < stickyLinks.length; i++) { /* stickyLinks[i].setAttribute('target', '_blank'); stickyLinks[i].setAttribute('rel', 'noopener'); */ } // Editorial 302 links, same here var editLinks = document.querySelectorAll( '.grid-item.category-editorial a' ); for (var i = 0; i < editLinks.length; i++) { editLinks[i].setAttribute('target', '_blank'); editLinks[i].setAttribute('rel', 'noopener'); } } // Add current year to copyright notices document.getElementById( 'js-current-year' ).textContent = new Date().getFullYear(); // Call functions after page load extLink(); //addAlly(); setTimeout(function() { //addAlly(); //ideally we should only need to run it once ↑ }, 5000); }; function closeCookieDialog (){ document.getElementById("cookie-consent").style.display = "none"; return false; } setTimeout ( function () { closeCookieDialog(); }, 15000); console.log(`%c 🚀🚀🚀 ███ █████ ███████ █████████ ███████████ █████████████ ███████████████ ███████ ███████ ███████ ┌───────────────────────────────────────────────────────────────────┐ │ │ │ Towards AI is looking for contributors! │ │ Join us in creating awesome AI content. │ │ Let's build the future of AI together → │ │ https://towardsai.net/contribute │ │ │ └───────────────────────────────────────────────────────────────────┘ `, `background: ; color: #00adff; font-size: large`); //Remove latest category across site document.querySelectorAll('a[rel="category tag"]').forEach(function(el) { if (el.textContent.trim() === 'Latest') { // Remove the two consecutive spaces (  ) if (el.nextSibling && el.nextSibling.nodeValue.includes('\u00A0\u00A0')) { el.nextSibling.nodeValue = ''; // Remove the spaces } el.style.display = 'none'; // Hide the element } }); // Add cross-domain measurement, anonymize IPs 'use strict'; //var ga = gtag; ga('config', 'G-9D3HKKFV1Q', 'auto', { /*'allowLinker': true,*/ 'anonymize_ip': true/*, 'linker': { 'domains': [ 'medium.com/towards-artificial-intelligence', 'datasets.towardsai.net', 'rss.towardsai.net', 'feed.towardsai.net', 'contribute.towardsai.net', 'members.towardsai.net', 'pub.towardsai.net', 'news.towardsai.net' ] } */ }); ga('send', 'pageview'); -->