Name: Towards AI Legal Name: Towards AI, Inc. Description: Towards AI is the world's leading artificial intelligence (AI) and technology publication. Read by thought-leaders and decision-makers around the world. Phone Number: +1-650-246-9381 Email: pub@towardsai.net
228 Park Avenue South New York, NY 10003 United States
Website: Publisher: https://towardsai.net/#publisher Diversity Policy: https://towardsai.net/about Ethics Policy: https://towardsai.net/about Masthead: https://towardsai.net/about
Name: Towards AI Legal Name: Towards AI, Inc. Description: Towards AI is the world's leading artificial intelligence (AI) and technology publication. Founders: Roberto Iriondo, , Job Title: Co-founder and Advisor Works for: Towards AI, Inc. Follow Roberto: X, LinkedIn, GitHub, Google Scholar, Towards AI Profile, Medium, ML@CMU, FreeCodeCamp, Crunchbase, Bloomberg, Roberto Iriondo, Generative AI Lab, Generative AI Lab VeloxTrend Ultrarix Capital Partners Denis Piffaretti, Job Title: Co-founder Works for: Towards AI, Inc. Louie Peters, Job Title: Co-founder Works for: Towards AI, Inc. Louis-François Bouchard, Job Title: Co-founder Works for: Towards AI, Inc. Cover:
Towards AI Cover
Logo:
Towards AI Logo
Areas Served: Worldwide Alternate Name: Towards AI, Inc. Alternate Name: Towards AI Co. Alternate Name: towards ai Alternate Name: towardsai Alternate Name: towards.ai Alternate Name: tai Alternate Name: toward ai Alternate Name: toward.ai Alternate Name: Towards AI, Inc. Alternate Name: towardsai.net Alternate Name: pub.towardsai.net
5 stars – based on 497 reviews

Frequently Used, Contextual References

TODO: Remember to copy unique IDs whenever it needs used. i.e., URL: 304b2e42315e

Resources

Our 15 AI experts built the most comprehensive, practical, 90+ lesson courses to master AI Engineering - we have pathways for any experience at Towards AI Academy. Cohorts still open - use COHORT10 for 10% off.

Publication

LAI #79: How LLMs Learn, Vertical Model Growth, and Smarter Evaluation
Artificial Intelligence   Latest   Machine Learning

LAI #79: How LLMs Learn, Vertical Model Growth, and Smarter Evaluation

Author(s): Towards AI Editorial Team

Originally published on Towards AI.

Featured

Good morning, AI enthusiasts,

This week’s issue is about getting back to first principles. We’re diving into how LLMs actually learn: what’s under the hood, and why it matters when you’re building or deploying anything serious. It’s also the perfect lead-in to the launch of our 10-Hour LLM Primer, designed to cut through the fog and help you make real technical decisions, fast.

In the curated section, you’ll find a breakdown of how Segment Any Text improves RAG structure, a primer on MCP and how it powers secure automation, and a look at why vertical LLMs are outpacing general-purpose models by 10x. You also have a tutorial on agent tracing with Langfuse, classification metrics, and a community-built Llama router.

And this week again, we have an amazing guest post on Miguel Otero Pedrido’s Neural Maze newsletter comparing Retrieval-Augmented Generation (RAG) and Cache-Augmented Generation (CAG). The findings were clear: CAG is far more efficient post-cache processing nearly 10× fewer tokens per query, and more disciplined in refusing to answer when information is out-of-scope. RAG, however, produced more detailed in-scope answers in direct comparisons. Cost-wise, CAG breaks even fast; in our tests, just six queries on a small cache made it cheaper than RAG. The full piece includes evaluation methods, examples, and cost breakdowns — you can read it here!

We’d love to hear your thoughts if you do check it out. Drop a quick comment if you found it helpful or if there’s more you’d like us to explore on these systems.

What’s AI Weekly

The lack of a fundamental understanding of how LLMs are made leads to people both underestimating AI capability in some areas (and therefore not pushing AI usage to its potential) and overestimating AI capability in other areas (which can lead to user frustration and premature rejection of these tools, or worse to AI mistakes that slip into work outputs). So this week in What’s AI, I am diving into how these models are designed and built. While a thorough understanding of the fundamental design of LLMs is no longer always necessary to utilize and build LLM products effectively, knowing some theoretical concepts, like LLM’s training objective, training data, how they generate words, and understanding embeddings, can make a big difference in getting the most out of them. So, read the complete article here or watch the video on YouTube.

— Louis-François Bouchard, Towards AI Co-founder & Head of Community

Introducing the 10-Hour Video Primer: Building and Operating LLMs

If you’ve ever gone five blog posts deep and still ended up with “it depends,” this course is for you.

LAI #79: How LLMs Learn, Vertical Model Growth, and Smarter Evaluation

The 10-Hour LLM Primer helps you cut through the noise and make real technical decisions — fast.

In five focused video sessions, you’ll learn:

  • What transformers actually do (and don’t)
  • When to use prompting, RAG, fine-tuning, or none of them
  • How real-world agent workflows are structured
  • What breaks at scale — and how to fix it
  • How to evaluate and maintain production-grade LLM systems

It’s bingeable in a day. Or treat it like an AI docuseries.

No fluff. No hype. Just the clarity we usually deliver in $25K+ enterprise workshops.

Launch price: $199, includes all future updates (dropping soon)

Price will go up once new lessons drop — grab it now.

👉 Course details here

Learn AI Together Community Section!

Featured Community post from the Discord

Jasonfpv_ has built a FlexLlama repo, a lightweight, extensible, and user-friendly self-hosted tool that efficiently runs multiple llama.cpp server instances with OpenAI v1 API compatibility. It’s designed to manage multiple models across different GPUs, making it a powerful solution for local AI development and deployment. Check it out on GitHub and support a fellow community member. If you have any questions or feedback, share them in the thread!

AI poll of the week!

70% of you voted for hybrid architectures, and that says a lot. We’re moving past the “bigger is better” phase of AI. Instead of one model to rule them all, builders are betting on coordination over domination: a strong generalist guiding a swarm of specialists. It’s more efficient, more modular, and maybe… more human.

What’s interesting is that very few chose “one giant model.” That’s a shift. The market may still reward massive models today, but the real-world is clearly thinking in systems, not silos.

What’s a real use case where you think the hybrid model (big generalist + small experts) would clearly outperform any single model alone? Share your thoughts in the thread!

Collaboration Opportunities

The Learn AI Together Discord community is flooding with collaboration opportunities. If you are excited to dive into applied AI, want a study partner, or even want to find a partner for your passion project, join the collaboration channel! Keep an eye on this section, too — we share cool opportunities every week!

1. Safar4352 is building a project from scratch and needs a project partner. The project will be Python-based, so if you want to understand GenAI and LLMs in depth, connect with him in the thread!

2. Tranquil_dolphin_27432 wants to collaborate with someone who typically sees AI as their passion project. If this sounds like you, reach out in the thread!

3. Zabka___ is looking for a small community of 10–30 high schoolers, students, or researchers to learn machine learning together. Their focus is on theoretical aspects, such as linear algebra and statistics. If you want to expand your knowledge in this space, contact them in the thread!

Meme of the week!

Meme shared by ghost_in_the_machine

TAI Curated Section

Article of the week

LangChain + Segment Any Text + RAG = The Key to Understanding Your Documents By Gao Dalie (高達烈)

This article addresses the common RAG system issue of semantic fragmentation from standard text splitting. It proposes using a Segment Any Text (SAT) approach, implemented via the ContextGem framework, to ensure text chunks are semantically whole before processing. The author demonstrates how to build a powerful agent that combines ContextGem’s structured data extraction with a traditional LangChain RAG pipeline using FAISS and OpenAI. This dual approach allows for more accurate and context-aware responses by querying both structured insights and unstructured document text.

Our must-read articles

1. MCP -The Golden Key for AI Automation By Alex Punnen

This blog explains the Model Context Protocol (MCP), a standard for letting LLMs interact with external APIs. It breaks down how MCP works by allowing an LLM to generate a JSON-RPC call for a specific tool after receiving its description. The author uses a simple calculator example to demonstrate the process. A significant portion covers MCP’s secure, OAuth2-based authorization, using the Zerodha Kite financial platform to illustrate how an authenticated session can be established without the client ever handling sensitive credentials.

2. $100B and Rising: Why Vertical LLMs Are Growing 10x Faster Than General AI Models By R. Thompson (PhD)

This article details the growing industry shift from general-purpose AI to vertical LLMs, which are tailored for specific sectors like finance, healthcare, and law. It highlights how general models can fall short in nuanced, regulated environments, whereas specialized models like BloombergGPT and Med-PaLM 2 provide superior accuracy, compliance, and contextual depth. The piece also covers the technical methods, such as RAG and QLoRA, that facilitate the development of these domain-specific models, which are projected to see significant market growth.

3. Monitor and Evaluate Open AI SDK Agents using Langfuse By Steve George

The author presents a method for monitoring an agentic workflow built with the OpenAI SDK. It outlines a three-agent system: an input guardrail to block sensitive data, an assist agent to generate responses, and a validation agent for fact-checking. The piece demonstrates how to use OpenTelemetry to send trace data to Langfuse for visualization. It also covers programmatically accessing traces with the Langfuse SDK to create analytical plots for evaluation.

4. The Essential Guide to Model Evaluation Metrics for Classification By Ayo Akinkugbe

This piece provides a comprehensive overview of evaluation metrics for classification models, starting with the foundational confusion matrix. It details common metrics like accuracy, precision, and recall, explaining their ideal use cases and limitations, particularly with imbalanced data. The guide also covers more advanced measures such as ROC AUC, Log Loss, and the Phi Coefficient, using practical case studies to illustrate their application. It stresses the importance of selecting metrics based on specific project needs and business goals.

If you are interested in publishing with Towards AI, check our guidelines and sign up. We will publish your work to our network if it meets our editorial policies and standards.

Join thousands of data leaders on the AI newsletter. Join over 80,000 subscribers and keep up to date with the latest developments in AI. From research to projects and ideas. If you are building an AI startup, an AI-related product, or a service, we invite you to consider becoming a sponsor.

Published via Towards AI


Take our 90+ lesson From Beginner to Advanced LLM Developer Certification: From choosing a project to deploying a working product this is the most comprehensive and practical LLM course out there!

Towards AI has published Building LLMs for Production—our 470+ page guide to mastering LLMs with practical projects and expert insights!


Discover Your Dream AI Career at Towards AI Jobs

Towards AI has built a jobs board tailored specifically to Machine Learning and Data Science Jobs and Skills. Our software searches for live AI jobs each hour, labels and categorises them and makes them easily searchable. Explore over 40,000 live jobs today with Towards AI Jobs!

Note: Content contains the views of the contributing authors and not Towards AI.