Llama 3.2 Vision Review
Author(s): Barhoumi Mosbeh Originally published on Towards AI. Ollama multi-modal Ollama has just announced its official support for the Llama 3.2 Vision models. The Llama 3.2 Vision models come in two sizes: 11 billion and 90 billion parameters. In this article, I …
Late Chunking In Long Context Embedding Models
Author(s): Barhoumi Mosbeh Originally published on Towards AI. source In a previous article, we looked at contextual retrieval from Anthropic, which is their context enhancement technique for improving RAG systems. But thereβs another technique called late chunking in long-context embedding models, which …
How Can GPTs Interact with Computers? OmniParser Explained
Author(s): Barhoumi Mosbeh Originally published on Towards AI. Microsoft has silently released OmniParser, an open-source tool designed to convert screenshots into structured, easy-to-interpret elements for Vision Agents. The goal of this tool is to advance the emerging field of enabling large language …
Fine-Tune LLMs with Unsloth
Author(s): Barhoumi Mosbeh Originally published on Towards AI. unsloth Why Fine-Tune When We Have RAG? Itβs a question I see a lot β with RAG (Retrieval-Augmented Generation) becoming increasingly popular, why bother with fine-tuning at all? While RAG is fantastic for many …
RAG From Scratch
Author(s): Barhoumi Mosbeh Originally published on Towards AI. Iβm working as a machine learning engineer, and I frequently use Claude or ChatGPT to help me write code. However, in some cases, the model starts to repeat itself or hallucinate, especially during complex …
Anthropicβs New RAG Approach
Author(s): Barhoumi Mosbeh Originally published on Towards AI. The Rise of LLMs LLMs are super-powerful tools. I remembered the first time ChatGPT was released, I was saying hell no, I will lose my job (before even getting one), butβ¦ then I saw …