DeepSeek R1 Distilled Models in Ollama: Not What You Think
Last Updated on January 31, 2025 by Editorial Team
Author(s): Kshitij Darwhekar
Originally published on Towards AI.
DeepSeek R1βs distilled models in Ollama sound like smaller versions of the original, but are they really?
This member-only story is on us. Upgrade to access all of Medium.
AI-generated using ChatGPT by AuthorDonβt have a paid Medium membership (yet)? You can read the entire article for free by clicking here with my friendβs link
DeepSeek recently introduced its models, including DeepSeek V3 and DeepSeek R1. These models have gained significant popularity in the AI community and on social media due to their impressive performance compared to models like OpenAIβs o1. Unlike OpenAIβs models, DeepSeekβs models are fully open-source and free to use.
DeepSeek Models in comparison with OpenAIβs models.Since DeepSeek models are open-source and licensed under MIT, they are free to use for both personal and commercial purposes, and you can even run them locally. However, unless you have an insanely powerful machine, you wonβt be able to run DeepSeek R1 on your local setup.
Thatβs where the smaller distilled models come in. DeepSeek has not only released the R1:671B model but also several dense models that are widely used in the research community.
All of these models are available on Hugging Face and Ollama, so you can choose whichever platform you prefer. In the next section, weβll dive deeper into these distilled models and their performance.
In simple terms, distillation is… Read the full blog for free on Medium.
Join thousands of data leaders on the AI newsletter. Join over 80,000 subscribers and keep up to date with the latest developments in AI. From research to projects and ideas. If you are building an AI startup, an AI-related product, or a service, we invite you to consider becoming aΒ sponsor.
Published via Towards AI