How DeepSeek Cuts AI Costs: From Homegrown Tech to Desert Power
Last Updated on January 30, 2025 by Editorial Team
Author(s): Don Lim
Originally published on Towards AI.
This member-only story is on us. Upgrade to access all of Medium.
DeepSeek doesnβt rely on NVIDIA chips for inference. They also donβt solely use Western AI development frameworks like PyTorch and TensorFlow. This was a big shock to me as well, as I hadnβt paid too much attention to Chinaβs software industry until recently. Iβve been developing software for several decades now, but the only software from China that I used was Filmora and face beautifying apps.
A Blue Whale Doll in My Hand β This doll was in my house for a long time. I donβt know where it came from.The Magnificent 7 (Mag7) companies can utilize similar optimization techniques as DeepSeek and save on electricity in remote locations. However, they wonβt be able to escape NVIDIAβs influence, since they all use critical AI development frameworks called PyTorch and TensorFlow, which are currently dependent on NVIDIA chips.
Almost all Large Language Models (LLMs) use PyTorch (developed by Facebook) and TensorFlow (developed by Google) as development platforms or frameworks. These frameworks are specifically designed to handle the intensive vector and matrix calculations required for training and running large-scale neural networks. The developers of PyTorch and TensorFlow have been working with NVIDIA for… Read the full blog for free on Medium.
Join thousands of data leaders on the AI newsletter. Join over 80,000 subscribers and keep up to date with the latest developments in AI. From research to projects and ideas. If you are building an AI startup, an AI-related product, or a service, we invite you to consider becoming aΒ sponsor.
Published via Towards AI