Putting The World’s Largest AI Supercomputer into Perspective
Last Updated on September 18, 2024 by Editorial Team
Author(s): Ignacio de Gregorio
Originally published on Towards AI.
Top highlight
This member-only story is on us. Upgrade to access all of Medium.
Source: AI-generated by Author using FluxElon Musk has just announced that xAI has finally connected their Colossus cluster, a 100,000-install base NVIDIA H100 GPU accelerated computer.
The biggest AI computer the world has ever seen (and it’s not even close) boasts some of the most astonishing numbers one can fathom and helps us toy with the idea of how large the next generation of models will be.
And let me tell you, they are huge!
Get news like this before anyone else by subscribing to my newsletter, the place where analysts and strategists get answers to AI’s most pressing questions.
The newsletter to stay ahead of the curve in AI
thetechoasis.beehiiv.com
AI workloads require an insane amount of computation for them to be viable. The reason is that they are huge digital files boasting billions of parameters occupying double or triple figures in GigaBytes and, in some cases, like frontier models, TeraBytes in size.
To make matters worse, due to their structure, these models are queried every single time they need to predict something, which, in text terms, means they are queried for every new word they predict.
Therefore, to prevent latency from becoming unbearable, we… Read the full blog for free on Medium.
Join thousands of data leaders on the AI newsletter. Join over 80,000 subscribers and keep up to date with the latest developments in AI. From research to projects and ideas. If you are building an AI startup, an AI-related product, or a service, we invite you to consider becoming a sponsor.
Published via Towards AI