China’s AI firms are cleverly innovating around chip bans

1 min read Original article ↗

Tweaks to software blunt the shortage of powerful hardware

TODAY’S TOP artificial-intelligence (AI) models rely on large numbers of cutting-edge processors known as graphics processing units (GPUs). Most Western companies have no trouble acquiring them. Llama 3, the newest model from Meta, a social-media giant, was trained on 16,000 H100 GPUs from Nvidia, an American chipmaker. Meta plans to stockpile 600,000 more before year’s end. XAI, a startup backed by Elon Musk, has built a data centre in Memphis powered by 100,000 H100s. And though OpenAI, the other big model-maker, is tight-lipped about its GPU stash, it had its latest processors hand-delivered by Jensen Huang, Nvidia’s boss, in April.

This article appeared in the Science & technology section of the print edition under the headline “Miniature model-building”

From the September 21st 2024 edition

Discover stories from this section and more in the list of contents

Explore the edition