SambaNova | The Fastest AI Inference Platform & Hardware

1 min read Original article ↗

DeepSeek

We support the groundbreaking DeepSeek models, including the 671-billion-parameter DeepSeek-R1, which excels in coding, reasoning, and mathematics at a fraction of the cost of other models.

On our SambaNova RDU, DeepSeek-R1 achieves remarkable speeds of up to 200 tokens/second, as measured independently by Artificial Analysis.

Llama

As a launch partner for Meta's Llama 4 series, we've been at the forefront of open-source AI innovation. SambaNova Cloud was the first platform to support all three variants of Llama 3.1 (8B, 70B, and 405B) with fast inference.

We are excited to work with Meta to deliver fast inference on both Scout and Maverick models.

OpenAI

OpenAI's Whisper model — supported on our SambaNova RDU — has become a cornerstone for audio-based AI applications. By leveraging our platform's fast processing speeds, developers can unlock new use cases in voice-based AI agents to create more immersive experiences.