Zubnet AI学习Wiki › SambaNova
公司

SambaNova

又名: SN40L chip, ultra-fast inference
设计为 AI 工作负载专门打造的自定义芯片(RDUs)的 AI 硬件公司。他们的 SambaNova Cloud 提供可得最快的推理速度之一,在“速度优先”的 AI 服务上和 Groq 竞争。

为什么重要

SambaNova 重要是因为 AI 算力不应只有 NVIDIA 一家独大,得有人证明专门为 AI 打造的芯片能在真实市场竞争,不只是在研究论文里。他们的 RDU 架构证明了当你专门为神经网络工作负载设计硅片时,能有有意义的性能提升,他们的云推理服务让开发者尝到后 GPU AI 基础设施的味道。无论 SambaNova 自己能不能成为主导替代,他们施加的竞争压力 — 和 Groq、Cerebras、云供应商的自定义芯片一起 — 对一个负担不起永久硬件单一文化的产业来说是健康的。

Deep Dive

SambaNova was founded in 2017 by Rodrigo Liang, Christopher Ré, and Kunle Olukotun at Stanford University. Ré is a MacArthur Fellow and one of the most influential figures in modern machine learning (his later work on state-space models and data-centric AI would spawn multiple companies), while Olukotun is a pioneer in chip architecture who helped develop the concept of multicore processors. The founding thesis was straightforward but ambitious: NVIDIA's GPUs, while dominant, were not designed specifically for AI workloads. A chip built from the ground up for AI — optimizing for the specific dataflow patterns, memory access requirements, and parallelism that neural networks demand — could deliver dramatically better performance per watt and per dollar. SambaNova raised over $1.1 billion in venture funding, including a massive $676 million Series D in 2021, making it one of the best-funded AI hardware startups in history.

The Reconfigurable Dataflow Unit

SambaNova's core technology is the Reconfigurable Dataflow Unit (RDU), most recently the SN40L chip. Unlike GPUs, which execute instructions in a relatively traditional fetch-decode-execute cycle adapted for parallel workloads, the RDU is a dataflow architecture — computation happens as data flows through the chip, with the processing pattern reconfigured for each model rather than following a fixed instruction stream. In theory, this eliminates many of the inefficiencies inherent in running neural networks on general-purpose hardware. The SN40L specifically was designed with a three-tiered memory hierarchy that can hold much larger models in on-chip memory than a typical GPU, reducing the expensive off-chip memory transfers that bottleneck inference. SambaNova has claimed that their architecture can serve models like Llama 2 70B and Llama 3.1 405B at speeds that rival or exceed NVIDIA's fastest offerings, and independent benchmarks have generally supported these claims for specific workloads.

The Pivot to Cloud Inference

SambaNova's business model has undergone a significant evolution. Initially, the company sold on-premise hardware appliances — full-rack systems running RDUs — to large enterprises and government agencies. These DataScale systems found customers in national laboratories, financial institutions, and defense applications where data sovereignty and performance mattered more than cost. But the enterprise hardware market proved challenging: long sales cycles, complex integration, and customers who were often not ready to deploy AI at the scale that justified custom hardware. In 2023, SambaNova pivoted toward cloud-based inference, launching SambaNova Cloud as an API service where developers could access models running on RDUs without buying hardware. This put them in direct competition with Groq, another AI chip startup that had made "fastest inference" its calling card, as well as the inference offerings from major cloud providers.

Speed as a Feature

The cloud inference pivot crystallized SambaNova's positioning: speed as the primary selling point. Their API consistently delivers some of the fastest tokens-per-second rates in the industry, particularly for larger models where the memory hierarchy advantages of the RDU architecture are most pronounced. They offered free tier access to popular open-source models like Llama and Qwen, using speed as the hook to attract developers who would then convert to paid usage. This strategy mirrored what Groq had done with their LPU chips, creating a two-horse race in the "fast inference" niche. For developers building latency-sensitive applications — real-time agents, voice assistants, interactive coding tools — the speed difference is not just a nice benchmark number but a genuine product differentiator that affects user experience.

The NVIDIA Problem

Every AI chip startup ultimately faces the same challenge: NVIDIA's ecosystem is extraordinarily deep, and CUDA is the de facto standard for AI development. SambaNova has mitigated this by focusing on inference rather than training — inference workloads are more standardized and less dependent on CUDA's full software stack — and by supporting popular open-source models out of the box so developers don't need to learn new tooling. But the company is swimming against a powerful current. NVIDIA continuously improves its own inference performance, and cloud providers are building custom inference chips (Google's TPUs, Amazon's Inferentia and Trainium, Microsoft's Maia). SambaNova's path to long-term success likely requires either a sustained performance advantage large enough to justify the ecosystem switching cost, or a partnership with a major cloud provider that bundles RDU-powered inference into an existing platform. With over a billion dollars raised and real technology behind the claims, SambaNova has a genuine shot — but the window to prove the thesis is narrowing as competition intensifies.

相关概念

In The News

Intel-SambaNova Blueprint: Finally, a Realistic Take on AI 基础设施
Apr 09, 2026
← 所有术语
← Runway Sampling →
ESC