Cerebras Systems is a U.S.-based AI compute company that designs and builds wafer-scale processors and integrated AI supercomputer systems for training and running very large models. Founded in 2016 and headquartered in Sunnyvale, California, the company takes a fundamentally different approach from GPU-based architectures: instead of stitching together thousands of small accelerators, Cerebras manufactures an entire 300mm wafer as a single chip—the Wafer-Scale Engine (WSE)—and packages it into turnkey systems (the CS-series) and a managed inference/training cloud.
The company’s third-generation system, CS-3, is built around the WSE-3 processor fabricated on TSMC’s 5nm node and designed specifically for large language models and other transformer workloads at “mega-scale.” CS-3 and the Cerebras Inference Cloud are positioned as an alternative for organizations that need very large models or high token throughput but face cost, power, or supply constraints with conventional GPU clusters. Cerebras targets national AI infrastructure (“Cerebras for Nations”), research institutions, and enterprises in sectors such as life sciences, energy, and financial services, where faster time-to-train and high-throughput inference can be economically meaningful.
Financially, Cerebras remains a private, venture-backed company. It filed for a U.S. IPO in 2024 but subsequently chose to raise a large private Series G round instead and defer listing. In September 2025, the company closed a $1.1 billion financing at an $8.1 billion post-money valuation and is using the capital to scale manufacturing and expand its global data center footprint from six sites toward as many as 15, including participation in the Stargate AI data center hub in the United Arab Emirates.