Cerebras Systems is dramatically expanding its AI inference capacity and strategically positioning itself to challenge Nvidia’s market dominance in the artificial intelligence infrastructure space. By adding six new data centers across North America and Europe and securing partnerships with major tech platforms, Cerebras is betting on the growing demand for high-speed AI inference services as enterprises seek faster alternatives to traditional GPU solutions. This expansion represents a significant development in the evolving AI hardware landscape, potentially reshaping how businesses access and deploy artificial intelligence capabilities.
The big picture: Cerebras Systems announced a massive twentyfold increase in its AI inference capacity, adding six new data centers across North America and Europe to deliver over 40 million tokens per second.
Strategic partnerships: Cerebras has secured integrations with two significant platforms that will expand its market reach.
Technical advantages: The company is positioning its Wafer-Scale Engine (WSE-3) processor as significantly faster than GPU-based alternatives for specific AI workloads.
Behind the numbers: Cerebras is pursuing a dual strategy of superior speed and cost-effectiveness.
Why this matters: With 85% of its inference capacity located in the United States, Cerebras is advancing domestic AI infrastructure at a time when processing capabilities are becoming a critical resource for businesses adopting AI technologies.
What they’re saying: “This year, our goal is to truly satisfy all the demand and all the new demand we expect will come online as a result of new models like Llama 4 and new DeepSeek models,” said James Wang of Cerebras.