×
AI Startup Cerebras Challenges Nvidia With 20x Faster Chips
Written by
Published on
Join our daily newsletter for breaking news, product launches and deals, research breakdowns, and other industry-leading AI coverage
Join Now

AI chip market disruption: Cerebras Systems, a US startup, is making waves in the AI chip industry with its Wafer Scale Engine (WSE) processors, positioning itself as a formidable challenger to Nvidia’s dominance.

  • Cerebras claims its chips can execute AI workloads up to 20 times faster and at one-fifth the cost compared to Nvidia’s GPUs, including the popular H100 model.
  • The company has introduced a new service called Cerebras Inference, designed specifically for running generative AI programs.
  • Cerebras’ chips have demonstrated impressive performance in running Meta’s Llama 3.1 language model, producing 1,800 tokens per second for the 8 billion parameter version and 450 tokens per second for the 70 billion parameter version.

Performance and pricing advantages: Cerebras asserts that its technology outperforms other major cloud providers and offers more competitive pricing for AI inference tasks.

  • The company claims its performance surpasses that of cloud giants like AWS and Azure for similar AI workloads.
  • Cerebras’ pricing model is notably aggressive, charging 10 cents per million tokens for Llama 3.1 8B and 60 cents per million for Llama 3.1 70B.
  • This pricing structure represents a significant cost advantage compared to OpenAI’s rates, which range from $2.50 to $15 per million tokens.

Technological advancements: The latest WSE-3 chip from Cerebras boasts impressive specifications that highlight the company’s focus on pushing the boundaries of AI chip technology.

  • The WSE-3 chip contains a staggering 4 trillion transistors and 900,000 AI cores.
  • Cerebras claims its chip offers 7,000 times more memory bandwidth than Nvidia’s H100, a key factor in AI processing capabilities.
  • The company’s chips are integrated into CS-3 hardware systems, which are priced at “a couple million per system,” compared to around $30,000 for an H100 GPU.

Market strategy and expansion: Cerebras is actively working to increase accessibility to its advanced chip technology, aiming to broaden its market presence.

  • The company is exploring partnerships with cloud providers to make its chips more widely available to potential customers.
  • By offering its technology through cloud services, Cerebras could lower the barrier to entry for organizations interested in leveraging its high-performance AI chips without the need for significant upfront hardware investments.

Competitive landscape: While Cerebras presents compelling performance and cost figures, it’s important to note the dynamic nature of the AI chip market.

  • The comparisons made by Cerebras are primarily against Nvidia’s current generation of chips, particularly the H100.
  • Nvidia has already announced its next-generation Blackwell architecture chips, slated for release later this year, which promise substantial performance improvements over the H100.
  • The impending release of Nvidia’s new chips could potentially narrow or alter the performance gap claimed by Cerebras.

Implications for the AI industry: Cerebras’ entry into the market with its high-performance chips could have far-reaching effects on the AI landscape and its applications.

  • Increased competition in the AI chip market may lead to accelerated innovation and potentially lower costs for AI processing, benefiting various industries relying on AI technologies.
  • The availability of more powerful and cost-effective AI chips could enable the development of more sophisticated AI models and applications, potentially advancing fields such as natural language processing, computer vision, and scientific research.
  • However, the true impact of Cerebras’ technology will depend on its ability to scale production, secure partnerships, and maintain its performance edge in a rapidly evolving market dominated by established players like Nvidia.
Startup Goes After Nvidia With Giant AI Chips Built for Speed

Recent News

Propaganda is everywhere, even in LLMS — here’s how to protect yourself from it

Recent tragedy spurs examination of AI chatbot safety measures after automated responses proved harmful to a teenager seeking emotional support.

How Anthropic’s Claude is changing the game for software developers

AI coding assistants now handle over 10% of software development tasks, with major tech firms reporting significant time and cost savings from their deployment.

AI-powered divergent thinking: How hallucinations help scientists achieve big breakthroughs

Meta's new AI model combines powerful performance with unusually permissive licensing terms for businesses and developers.