×
Written by
Published on
Written by
Published on
Join our daily newsletter for breaking news, product launches and deals, research breakdowns, and other industry-leading AI coverage
Join Now

AI chip market disruption: Cerebras Systems, a US startup, is making waves in the AI chip industry with its Wafer Scale Engine (WSE) processors, positioning itself as a formidable challenger to Nvidia’s dominance.

  • Cerebras claims its chips can execute AI workloads up to 20 times faster and at one-fifth the cost compared to Nvidia’s GPUs, including the popular H100 model.
  • The company has introduced a new service called Cerebras Inference, designed specifically for running generative AI programs.
  • Cerebras’ chips have demonstrated impressive performance in running Meta’s Llama 3.1 language model, producing 1,800 tokens per second for the 8 billion parameter version and 450 tokens per second for the 70 billion parameter version.

Performance and pricing advantages: Cerebras asserts that its technology outperforms other major cloud providers and offers more competitive pricing for AI inference tasks.

  • The company claims its performance surpasses that of cloud giants like AWS and Azure for similar AI workloads.
  • Cerebras’ pricing model is notably aggressive, charging 10 cents per million tokens for Llama 3.1 8B and 60 cents per million for Llama 3.1 70B.
  • This pricing structure represents a significant cost advantage compared to OpenAI’s rates, which range from $2.50 to $15 per million tokens.

Technological advancements: The latest WSE-3 chip from Cerebras boasts impressive specifications that highlight the company’s focus on pushing the boundaries of AI chip technology.

  • The WSE-3 chip contains a staggering 4 trillion transistors and 900,000 AI cores.
  • Cerebras claims its chip offers 7,000 times more memory bandwidth than Nvidia’s H100, a key factor in AI processing capabilities.
  • The company’s chips are integrated into CS-3 hardware systems, which are priced at “a couple million per system,” compared to around $30,000 for an H100 GPU.

Market strategy and expansion: Cerebras is actively working to increase accessibility to its advanced chip technology, aiming to broaden its market presence.

  • The company is exploring partnerships with cloud providers to make its chips more widely available to potential customers.
  • By offering its technology through cloud services, Cerebras could lower the barrier to entry for organizations interested in leveraging its high-performance AI chips without the need for significant upfront hardware investments.

Competitive landscape: While Cerebras presents compelling performance and cost figures, it’s important to note the dynamic nature of the AI chip market.

  • The comparisons made by Cerebras are primarily against Nvidia’s current generation of chips, particularly the H100.
  • Nvidia has already announced its next-generation Blackwell architecture chips, slated for release later this year, which promise substantial performance improvements over the H100.
  • The impending release of Nvidia’s new chips could potentially narrow or alter the performance gap claimed by Cerebras.

Implications for the AI industry: Cerebras’ entry into the market with its high-performance chips could have far-reaching effects on the AI landscape and its applications.

  • Increased competition in the AI chip market may lead to accelerated innovation and potentially lower costs for AI processing, benefiting various industries relying on AI technologies.
  • The availability of more powerful and cost-effective AI chips could enable the development of more sophisticated AI models and applications, potentially advancing fields such as natural language processing, computer vision, and scientific research.
  • However, the true impact of Cerebras’ technology will depend on its ability to scale production, secure partnerships, and maintain its performance edge in a rapidly evolving market dominated by established players like Nvidia.
Startup Goes After Nvidia With Giant AI Chips Built for Speed

Recent News

71% of Investment Bankers Now Use ChatGPT, Survey Finds

Investment banks are increasingly adopting AI, with smaller firms leading the way and larger institutions seeing higher potential value per employee.

Scientists are Designing “Humanity’s Last Exam” to Assess Powerful AI

The unprecedented test aims to assess AI capabilities across diverse fields, from rocketry to philosophy, with experts submitting challenging questions beyond current benchmarks.

Hume Launches ‘EVI 2’ AI Voice Model with Emotional Responsiveness

The new AI voice model offers improved naturalness, faster response times, and customizable voices, potentially enhancing AI-human interactions across various industries.