×
Why reasoning models like DeepSeek-R1 are increasing (not decreasing) demand for GPUs
Written by
Published on
Join our daily newsletter for breaking news, product launches and deals, research breakdowns, and other industry-leading AI coverage
Join Now

The rise of advanced AI reasoning models like DeepSeek-R1 has transformed the computational landscape for AI infrastructure providers. Together AI has secured significant funding to meet growing enterprise demands for high-performance AI model deployment and reasoning capabilities.

Recent funding and growth: Together AI has raised $305 million in Series B funding led by General Catalyst and Prosperity7, marking a significant expansion of its AI deployment platform.

  • The company has experienced 6X year-over-year growth with over 450,000 registered developers
  • Notable customers include AI startups like Krea AI, Captions, and Pika Labs
  • The platform now supports multiple AI modalities including language, reasoning, images, audio, and video

Infrastructure demands: Contrary to initial expectations, DeepSeek-R1 and similar reasoning models are driving increased infrastructure requirements.

  • DeepSeek-R1’s 671 billion parameters require distribution across multiple servers
  • User requests can last up to 2-3 minutes, necessitating dedicated computational resources
  • Together AI has introduced “reasoning clusters” offering 128 to 2,000 chips for optimal performance

Key applications: Reasoning models are finding practical applications across various use cases.

  • Enhanced coding capabilities through problem decomposition
  • Reduced model hallucinations through verification processes
  • Quality improvements in non-reasoning models
  • Self-improvement capabilities using reinforcement learning

Technical innovations: Together AI has made strategic moves to enhance its platform capabilities.

  • Acquisition of CodeSandbox for lightweight VM deployment
  • Implementation of Nvidia’s Blackwell GPUs, offering 2X performance improvement
  • Optimization of inference speeds, achieving 85 tokens per second compared to Azure’s 7 tokens per second

Competitive landscape: Together AI operates in an increasingly crowded market for AI infrastructure.

  • Major cloud providers Microsoft, AWS, and Google offer competing platforms
  • Specialized AI infrastructure providers like Groq and Samba Nova target similar markets
  • Together AI differentiates through full-stack offerings and superior performance metrics

Market implications: The increasing demand for AI infrastructure, particularly for reasoning models, suggests a continuing trend toward more sophisticated and resource-intensive AI applications, potentially creating new opportunities and challenges for infrastructure providers and enterprises alike.

Together AI’s $305M bet: Reasoning models like DeepSeek-R1 are increasing, not decreasing, GPU demand

Recent News

AI courses from Google, Microsoft and more boost skills and résumés for free

As AI becomes critical to business decision-making, professionals can enhance their marketability with free courses teaching essential concepts and applications without requiring technical backgrounds.

Veo 3 brings audio to AI video and tackles the Will Smith Test

Google's latest AI video generation model introduces synchronized audio capabilities, though still struggles with realistic eating sounds when depicting the celebrity in its now-standard benchmark test.

How subtle biases derail LLM evaluations

Study finds language models exhibit pervasive positional preferences and prompt sensitivity when making judgments, raising concerns for their reliability in high-stakes decision-making contexts.