×
Global AI computing to require eight New York’s worth of power by 2026
Written by
Published on
Join our daily newsletter for breaking news, product launches and deals, research breakdowns, and other industry-leading AI coverage
Join Now

The expansion of artificial intelligence computing infrastructure is creating unprecedented demands on power grids, with projections showing dramatic increases in energy consumption over the next few years.

Current state of AI computing: The artificial intelligence industry is entering a new phase where the focus is shifting from model training to widespread deployment of AI systems in production environments.

  • AI inference workloads, which involve running trained models in real-world applications, are becoming a major driver of computing demand
  • At least twelve new AI data centers are currently planned or under construction, each requiring approximately one gigawatt of power to operate
  • The scale of these facilities highlights the massive infrastructure requirements for supporting AI deployment at scale

Power consumption projections: By 2026, global AI processing is expected to consume approximately 40 gigawatts of power, equivalent to powering eight cities the size of New York.

  • This substantial increase in power consumption reflects the growing deployment of AI models across industries
  • The energy demands specifically relate to AI data centers, suggesting this is additional consumption beyond traditional computing infrastructure
  • These projections assume current technological approaches and efficiency levels remain relatively constant

Technological solutions: Companies are developing innovative approaches to address the growing energy and computing challenges in AI infrastructure.

  • Lightmatter, valued at $4.4 billion after receiving $400 million in venture funding, is developing chip technology using optical connections to link processors
  • These optical connections could potentially replace conventional network links in AI data centers, offering improved efficiency
  • The company’s significant valuation reflects investor confidence in the growing market for AI infrastructure solutions

Industry perspective: Thomas Graham, Lightmatter’s co-founder, emphasizes that the demand for AI computing will continue to grow primarily due to deployment rather than research needs.

  • Graham characterizes model training as research and development, while inference represents the actual deployment phase
  • The industry’s growth trajectory appears stable unless researchers develop dramatically more efficient AI algorithms
  • The focus on deployment suggests a maturing AI industry moving beyond experimental phases to practical applications

Future implications: The projected power consumption raises important questions about the sustainability of AI infrastructure and its impact on global energy systems.

  • The massive energy requirements could influence the geographic distribution of AI facilities, favoring locations with abundant power resources
  • These projections may accelerate development of more energy-efficient computing solutions
  • The energy demands could also impact the accessibility and cost structure of AI services, potentially affecting widespread adoption
Global AI computing will use 'multiple NYCs' worth of power by 2026, says founder

Recent News

Veo 2 vs. Sora: A closer look at Google and OpenAI’s latest AI video tools

Tech companies unveil AI tools capable of generating realistic short videos from text prompts, though length and quality limitations persist as major hurdles.

7 essential ways to use ChatGPT’s new mobile search feature

OpenAI's mobile search upgrade enables business users to access current market data and news through conversational queries, marking a departure from traditional search methods.

FastVideo is an open-source framework that accelerates video diffusion models

New optimization techniques reduce the computing power needed for AI video generation from days to hours, though widespread adoption remains limited by hardware costs.