×
Groundbreaking AI Model Slashes Energy Use, Matches Top Performance
Written by
Published on
Join our daily newsletter for breaking news, product launches and deals, research breakdowns, and other industry-leading AI coverage
Join Now

UC Santa Cruz researchers have developed a highly energy-efficient large language model that maintains state-of-the-art performance while drastically reducing computational costs.

Key innovation: Eliminating matrix multiplication in neural networks; The researchers eliminated the most computationally expensive element of large language models, matrix multiplication, by using ternary numbers and a new communication strategy between matrices:

  • Instead of real numbers, the matrices use ternary numbers (-1, 0, 1), reducing computation to summing rather than multiplying.
  • The matrices are overlaid and only the most important operations are performed, further reducing computational overhead.
  • Time-based computation is introduced during training to maintain performance despite the reduced operations.

Impressive performance and efficiency gains: The new open-source model matches the performance of state-of-the-art models like Meta’s Llama while achieving significant energy and cost savings:

  • On standard GPUs, the model uses 10 times less memory and operates 25% faster than other models.
  • Custom hardware designed for the model enables it to surpass human-readable throughput on just 13 watts, over 50 times more efficient than GPUs.

Implications for AI accessibility and sustainability: The drastic reduction in energy consumption and memory requirements opens up new possibilities for large language models:

  • Powerful AI models could potentially run at full capacity on memory-constrained devices like smartphones.
  • The massive energy and monetary costs associated with running models like ChatGPT could be significantly reduced, making AI more sustainable and accessible.

Analyzing deeper: While the researchers’ approach represents a significant breakthrough in efficiency, it remains to be seen how well the techniques scale to even larger models and more complex tasks. Further optimization of the custom hardware could yield even greater gains. This work highlights the importance of rethinking fundamental aspects of AI algorithms and hardware to make the technology more sustainable and widely accessible as it continues to advance.

Researchers run high-performing large language model on the energy needed to power a lightbulb

Recent News

Robin Williams’ daughter Zelda slams OpenAI’s Ghibli-style images amid artistic and ethical concerns

Robin Williams' daughter condemns OpenAI's AI-generated Ghibli-style images, highlighting both environmental costs and the contradiction with Miyazaki's well-documented opposition to artificial intelligence in creative work.

AI search tools provide wrong answers up to 60% of the time despite growing adoption

Independent testing reveals AI search tools frequently provide incorrect information, with error rates ranging from 37% to 94% across major platforms despite their growing popularity as Google alternatives.

Have at it! LessWrong forum encourages “crazy” ideas to solve AI safety challenges

The online community fosters unorthodox thinking about AI development risks, challenging orthodox research methods that may overlook critical safety solutions.