×
New Research Suggests AI Models Can’t Learn as They Go Along
Written by
Published on
Join our daily newsletter for breaking news, product launches and deals, research breakdowns, and other industry-leading AI coverage
Join Now

AI models face limitations in continuous learning: Recent research reveals that current artificial intelligence systems, including large language models like ChatGPT, are unable to update and learn from new data after their initial training phase.

  • A study by researchers at the University of Alberta in Canada has uncovered an inherent problem in the design of AI models that prevents them from learning continuously.
  • This limitation forces tech companies to spend billions of dollars training new models from scratch when new data becomes available.
  • The inability to incorporate new knowledge after initial training has been a long-standing concern in the AI industry.

Understanding the problem: The issue stems from the way neural networks, which form the basis of most modern AI systems, are designed and trained.

  • AI models typically go through distinct phases: training, where artificial neurons are fine-tuned to reflect a given dataset, and usage, where the model responds to new inputs.
  • Once the training phase is complete, the model’s neurons are set and cannot update or learn from new data.
  • This limitation means that large AI models must be retrained entirely when new information becomes available, a process that can be prohibitively expensive.

Research findings: The study conducted by Shibhansh Dohare and his colleagues tested whether common AI models could be adapted for continuous learning.

  • The team found that AI systems quickly lose the ability to learn new information, with a large number of artificial neurons becoming inactive or “dead” after exposure to new data.
  • In their experiments, after a few thousand retraining cycles, the networks performed poorly and appeared unable to learn.
  • This problem was observed across various learning algorithms, including those used for image recognition and reinforcement learning.

Implications for AI development: The research highlights a significant challenge in the field of artificial intelligence and machine learning.

  • The inability of AI models to learn continuously limits their adaptability and increases the cost of maintaining up-to-date systems.
  • This limitation could potentially hinder the progress of AI in areas where rapid adaptation to new information is crucial.
  • The findings underscore the need for innovative approaches to AI design that can overcome these inherent limitations.

Potential solution: The researchers have proposed a possible workaround to address the continuous learning problem.

  • They developed an algorithm that randomly reactivates some neurons after each training round, which appeared to reduce the poor performance associated with “dead” neurons.
  • This approach essentially “revives” inactive neurons, allowing the system to learn again.
  • While promising, this solution needs to be tested on much larger systems before its effectiveness can be confirmed for real-world applications.

Industry perspective: The inability of AI models to learn continuously has significant implications for the tech industry and AI research.

  • Mark van der Wilk from the University of Oxford describes a solution to continuous learning as a “billion-dollar question.”
  • A comprehensive solution that allows continuous model updates could significantly reduce the cost of training and maintaining AI systems.
  • This could potentially lead to more efficient and adaptable AI technologies across various applications.

Looking ahead: Challenges and opportunities in AI learning: The study’s findings open up new avenues for research and development in artificial intelligence.

  • The identified limitations in current AI models present both a challenge and an opportunity for innovation in the field.
  • Future research may focus on developing new architectures or training methods that enable continuous learning without compromising performance.
  • As AI continues to evolve, addressing these fundamental limitations could lead to more flexible, efficient, and human-like artificial intelligence systems.
AI models can't learn as they go along like humans do

Recent News

AI agents and the rise of Hybrid Organizations

Meta makes its improved AI image generator free to use while adding visible watermarks and daily limits to prevent misuse.

Adobe partnership brings AI creativity tools to Box’s content management platform

Box users can now access Adobe's AI-powered editing tools directly within their secure storage environment, eliminating the need to download files or switch between platforms.

Nvidia’s new ACE platform aims to bring more AI to games, but not everyone’s sold

Gaming companies are racing to integrate AI features into mainstream titles, but high hardware requirements and artificial interactions may limit near-term adoption.