×
Why some software developers believe neural networks are a dead end
Written by
Published on
Join our daily newsletter for breaking news, product launches and deals, research breakdowns, and other industry-leading AI coverage
Join Now

AI systems based on large neural networks present significant software engineering challenges that raise serious concerns about their reliability and responsible deployment, according to Professor Eerke Boiten of De Montfort University Leicester.

Core argument: Current AI systems, particularly those based on large neural networks, are fundamentally unmanageable from a software engineering perspective, making their use in critical applications irresponsible.

  • The primary challenge stems from the inability to apply traditional software engineering tools and principles to manage complexity and scale
  • These systems lack transparency and accountability, two essential elements for trustworthy software development
  • The development of AI has coincided with a concerning trend of diminished responsibility regarding data sources and algorithmic outcomes

Technical fundamentals: Large neural networks, which power most modern AI systems including generative AI and large language models (LLMs), operate through a complex web of interconnected nodes that process information in ways that are difficult to predict or control.

  • Neural networks contain millions of nodes, each processing multiple inputs through weighted connections and activation functions
  • Training these networks requires enormous computational resources, often costing millions of dollars
  • The training process is largely unsupervised, with minimal human input beyond potential post-training adjustments

Key limitations: The emergent behavior of neural networks fundamentally conflicts with traditional software engineering principles, particularly compositionality.

  • Neural networks lack internal structure that meaningfully relates to their functionality
  • They cannot be developed or reused as components
  • These systems do not create explicit models of knowledge
  • The absence of intermediate models prevents stepwise development
  • Explanation of system behavior becomes extremely difficult due to the lack of reasoning representation

Verification challenges: Traditional software testing and verification methods prove inadequate for current AI systems.

  • Input and state spaces are too vast for exhaustive testing
  • Stochastic behavior means correct outputs in testing don’t guarantee consistent performance
  • Component-level testing is impossible
  • Meaningful test coverage metrics cannot be established
  • The only available verification method – whole system testing – provides insufficient confidence

Fault management: The handling of errors and system improvements presents significant obstacles.

  • Error behavior is emergent and unpredictable
  • The scale of unsupervised training versus human error correction creates inherent reliability issues
  • Error fixes through retraining can introduce new problems that are difficult to detect
  • Regression testing becomes effectively impossible

Looking ahead: While current AI architecture may represent a developmental dead end, alternative approaches could offer more promising paths forward.

  • Hybrid systems combining symbolic and intuition-based AI might provide better reliability
  • AI systems could be valuable in limited contexts where errors can be detected and managed
  • Applications like weather prediction, where probabilistic outputs are expected, might be more suitable use cases
  • The development of compositional approaches to neural networks, though challenging, could address current limitations
Does current AI represent a dead end?

Recent News

How organizations are thinking about AI implementation in 2025

Large consulting firms are formalizing executive oversight of AI strategy as enterprise adoption accelerates, with PwC installing its first Chief AI Officer to guide both internal and client implementations.

Why artist Jeff Koons says he won’t use AI anytime soon

Leading contemporary artist positions AI as auxiliary tool while maintaining biology-driven creative process, signaling persistent traditionalism in high-value art production.

UTSA unveils plans for new AI and cybersecurity college

The Texas university will merge its data science, cybersecurity and computing programs to create a pipeline of tech talent for San Antonio's rapidly growing security sector.