AI systems based on large neural networks present significant software engineering challenges that raise serious concerns about their reliability and responsible deployment, according to Professor Eerke Boiten of De Montfort University Leicester.
Core argument: Current AI systems, particularly those based on large neural networks, are fundamentally unmanageable from a software engineering perspective, making their use in critical applications irresponsible.
- The primary challenge stems from the inability to apply traditional software engineering tools and principles to manage complexity and scale
- These systems lack transparency and accountability, two essential elements for trustworthy software development
- The development of AI has coincided with a concerning trend of diminished responsibility regarding data sources and algorithmic outcomes
Technical fundamentals: Large neural networks, which power most modern AI systems including generative AI and large language models (LLMs), operate through a complex web of interconnected nodes that process information in ways that are difficult to predict or control.
- Neural networks contain millions of nodes, each processing multiple inputs through weighted connections and activation functions
- Training these networks requires enormous computational resources, often costing millions of dollars
- The training process is largely unsupervised, with minimal human input beyond potential post-training adjustments
Key limitations: The emergent behavior of neural networks fundamentally conflicts with traditional software engineering principles, particularly compositionality.
- Neural networks lack internal structure that meaningfully relates to their functionality
- They cannot be developed or reused as components
- These systems do not create explicit models of knowledge
- The absence of intermediate models prevents stepwise development
- Explanation of system behavior becomes extremely difficult due to the lack of reasoning representation
Verification challenges: Traditional software testing and verification methods prove inadequate for current AI systems.
- Input and state spaces are too vast for exhaustive testing
- Stochastic behavior means correct outputs in testing don’t guarantee consistent performance
- Component-level testing is impossible
- Meaningful test coverage metrics cannot be established
- The only available verification method – whole system testing – provides insufficient confidence
Fault management: The handling of errors and system improvements presents significant obstacles.
- Error behavior is emergent and unpredictable
- The scale of unsupervised training versus human error correction creates inherent reliability issues
- Error fixes through retraining can introduce new problems that are difficult to detect
- Regression testing becomes effectively impossible
Looking ahead: While current AI architecture may represent a developmental dead end, alternative approaches could offer more promising paths forward.
- Hybrid systems combining symbolic and intuition-based AI might provide better reliability
- AI systems could be valuable in limited contexts where errors can be detected and managed
- Applications like weather prediction, where probabilistic outputs are expected, might be more suitable use cases
- The development of compositional approaches to neural networks, though challenging, could address current limitations
Does current AI represent a dead end?