IBM’s David Cox champions open innovation in enterprise generative AI, emphasizing the importance of transparency, collaboration, and the integration of proprietary business data into AI models.
Nuanced view of openness in AI: Cox challenges the notion that openness in AI is a simple binary concept, highlighting the growing ecosystem of open models from various sources, including tech giants, universities, and nation-states:
- He raises concerns about the quality of openness in many large language models (LLMs), noting that some provide only a “bag of numbers” without clear information on how they were produced, making reproducibility difficult or impossible.
- Cox outlines key characteristics of successful open-source projects, such as frequent updates, structured release cycles, regular security fixes, and active community contributions, and argues that many current open LLMs lack these properties.
Integrating enterprise data into LLMs: Cox proposes a novel perspective on LLMs, framing them primarily as data representations rather than just conversational tools, and suggests a mission to represent enterprise data within foundation models:
- He points out a significant gap in current LLMs: the proprietary “secret sauce” of enterprises remains largely unrepresented, limiting the potential value of these models for businesses.
- To address this, Cox outlines a three-step approach for enterprises: finding an open, trusted base model; creating a new representation of business data; and deploying, scaling, and creating value.
InstructLab: Practical implementation of enterprise AI adoption: Cox introduces InstructLab, a collaborative project between IBM and Red Hat that brings his vision for integrating enterprise data with open-source LLMs to life:
- InstructLab addresses the challenge of incorporating proprietary enterprise knowledge into AI models by offering a “genuinely open-source contribution model for LLMs.”
- The project’s methodology revolves around a taxonomy of world knowledge and skills, enabling users to precisely target areas for model enhancement and facilitating the integration of enterprise-specific expertise.
- InstructLab’s use of a “teacher” model to generate synthetic training data allows for the integration of proprietary data while maintaining model performance and adding enterprise-specific capabilities.
Broader implications for the future of enterprise AI: Cox’s insights and IBM’s InstructLab point to a shift in enterprise AI adoption, moving from generic, off-the-shelf models to tailored solutions that reflect each company’s unique expertise. As this technology matures, the competitive edge may well belong to those who can most effectively turn their institutional knowledge into AI-powered insights, suggesting that the next chapter of AI is not just about smarter machines but about machines that understand businesses as well as their creators do.
Recent Stories
DOE fusion roadmap targets 2030s commercial deployment as AI drives $9B investment
The Department of Energy has released a new roadmap targeting commercial-scale fusion power deployment by the mid-2030s, though the plan lacks specific funding commitments and relies on scientific breakthroughs that have eluded researchers for decades. The strategy emphasizes public-private partnerships and positions AI as both a research tool and motivation for developing fusion energy to meet data centers' growing electricity demands. The big picture: The DOE's roadmap aims to "deliver the public infrastructure that supports the fusion private sector scale up in the 2030s," but acknowledges it cannot commit to specific funding levels and remains subject to Congressional appropriations. Why...
Oct 17, 2025Tying it all together: Credo’s purple cables power the $4B AI data center boom
Credo, a Silicon Valley semiconductor company specializing in data center cables and chips, has seen its stock price more than double this year to $143.61, following a 245% surge in 2024. The company's signature purple cables, which cost between $300-$500 each, have become essential infrastructure for AI data centers, positioning Credo to capitalize on the trillion-dollar AI infrastructure expansion as hyperscalers like Amazon, Microsoft, and Elon Musk's xAI rapidly build out massive computing facilities. What you should know: Credo's active electrical cables (AECs) are becoming indispensable for connecting the massive GPU clusters required for AI training and inference. The company...
Oct 17, 2025Vatican launches Latin American AI network for human development
The Vatican hosted a two-day conference bringing together 50 global experts to explore how artificial intelligence can advance peace, social justice, and human development. The event launched the Latin American AI Network for Integral Human Development and established principles for ethical AI governance that prioritize human dignity over technological advancement. What you should know: The Pontifical Academy of Social Sciences, the Vatican's research body for social issues, organized the "Digital Rerum Novarum" conference on October 16-17, combining academic research with practical AI applications. Participants included leading experts from MIT, Microsoft, Columbia University, the UN, and major European institutions. The conference...