Liquid AI unveils groundbreaking non-transformer AI models: Liquid AI, a startup with roots in MIT’s CSAIL, has introduced a new class of AI models that challenge the dominance of transformer-based architectures in the field of artificial intelligence.
Revolutionary approach to AI model design: Liquid AI’s new Liquid Foundation Models (LFMs) are built from first principles, eschewing the transformer architecture that has been the cornerstone of most recent AI advancements.
- The company’s goal is to explore alternative methods for building foundation models beyond Generative Pre-trained Transformers (GPTs).
- LFMs are based on computational units grounded in dynamical systems theory, signal processing, and numerical linear algebra.
- This novel approach allows for the creation of general-purpose AI models capable of handling various types of sequential data, including video, audio, text, time series, and signals.
Impressive performance and efficiency: Early benchmarks suggest that LFMs are already outperforming comparable transformer-based models while using significantly less memory.
- The smallest model, LFM 1.3B, has surpassed Meta’s Llama 3.2-1.2B and Microsoft’s Phi-1.5 on several third-party benchmarks, including the Massive Multitask Language Understanding (MMLU) test.
- LFM-3B requires only 16 GB of memory, compared to the 48 GB needed by Meta’s Llama-3.2-3B model, showcasing superior memory efficiency.
- The largest model, LFM 40B MoE, employs a Mixture-of-Experts approach similar to Mistral’s Mixtral.
Multimodal capabilities and industry applications: Liquid AI has designed its foundation models to be versatile across multiple data modalities, positioning them for use in various industries.
- The models can handle audio, video, and text data, making them suitable for a wide range of applications.
- Target industries include financial services, biotechnology, and consumer electronics.
- The models’ efficiency makes them ideal for deployment on edge devices and in enterprise-level applications.
Technical innovations and adaptability: LFMs build upon Liquid Neural Networks (LNNs), an architecture developed at CSAIL that aims to make artificial neurons more efficient and adaptable.
- LNNs demonstrate that fewer neurons, combined with innovative mathematical formulations, can achieve results comparable to traditional deep learning models with thousands of neurons.
- This approach allows for real-time adjustments during inference without the computational overhead associated with traditional models.
- LFMs can handle up to 1 million tokens efficiently while maintaining minimal memory usage.
Accessibility and future developments: While not open-source, Liquid AI is making its models available through various platforms and is actively seeking feedback from early adopters.
- Users can access the models through Liquid’s inference playground, Lambda Chat, or Perplexity AI.
- The company is optimizing its models for deployment on hardware from NVIDIA, AMD, Apple, Qualcomm, and Cerebras.
- Liquid AI plans to release technical blog posts and engage in red-teaming efforts to improve future iterations of the models.
Anticipation builds for official launch: Liquid AI is preparing for a full launch event on October 23, 2024, at MIT’s Kresge Auditorium in Cambridge, MA.
- The company is accepting RSVPs for in-person attendance at the event.
- This launch will mark Liquid AI’s official entry as a key player in the foundation model space.
Potential industry impact: The introduction of Liquid Foundation Models represents a significant development in AI technology, offering a compelling alternative to traditional transformer-based models.
- By combining state-of-the-art performance with unprecedented memory efficiency, LFMs could potentially disrupt the current AI landscape.
- The success of these models may encourage further research into non-transformer architectures, potentially leading to new breakthroughs in AI capabilities and efficiency.
Recent Stories
DOE fusion roadmap targets 2030s commercial deployment as AI drives $9B investment
The Department of Energy has released a new roadmap targeting commercial-scale fusion power deployment by the mid-2030s, though the plan lacks specific funding commitments and relies on scientific breakthroughs that have eluded researchers for decades. The strategy emphasizes public-private partnerships and positions AI as both a research tool and motivation for developing fusion energy to meet data centers' growing electricity demands. The big picture: The DOE's roadmap aims to "deliver the public infrastructure that supports the fusion private sector scale up in the 2030s," but acknowledges it cannot commit to specific funding levels and remains subject to Congressional appropriations. Why...
Oct 17, 2025Tying it all together: Credo’s purple cables power the $4B AI data center boom
Credo, a Silicon Valley semiconductor company specializing in data center cables and chips, has seen its stock price more than double this year to $143.61, following a 245% surge in 2024. The company's signature purple cables, which cost between $300-$500 each, have become essential infrastructure for AI data centers, positioning Credo to capitalize on the trillion-dollar AI infrastructure expansion as hyperscalers like Amazon, Microsoft, and Elon Musk's xAI rapidly build out massive computing facilities. What you should know: Credo's active electrical cables (AECs) are becoming indispensable for connecting the massive GPU clusters required for AI training and inference. The company...
Oct 17, 2025Vatican launches Latin American AI network for human development
The Vatican hosted a two-day conference bringing together 50 global experts to explore how artificial intelligence can advance peace, social justice, and human development. The event launched the Latin American AI Network for Integral Human Development and established principles for ethical AI governance that prioritize human dignity over technological advancement. What you should know: The Pontifical Academy of Social Sciences, the Vatican's research body for social issues, organized the "Digital Rerum Novarum" conference on October 16-17, combining academic research with practical AI applications. Participants included leading experts from MIT, Microsoft, Columbia University, the UN, and major European institutions. The conference...