Zubnet AIApprendreWiki › Liquid AI
Entreprises

Liquid AI

Aussi connu sous: Liquid Foundation Models, liquid neural networks
Spinout du MIT explorant des architectures de réseaux de neurones fondamentalement différentes inspirées des circuits neuronaux biologiques. Leurs Liquid Foundation Models utilisent des dynamiques continuous-time plutôt que des transformers à poids fixes, promettant meilleure efficacité et adaptabilité.

Pourquoi c'est important

Liquid AI représente le défi le plus sérieusement financé à l'hypothèse que les transformers sont la seule architecture qui compte. En construisant des foundation models de niveau production sur des dynamiques continuous-time bio-inspirées, ils testent si le pari all-in de l'industrie IA sur les mécanismes d'attention était prématuré. Même si les LFM ne détrônent pas les transformers carrément, leurs avantages d'efficacité pour le déploiement edge et le traitement de longues séquences pourraient tailler des niches critiques en robotique, IA mobile et systèmes embarqués — des marchés où faire tourner un transformer 70B n'est simplement pas une option.

Deep Dive

Liquid AI grew out of research at MIT's Computer Science and Artificial Intelligence Laboratory (CSAIL), specifically from the work of Ramin Hasani, Mathias Lechner, and Daniela Rus. Hasani and Lechner had been studying the nervous systems of C. elegans — a tiny roundworm with exactly 302 neurons — and discovered that the mathematical equations governing these biological neural circuits could be adapted into a new kind of artificial neural network. Unlike standard networks where connection weights are fixed after training, these "liquid" networks use continuous-time differential equations that allow parameters to adapt dynamically based on input. The company was formally founded in 2023 and quickly raised over $250 million at a $2 billion-plus valuation, with backing from AMD Ventures and other investors who saw the potential for an architecture that fundamentally breaks from the transformer paradigm.

Liquid Foundation Models: A Different Bet

Liquid AI's core product line — the Liquid Foundation Models (LFMs) — launched with three sizes: LFM-1B, LFM-3B, and LFM-40B. What makes these architecturally distinct is that they are not transformers, and they are not state-space models in the Mamba sense either. LFMs use a hybrid approach combining structured state-space layers with attention-like mechanisms, but the underlying mathematics is rooted in those continuous-time dynamics from the biological research. In practice, this means LFMs can handle very long sequences more efficiently than standard transformers — their memory footprint doesn't blow up quadratically with sequence length. The LFM-1B model, in particular, attracted attention for outperforming several transformer-based models of similar size on standard benchmarks, suggesting that the architectural differences translated into real capability gains rather than just theoretical elegance.

Edge AI and the Efficiency Argument

One of Liquid AI's most compelling claims is efficiency at the edge. Because liquid networks can represent complex dynamics with fewer parameters than transformers, they're naturally suited for deployment on devices with limited compute — phones, robots, IoT sensors, autonomous vehicles. The company has been explicit about targeting these use cases, positioning themselves not as another chatbot company but as the architecture provider for AI that runs everywhere. This is a fundamentally different market from the cloud-first approach of most AI labs. If your model can run meaningfully on a phone's neural processing unit without constant server calls, you unlock applications that are impossible with cloud-dependent AI: real-time robotics, offline processing, privacy-preserving inference on device. Liquid AI partnered with Qualcomm and other hardware vendors to optimize their models for specific chip architectures, a move that signals serious intent about the edge deployment story.

The Architecture Diversity Thesis

Liquid AI's existence is a bet on architectural diversity — the idea that transformers, despite their dominance, are not the final word in neural network design. This thesis has gained credibility as the limitations of transformers have become clearer: quadratic attention costs, difficulty with very long sequences, massive energy consumption during inference. The state-space model community (Mamba, RWKV, and others) has already proven that competitive alternatives exist; Liquid AI pushes further by arguing that biologically inspired dynamics offer advantages that even SSMs miss, particularly in temporal reasoning and adaptive behavior. Whether this is true at frontier scale remains unproven — the LFM-40B is competitive but not dominant against the best transformer models of comparable size — but the theoretical foundations are rigorous enough that the AI research community takes the work seriously.

Challenges and Skepticism

The obvious risk for Liquid AI is that the transformer ecosystem is enormously entrenched. The software stack (PyTorch, CUDA kernels, inference servers) is overwhelmingly optimized for transformer architectures. Every major cloud provider has spent billions building infrastructure tuned for attention-based models. Switching to a fundamentally different architecture means rebuilding tooling, retraining engineers, and convincing customers that the efficiency gains justify the transition costs. Liquid AI has addressed this partially by providing drop-in API compatibility — from the user's perspective, calling an LFM looks identical to calling any other model. But the deeper challenge is whether they can demonstrate a clear, sustained advantage at the scales that matter for enterprise adoption. With $250 million in funding and strong academic credentials, they have more runway than most architecture challengers. The next year will determine whether liquid neural networks become a real force in production AI or remain one of the most intellectually interesting footnotes in the field's history.

Concepts liés

In The News

Liquid AI's 350M Model Outperforms 1B+ Models With Hybrid Architecture
Apr 01, 2026
← Tous les termes
← Leonardo.ai llama.cpp →
ESC