What is included in this article: |
Performance breakthrough – Faster reasoning speed 2 times faster, faster training speed Technical Architecture – Mixed design of convolution and attention blocks Model specifications – Three size variants (350m, 700m, 1.2b parameters) Benchmark results – Excellent performance compared to similar-sized models Deployment Optimization – Edge design for various hardware Open source accessibility – Licensing based on Apache 2.0 Market meaning – Impact on Edge AI adoption |
Liquid AI Release LFM2 (its second-generation liquid foundation model) releases liquid AI release, thus making a significant leap in the landscape of artificial intelligence on devices. This new generative AI model represents a paradigm shift in edge computing, providing unprecedented performance optimizations designed specifically to maintain competitive quality standards while maintaining competitive quality standards.
Revolutionary performance growth
LFM2 establishes new benchmarks in edge AI space by achieving significant efficiency improvements across multiple dimensions. These models provide faster decoding and pre-fill performance compared to QWEN3 on CPU architectures, which is a key advancement in real-time applications. Perhaps even more impressive is that the training process itself has been optimized to achieve 3x faster training compared to previous LFM generation, making LFM2 the most cost-effective way to build functional, universal universal AI systems.
These performance improvements are not only incremental, but also represent fundamental breakthroughs in accessing powerful AI on resource-constrained devices. These models are specially designed to unlock millisecond latency, offline resilience and data protection privacy – cell phones, laptops, cars, robots, wearables, satellites and other features that must be understood in real time.
Hybrid building innovation
The technical basis of LFM2 lies in its new hybrid architecture, combining the best aspects of convolution and attention mechanisms. The model adopts a complex 16-block structure consisting of 10 bi-gate short-term convolution blocks and 6 grouped query attention (GQA). This hybrid approach comes from the pioneering work of liquid AI on liquid time constant network (LTC), which introduces continuous time recurrent neural networks of linear dynamic systems regulated by nonlinear input-related gates.
At the heart of this architecture is the Linear Input Change (LIV) operator framework, which directly generates weights through the inputs it acts as. This allows convolution, recurrence, attention and other structured layers to fall under a unified input-aware framework. The LFM2 convolution block implements multiplication gates and short convolutions, creating a linear first-order system that converges to zero after a finite time.
The architecture selection process utilizes the Neural Architecture Search Engine of Stellar, Liquid AI, which has been modified to evaluate language modeling capabilities beyond traditional verification loss and confusion metrics. Instead, it uses a comprehensive suite of over 50 internal assessments that can evaluate multiple features including knowledge recollection, multi-hop on reasoning, understanding of low-resource languages, the following teaching and tools use.

Comprehensive model lineup
The LFM2 is available in three strategic size configurations: 350m, 700m and 1.2b parameters, each of which is optimized for different deployment plans while maintaining the core efficiency advantages. All models are trained in 10 trillion tokens, and these tokens are taken from a well-planned pre-trained corpus that includes about 75% in English, 20% in multilingual content and 5% in code data from web and licensed materials.
The training method combines knowledge distillation using existing LFM1-7B as teacher model, and the cross-training between LFM2 student outcomes and teacher outcomes throughout the 10T token training process is the primary training signal. During preprocessing, the context length extends to 32K, allowing the model to effectively process longer sequences.

Superior benchmark performance
The evaluation results show that LFM2 is significantly better than similar-sized models in multiple benchmark categories. Although the parameters are 47%, the LFM2-1.2B model is competitive with QWEN3-1.7B. Similarly, the LFM2-700M outperformed the Gemma 3 1b IT, while the smallest LFM2-350M checkpoint remained competitive with the QWEN3-0.6B and LLAMA 3.2 1B indications.
In addition to automated benchmarks, LFM2 also shows excellent conversational capabilities in multi-transfer conversations. LFM2-1.2B uses the Wildchat dataset and the LLM-AS-AAA-Gudge evaluation framework to match QWEN3-1.7B performance, while LFM2-1.2B shows significant preference advantages over Llama 3.2 1B instructions and Gemma 3 1B, albeit much smaller and faster.

Deployment of edge optimization
These models perform well in real-world deployment scenarios and have been exported to multiple inference frameworks including Pytorch’s executorch and the open source llama.cpp library. Testing target hardware, including the Samsung Galaxy S24 Ultra and AMD Ryzen platforms, showed that LFM2 dominates the speed of pre-filling and decoding inference relative to model size.
Powerful CPU performance is effectively converted into accelerators such as GPU and NPU after kernel optimization, making LFM2 suitable for a variety of hardware configurations. This flexibility is crucial for various ecosystems of devices that require the edge of device AI capabilities.
in conclusion
The launch of LFM2 addresses a key gap in the field of AI deployment, where the transition from cloud-based edge-based reasoning is accelerating. By enabling millisecond delay, offline operations and data to enable human privacy, LFM2 unlocks new possibilities for AI integration in consumer electronics, robotics, smart appliances, finance, e-commerce and education.
The technological achievements represented in LFM2 show the maturity of Edge AI technology, where the trade-off between model capability and deployment efficiency is successfully optimized. As businesses move from cloud LLM to a hub of cost-effective, fast, private and on-premises intelligence, LFM2 positiones itself as the fundamental technology for next-generation AI-powered devices and applications.

Asif Razzaq is CEO of Marktechpost Media Inc. As a visionary entrepreneur and engineer, ASIF is committed to harnessing the potential of artificial intelligence to achieve social benefits. His recent effort is to launch Marktechpost, an artificial intelligence media platform that has an in-depth coverage of machine learning and deep learning news that can sound both technically, both through technical voices and be understood by a wide audience. The platform has over 2 million views per month, demonstrating its popularity among its audience.