Tooliax Logo
ExploreCompareCategoriesSubmit Tool
News
Tooliax Logo
ExploreCompareCategoriesSubmit Tool
News
Liquid AI Unveils LFM2.5: A New Era for Compact, On-Device AI
Back to News
Wednesday, January 7, 20263 min read

Liquid AI Unveils LFM2.5: A New Era for Compact, On-Device AI

Liquid AI has introduced LFM2.5, a next-generation family of small foundation models built upon its innovative LFM2 architecture. These models are engineered with a specific focus on efficient deployment in on-device and edge computing environments. The diverse LFM2.5 series includes foundational and instruction-tuned variants, alongside specialized models for Japanese language, vision-language, and audio-language tasks. All models are openly available, with weights accessible on Hugging Face and integration facilitated through the LEAP platform.

LFM2.5: Engineered for Efficiency and Performance

The LFM2.5 models retain the hybrid LFM2 architecture, which is inherently optimized for rapid and memory-efficient inference on various processors, including CPUs and NPUs. A significant enhancement in this iteration is the scaling of the data and post-training pipelines. The pretraining phase for the 1.2 billion parameter core model was substantially expanded, progressing from 10 trillion to 28 trillion tokens. Furthermore, the instruct-tuned variant undergoes supervised fine-tuning, preference alignment, and extensive multi-stage reinforcement learning. This rigorous process concentrates on refining abilities such as instruction following, tool utilization, mathematical reasoning, and general knowledge inference.

Demonstrated Text Model Performance

The LFM2.5-1.2B-Instruct model serves as the primary general-purpose text solution within the family. Liquid AI has reported compelling benchmark results, positioning this model favorably against competitors in its class. On complex reasoning tasks, it achieved scores of 38.89 on GPQA and 44.35 on MMLU Pro. These figures notably surpass those of other 1-billion-parameter open models, including Llama-3.2-1B Instruct and Gemma-3-1B IT. For multi-step instruction adherence and function calling quality, the LFM2.5-1.2B-Instruct model posted strong results of 86.23 on IFEval and 47.33 on IFBench, again leading other comparable models.

Specialized Language and Multimodal Variants

Optimized for Japanese Language

A specialized version, LFM2.5-1.2B-JP, has been developed for the Japanese market, derived from the same robust backbone. This model is tailored for Japanese-specific tasks like JMMLU, M-IFEval in Japanese, and GSM8K in Japanese. It demonstrates superior performance on localized benchmarks compared to the general instruct model and either matches or exceeds other small multilingual models such as Qwen3-1.7B, Llama 3.2-1B Instruct, and Gemma 3-1B IT.

Vision-Language for Real-World Edge Applications

The updated LFM2.5-VL-1.6B model brings enhanced vision-language capabilities to the series. It integrates a dedicated vision tower for image comprehension with the LFM2.5-1.2B-Base as its language foundation. This model has undergone tuning across a broad spectrum of visual reasoning and optical character recognition (OCR) benchmarks, including MMStar, MM IFEval, and OCRBench v2, among others. It shows significant improvements over its predecessor, LFM2-VL-1.6B, and is designed for practical applications such as document analysis, user interface interpretation, and multi-image reasoning in constrained edge environments.

Native Audio Interaction and Generation

LFM2.5-Audio-1.5B stands out as a native audio language model, supporting both audio and text inputs and outputs. Presented as an Audio-to-Audio model, it incorporates an audio detokenizer reportedly eight times faster than previous versions while maintaining precision on hardware with limited resources. The model offers two distinct generation modes: interleaved generation for real-time speech-to-speech conversational agents where low latency is critical, and sequential generation for tasks such as automatic speech recognition (ASR) and text-to-speech (TTS), allowing modality switching without reinitialization. Its audio stack benefits from quantization-aware training at low precision, ensuring performance metrics remain consistent with full-precision baselines, thereby facilitating deployment on devices with restricted computational power.

Availability and Impact

The release of the LFM2.5 family as open weights on Hugging Face and through the LEAP platform signifies Liquid AI's commitment to advancing accessible, high-performance AI for on-device applications. By offering a range of specialized models—from general text to multimodal and localized variants—the LFM2.5 series is poised to empower developers and organizations to create sophisticated AI agents capable of operating effectively at the edge.

This article is a rewritten summary based on publicly available reporting. For the original story, visit the source.

Source: MarkTechPost
Share this article

Latest News

From Political Chaos to Policy Crossroads: Albanese Navigates Shifting Sands

From Political Chaos to Policy Crossroads: Albanese Navigates Shifting Sands

Feb 3

Historic Reimagining: Barnsley Crowned UK's First 'Tech Town' with Major Global Partnerships

Historic Reimagining: Barnsley Crowned UK's First 'Tech Town' with Major Global Partnerships

Feb 3

OpenClaw: Viral AI Assistant's Autonomy Ignites Debate Amidst Expert Warnings

OpenClaw: Viral AI Assistant's Autonomy Ignites Debate Amidst Expert Warnings

Feb 3

Adobe Sunsets Animate: A Generative AI Strategy Claims a Legacy Tool

Adobe Sunsets Animate: A Generative AI Strategy Claims a Legacy Tool

Feb 3

Palantir CEO Alex Karp: ICE Protesters Should Demand *More* AI Surveillance

Palantir CEO Alex Karp: ICE Protesters Should Demand *More* AI Surveillance

Feb 3

View All News

More News

Crafting Enterprise AI: Five Pillars for Scalability and Resilience

February 2, 2026

Crafting Enterprise AI: Five Pillars for Scalability and Resilience

East London Cafe Transforms Orders into Conversations, Fostering Connection Through British Sign Language

February 2, 2026

East London Cafe Transforms Orders into Conversations, Fostering Connection Through British Sign Language

Exposed: The 'AI-Washing' Phenomenon Masking Traditional Layoffs

February 2, 2026

Exposed: The 'AI-Washing' Phenomenon Masking Traditional Layoffs

Tooliax LogoTooliax

Your comprehensive directory for discovering, comparing, and exploring the best AI tools available.

Quick Links

  • Explore Tools
  • Compare
  • Submit Tool
  • About Us

Legal

  • Privacy Policy
  • Terms of Service
  • Cookie Policy
  • Contact

© 2026 Tooliax. All rights reserved.