×
Liquid AI’s new vision-language models run 2x faster on smartphones
Written by
Published on
Join our daily newsletter for breaking news, product launches and deals, research breakdowns, and other industry-leading AI coverage
Join Now

Liquid AI has launched LFM2-VL, a new family of vision-language foundation models designed for efficient deployment on smartphones, laptops, wearables, and embedded systems. The models promise up to twice the GPU inference speed of comparable vision-language models while maintaining competitive accuracy, addressing the growing demand for on-device AI that can process both text and images without relying on cloud infrastructure.

What you should know: LFM2-VL represents a significant step toward making multimodal AI accessible for resource-constrained devices through architectural innovations that prioritize efficiency.

  • The models can process images at native resolutions up to 512×512 pixels without distortion, using smart patching for larger images that preserves both fine detail and global context.
  • Two variants are available: LFM2-VL-450M with less than half a billion parameters for highly constrained environments, and LFM2-VL-1.6B for more capable single-GPU deployment.
  • Both models are built on Liquid AI’s foundation architecture based on dynamical systems and signal processing principles, moving beyond traditional transformer models.

How it works: The models use a modular architecture combining multiple specialized components to achieve their efficiency gains.

  • LFM2-VL integrates a language model backbone, SigLIP2 NaFlex vision encoder, and a multimodal projector with pixel unshuffle technology that reduces image tokens and improves throughput.
  • Users can adjust parameters like maximum image tokens or patches to balance speed and quality for specific deployment scenarios.
  • Training involved approximately 100 billion multimodal tokens from open datasets and synthetic data.

Performance benchmarks: LFM2-VL achieves competitive results across vision-language evaluations while delivering superior processing speeds.

  • The 1.6B model scores 65.23 in RealWorldQA, 58.68 in InfoVQA, and 742 in OCRBench, maintaining solid performance in multimodal reasoning tasks.
  • In inference testing, LFM2-VL achieved the fastest GPU processing times in its class when tested on a standard workload of a 1024×1024 image with short prompt.

The bigger picture: This launch builds on Liquid AI’s broader strategy to decentralize AI execution and reduce cloud dependency.

  • In July 2025, the company launched the Liquid Edge AI Platform (LEAP), a cross-platform SDK enabling developers to run small language models directly on mobile devices.
  • LEAP offers OS-agnostic support for iOS and Android with models as small as 300MB, accompanied by Apollo, a companion app for offline model testing.
  • The approach reflects growing industry interest in privacy-preserving, low-latency AI that operates independently of internet connectivity.

What they’re saying: Liquid AI co-founder and CEO Ramin Hasani emphasized the company’s core value proposition in announcing the release.

  • “Efficiency is our product,” Hasani wrote on X, highlighting the models’ “up to 2× faster on GPU with competitive accuracy” and “smart patching for big images.”

Company background: Liquid AI was founded by former MIT CSAIL researchers focused on building alternatives to transformer-based architectures.

  • The company’s Liquid Foundation Models are based on principles from dynamical systems, signal processing, and numerical linear algebra.
  • Their approach aims to deliver competitive performance using significantly fewer computational resources while enabling real-time adaptability during inference.

Availability and licensing: The models are immediately accessible through standard development channels with custom licensing terms.

  • LFM2-VL models are available on Hugging Face with example fine-tuning code in Colab, compatible with Hugging Face transformers and TRL.
  • They’re released under a custom “LFM1.0 license” based on Apache 2.0 principles, with commercial use permitted under different terms for companies above and below $10 million in annual revenue.
Liquid AI wants to give smartphones small, fast AI that can see with new LFM2-VL model

Recent News

CoreWeave stock drops 14% on earnings miss despite $1.21B revenue beat

A $30.1 billion backlog suggests strong demand despite near-term execution concerns.

Hidden Door launches AI role-playing platform with creator revenue sharing

Players can add vampires to Pride and Prejudice while creators earn from their IP.

AI helps manufacturers maintain lean inventories amid tariff uncertainty

AI agents suggest decisions like moving 100 tonnes between plants, but humans still hit accept.