File size: 2,468 Bytes
b9562f8
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
---
base_model:
- LiquidAI/LFM2-1.2B
---

# LFM2-1.2B
Run **LFM2-1.2B** on Qualcomm NPU with [NexaSDK](https://sdk.nexa.ai).

## Quickstart

1. **Install NexaSDK** and create a free account at [sdk.nexa.ai](https://sdk.nexa.ai)
2. **Activate your device** with your access token:

   ```bash
   nexa config set license '<access_token>'
   ```
3. Run the model locally in one line:

   ```bash
   nexa infer NexaAI/LFM2-1.2B-npu
   ```

## Model Description  
**LFM2-1.2B** is part of Liquid AI’s second-generation **LFM2** family, designed specifically for **on-device and edge AI deployment**.  
With **1.2 billion parameters**, it strikes a balance between compact size, strong reasoning, and efficient compute utilization—ideal for running on CPUs, GPUs, or NPUs.

LFM2 introduces a **hybrid Liquid architecture** with **multiplicative gates and short convolutions**, enabling faster convergence and improved contextual reasoning.  
It demonstrates up to **3× faster training** and **2× faster inference** on CPU compared to Qwen3, while maintaining superior accuracy across multilingual and instruction-following benchmarks.

## Features  
-**Speed & Efficiency** – 2× faster inference and prefill].  
- 🧠 **Hybrid Liquid Architecture** – Combines multiplicative gating with convolutional layers for better reasoning and token reuse.  
- 🌍 **Multilingual Competence** – Supports diverse languages for global use cases.  
- 🛠 **Flexible Deployment** – Runs efficiently on CPU, GPU, and NPU hardware.  
- 📈 **Benchmark Performance** – Outperforms similarly-sized models in math, knowledge, and reasoning tasks.

## Use Cases  
- Edge AI assistants and voice agents  
- Offline reasoning and summarization on mobile or automotive devices  
- Local code and text generation tools  
- Lightweight multimodal or RAG pipelines  
- Domain-specific fine-tuning for vertical applications (e.g., finance, robotics)

## Inputs and Outputs  
**Input**  
- Text prompts or structured instructions (tokenized sequences for API use).  

**Output**  
- Natural-language or structured text generations.  
- Optionally: logits or embeddings for advanced downstream integration.

## License  
This model is released under the **Creative Commons Attribution–NonCommercial 4.0 (CC BY-NC 4.0)** license.  
Non-commercial use, modification, and redistribution are permitted with attribution.  
For commercial licensing, please contact **dev@nexa.ai**.