mlabonne commited on
Commit
9747fc8
·
verified ·
1 Parent(s): 2a067a2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +16 -1
README.md CHANGED
@@ -51,6 +51,14 @@ Alternatively, try the API model on the [Playground](https://playground.liquid.a
51
 
52
  ## 📄 Model details
53
 
 
 
 
 
 
 
 
 
54
  LFM2.5-VL-1.6B is a general-purpose vision-language model with the following features:
55
 
56
  - **LM Backbone**: LFM2.5-1.2B-Base
@@ -65,6 +73,13 @@ LFM2.5-VL-1.6B is a general-purpose vision-language model with the following fea
65
  - text: `temperature=0.1`, `min_p=0.15`, `repetition_penalty=1.05`
66
  - vision: `min_image_tokens=64` `max_image_tokens=256`, `do_image_splitting=True`
67
 
 
 
 
 
 
 
 
68
  We recommend using it for general vision-language workloads, OCR or document comprehension. It’s not well-suited for knowledge-intensive tasks.
69
 
70
  ### Chat Template
@@ -174,7 +189,7 @@ We recommend fine-tuning LFM2.5-VL-1.6B model on your use cases to maximize perf
174
 
175
  | Notebook | Description | Link |
176
  |-----------|----------------------------------------------------------------------|------|
177
- | SFT (TRL) | Supervised Fine-Tuning (SFT) notebook with a LoRA adapter using TRL. | <a href="https://colab.research.google.com/drive/10530_jt_Joa5zH2wgYlyXosypq1R7PIz?usp=sharing"><img src="https://cdn-uploads.huggingface.co/production/uploads/61b8e2ba285851687028d395/vlOyMEjwHa_b_LXysEu2E.png" width="110" alt="Colab link"></a> |
178
 
179
 
180
  ## 📊 Performance
 
51
 
52
  ## 📄 Model details
53
 
54
+ | Model | Parameters | Description |
55
+ |-------|------------|-------------|
56
+ | [LFM2.5-1.2B-Base](https://huggingface.co/LiquidAI/LFM2.5-1.2B-Base) | 1.2B | Pre-trained base model for fine-tuning |
57
+ | [LFM2.5-1.2B-Instruct](https://huggingface.co/LiquidAI/LFM2.5-1.2B-Instruct) | 1.2B | General-purpose instruction-tuned model |
58
+ | [LFM2.5-1.2B-JP](https://huggingface.co/LiquidAI/LFM2.5-1.2B-JP) | 1.2B | Japanese-optimized chat model |
59
+ | [**LFM2.5-VL-1.6B**](https://huggingface.co/LiquidAI/LFM2.5-VL-1.6B) | 1.6B | Vision-language model with fast inference |
60
+ | [LFM2.5-Audio-1.5B](https://huggingface.co/LiquidAI/LFM2.5-Audio-1.5B) | 1.5B | Audio-language model for speech and text I/O |
61
+
62
  LFM2.5-VL-1.6B is a general-purpose vision-language model with the following features:
63
 
64
  - **LM Backbone**: LFM2.5-1.2B-Base
 
73
  - text: `temperature=0.1`, `min_p=0.15`, `repetition_penalty=1.05`
74
  - vision: `min_image_tokens=64` `max_image_tokens=256`, `do_image_splitting=True`
75
 
76
+ | Model | Description |
77
+ |-------|-------------|
78
+ | [**LFM2.5-VL-1.6B**](https://huggingface.co/LiquidAI/LFM2.5-VL-1.6B) | Original model checkpoint in native format. Best for fine-tuning or inference with Transformers and vLLM. |
79
+ | [LFM2.5-VL-1.6B-GGUF](https://huggingface.co/LiquidAI/LFM2.5-VL-1.6B-GGUF) | Quantized format for llama.cpp and compatible tools. Optimized for CPU inference and local deployment with reduced memory usage. |
80
+ | [LFM2.5-VL-1.6B-ONNX](https://huggingface.co/LiquidAI/LFM2.5-VL-1.6B-ONNX) | ONNX Runtime format for cross-platform deployment. Enables hardware-accelerated inference across diverse environments (cloud, edge, mobile). |
81
+ | [LFM2.5-VL-1.6B-MLX](https://huggingface.co/mlx-community/LFM2.5-VL-1.6B-8bit) | MLX format for Apple Silicon. Optimized for fast inference on Mac devices using the MLX framework. |
82
+
83
  We recommend using it for general vision-language workloads, OCR or document comprehension. It’s not well-suited for knowledge-intensive tasks.
84
 
85
  ### Chat Template
 
189
 
190
  | Notebook | Description | Link |
191
  |-----------|----------------------------------------------------------------------|------|
192
+ | SFT (TRL) | Supervised Fine-Tuning with LoRA using TRL. | <a href="https://colab.research.google.com/drive/10530_jt_Joa5zH2wgYlyXosypq1R7PIz?usp=sharing"><img src="https://cdn-uploads.huggingface.co/production/uploads/61b8e2ba285851687028d395/vlOyMEjwHa_b_LXysEu2E.png" width="110" alt="Colab link"></a> |
193
 
194
 
195
  ## 📊 Performance