fmasterpro27's picture
Update README.md
b1e95f3 verified
---
license: other
language:
- en
- ar
- zh
- fr
- de
- ja
- ko
- es
base_model:
- LiquidAI/LFM2.5-1.2B-Base
library_name: transformers
tags:
- liquid
- lfm2
- lfm2.5
pipeline_tag: text-generation
---
# Open4bits / LFM2.5-1.2B-Base-Quantized
This repository provides **multiple quantized variants** of the **LFM 2.5 Base (1.2B parameters)** model for efficient inference and deployment.
The **original model** is developed and released by **LiquidAI**:
Original model:
https://huggingface.co/LiquidAI/LFM2.5-1.2B-Base
These quantizations are maintained and published by **ArkAiLab** under the **Open4bits** organization to improve accessibility across a wide range of hardware.
---
## Available Quantization Formats
Each format is stored in a **separate directory**:
- **FP16** – Baseline half-precision weights
- **FP8** – High-performance low-precision format (GPU support required)
- **INT8** – Balanced performance and memory usage (BitsAndBytes)
- **NF4 (4-bit)** – Maximum compression using BitsAndBytes double quant
---
## Model Information
- **Model Name:** LFM 2.5 Base
- **Parameters:** ~1.2B
- **Architecture:** Custom LiquidAI architecture
- **Original Author:** LiquidAI
- **Quantized By:** ArkAiLab (Open4bits)
This model **requires** `trust_remote_code=True` when loading.
---
## Quantization Details
- Quantized using **PyTorch** and **Hugging Face Transformers**
- INT8 and NF4 formats use **BitsAndBytes**
- FP8 provided where hardware support allows
- No GPTQ, AWQ, or llama.cpp used
- Safe for **Google Colab** and **Kaggle**
---
## Usage Example
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model_id = "Open4bits/LFM2.5-1.2B-Base-Quantized"
tokenizer = AutoTokenizer.from_pretrained(
model_id,
trust_remote_code=True
)
model = AutoModelForCausalLM.from_pretrained(
model_id,
trust_remote_code=True,
device_map="auto"
)
inputs = tokenizer("Hello, world!", return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=50)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
```
---
## Organization
This repository is maintained by **ArkAiLab** under the **Open4bits** initiative.
ArkAiLab (Main Organization):
https://huggingface.co/ArkAiLab-Adl
Open4bits (Quantization Projects):
https://huggingface.co/Open4bits
---
## License
This repository follows the **same license** as the original LiquidAI model.
Please refer to the original model repository for full licensing details.
---
## Disclaimer
This is an **unofficial quantized release**.
All credit for the original model architecture and training goes to **LiquidAI**.