File size: 3,157 Bytes
a5dc8ab 233897f a5dc8ab 233897f a5dc8ab 233897f a5dc8ab 233897f a5dc8ab 233897f a5dc8ab 233897f a5dc8ab 233897f a5dc8ab 233897f a5dc8ab 233897f a5dc8ab ac5b2d0 a5dc8ab 233897f a5dc8ab 233897f a5dc8ab 233897f a5dc8ab 233897f a5dc8ab 233897f a5dc8ab 233897f a5dc8ab 16a3d81 a5dc8ab 233897f a5dc8ab 233897f a5dc8ab 233897f a5dc8ab 233897f 16a3d81 233897f a5dc8ab 233897f a5dc8ab 233897f a5dc8ab 233897f 16a3d81 233897f a5dc8ab 233897f a5dc8ab 233897f 16a3d81 233897f a5dc8ab 233897f a5dc8ab 233897f ac5b2d0 233897f a5dc8ab 233897f a5dc8ab 233897f a5dc8ab 233897f 16a3d81 233897f a5dc8ab 233897f a5dc8ab 233897f 16a3d81 233897f a5dc8ab 233897f a5dc8ab ac5b2d0 a5dc8ab 233897f a5dc8ab e4f18b8 233897f a5dc8ab 233897f a5dc8ab ac5b2d0 a5dc8ab 233897f 5020392 233897f ac5b2d0 233897f |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 |
---
library_name: transformers
tags:
- falcon-h1
- edge
license: other
license_name: falcon-llm-license
license_link: https://falconllm.tii.ae/falcon-terms-and-conditions.html
---
<img src="https://cdn-uploads.huggingface.co/production/uploads/62441d1d9fdefb55a0b7d12c/l1du02RjuAZJcksI5tQ-F.png" alt="drawing" width="800"/>
# Table of Contents
0. [TL;DR](#TL;DR)
1. [Model Details](#model-details)
2. [Training Details](#training-details)
3. [Usage](#usage)
4. [Evaluation](#evaluation)
5. [Citation](#citation)
# TL;DR
# Model Details
## Model Description
- **Developed by:** [https://www.tii.ae](https://www.tii.ae)
- **Model type:** Causal decoder-only
- **Architecture:** Hybrid Transformers + Mamba architecture
- **Language(s) (NLP):** English
- **Number of Parameters:** 90M
- **License:** Falcon-LLM License
# Training details
For more details about the training protocol of this model, please refer to the [Falcon-H1-Tiny technical blogpost](https://huggingface.co/spaces/tiiuae/tiny-h1-blogpost).
# Usage
Currently to use this model you can either rely on Hugging Face `transformers`, `vLLM`, `sglang`, `llama.cpp`, `ollama` or `mlx` library.
## Inference
### 🤗 transformers
Refer to the snippet below to run H1 models using 🤗 transformers:
```python
import torch
from transformers import AutoModelForCausalLM, AutoTokenizer
model_id = "tiiuae/Falcon-H1-Tiny-R-90M"
model = AutoModelForCausalLM.from_pretrained(
model_id,
torch_dtype=torch.bfloat16,
device_map="auto"
)
# Perform text generation
```
or
```bash
transformers serve tiiuae/Falcon-H1-Tiny-R-90M
```
### `llama.cpp`
You can find all GGUF files compatible with `llama.cpp` under [our official collection]() - an example setup could be:
```bash
brew install llama.cpp
pip install huggingface_hub
hf download tiiuae/Falcon-H1-Tiny-R-90M-GGUF Falcon-H1-Tiny-R-90M-Q8_0.gguf --local-dir ./
llama-cli ./Falcon-H1-Tiny-R-90M-Q8_0.gguf -cnv
```
### `ollama`
```bash
ollama run hf.co/tiiuae/Falcon-H1-Tiny-R-90M:Q8_0
```
### Apple `mlx`
```bash
mlx_lm.chat --model tiiuae/Falcon-H1-Tiny-R-90M
```
### vLLM
For vLLM, simply start a server by executing the command below:
```bash
# pip install vllm>=0.9.0
vllm serve tiiuae/Falcon-H1-Tiny-R-90M --tensor-parallel-size 2 --data-parallel-size 1
```
### sglang
```bash
python -m sglang.launch_server \
--model ttiiuae/Falcon-H1-Tiny-R-90M \
--tensor-parallel-size 1
```
# Evaluation
For detailed evaluation of Falcon-H1-Tiny series, please refer to our [technical blogpost](https://huggingface.co/spaces/tiiuae/tiny-h1-blogpost)
# Useful links
- View [our release blogpost](https://huggingface.co/spaces/tiiuae/tiny-h1-blogpost).
- Feel free to join [our discord server](https://discord.gg/trwMYP9PYm) if you have any questions or to interact with our researchers and developers.
# Citation
If the Falcon-H1-Tiny family of models were helpful to your work, feel free to give us a cite.
```
@misc{falcon_h1_tiny,
title={Falcon-H1-Tiny: A series of extremely small, yet powerful language models redefining capabilities at small scale},
author={Falcon-LLM Team},
year={2026},
}
``` |