Improve model card: Add pipeline tag, enrich tags, update license, and add top-level links

#3
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +16 -7
README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- library_name: transformers
3
  language:
4
  - ar
5
  - cs
@@ -19,18 +19,25 @@ language:
19
  - sv
20
  - ur
21
  - zh
 
 
22
  tags:
23
  - falcon-h1
24
- license: other
25
- license_name: falcon-llm-license
26
- license_link: https://falconllm.tii.ae/falcon-terms-and-conditions.html
27
- base_model: tiiuae/Falcon-H1-7B-Base
 
 
28
  inference: true
 
29
  ---
30
 
 
 
31
  <img src="https://huggingface.co/datasets/tiiuae/documentation-images/resolve/main/falcon_mamba/falcon-h1-logo.png" alt="drawing" width="800"/>
32
 
33
- # Table of Contents
34
 
35
  0. [TL;DR](#TL;DR)
36
  1. [Model Details](#model-details)
@@ -40,6 +47,7 @@ inference: true
40
  5. [Citation](#citation)
41
 
42
  # TL;DR
 
43
 
44
  # Model Details
45
 
@@ -49,7 +57,7 @@ inference: true
49
  - **Model type:** Causal decoder-only
50
  - **Architecture:** Hybrid Transformers + Mamba architecture
51
  - **Language(s) (NLP):** English, Multilingual
52
- - **License:** Falcon-LLM License
53
 
54
  # Training details
55
 
@@ -147,6 +155,7 @@ You can check more in detail on our [our release blogpost](https://falcon-lm.git
147
 
148
  - View [our release blogpost](https://falcon-lm.github.io/blog/falcon-h1/).
149
  - View [our technical report](https://arxiv.org/abs/2507.22448).
 
150
  - Feel free to join [our discord server](https://discord.gg/trwMYP9PYm) if you have any questions or to interact with our researchers and developers.
151
 
152
  # Citation
 
1
  ---
2
+ base_model: tiiuae/Falcon-H1-7B-Base
3
  language:
4
  - ar
5
  - cs
 
19
  - sv
20
  - ur
21
  - zh
22
+ library_name: transformers
23
+ license: apache-2.0
24
  tags:
25
  - falcon-h1
26
+ - multilingual
27
+ - code-generation
28
+ - math
29
+ - reasoning
30
+ - instruction-tuned
31
+ - science
32
  inference: true
33
+ pipeline_tag: text-generation
34
  ---
35
 
36
+ [\ud83d\udcc4 Paper](https://huggingface.co/papers/2507.22448) | [\ud83d\udcbb Code](https://github.com/tiiuae/Falcon-H1) | [\ud83c\udfe0 Project Page](https://tiiuae.github.io/Falcon-H1/)
37
+
38
  <img src="https://huggingface.co/datasets/tiiuae/documentation-images/resolve/main/falcon_mamba/falcon-h1-logo.png" alt="drawing" width="800"/>
39
 
40
+ # Table of Contents
41
 
42
  0. [TL;DR](#TL;DR)
43
  1. [Model Details](#model-details)
 
47
  5. [Citation](#citation)
48
 
49
  # TL;DR
50
+ Falcon-H1 is a new family of hybrid-head Large Language Models (LLMs) ranging from 0.5B to 34B parameters. They combine Transformer-based attention with State Space Models (SSMs) for superior long-context memory and computational efficiency. These models achieve state-of-the-art performance across reasoning, mathematics, multilingual tasks, instruction following, and scientific knowledge, outperforming larger models in efficiency and performance. They support up to 256K context tokens and 18 languages, and are released under a permissive open-source license.
51
 
52
  # Model Details
53
 
 
57
  - **Model type:** Causal decoder-only
58
  - **Architecture:** Hybrid Transformers + Mamba architecture
59
  - **Language(s) (NLP):** English, Multilingual
60
+ - **License:** Apache 2.0
61
 
62
  # Training details
63
 
 
155
 
156
  - View [our release blogpost](https://falcon-lm.github.io/blog/falcon-h1/).
157
  - View [our technical report](https://arxiv.org/abs/2507.22448).
158
+ - View [our official documentation](https://tiiuae.github.io/Falcon-H1/).
159
  - Feel free to join [our discord server](https://discord.gg/trwMYP9PYm) if you have any questions or to interact with our researchers and developers.
160
 
161
  # Citation