Add library_name and pipeline_tag to metadata
Browse filesThis PR adds `library_name` and `pipeline_tag` to the model card metadata.
The `library_name` is set to `transformers` as the models are compatible with the Hugging Face Transformers library. The `pipeline_tag` is set to `text-generation` as these are language models.
README.md
CHANGED
|
@@ -1,6 +1,11 @@
|
|
| 1 |
---
|
| 2 |
license: apache-2.0
|
|
|
|
|
|
|
|
|
|
|
|
|
| 3 |
---
|
|
|
|
| 4 |
**Paper**: [https://arxiv.org/pdf/2502.07780](https://arxiv.org/pdf/2502.07780)
|
| 5 |
**Code**: https://github.com/IST-DASLab/DarwinLM
|
| 6 |
**Models**: [DarwinLM-2.7B](https://huggingface.co/Shengkun/DarwinLM-2.7B), [DarwinLM-4.6B](https://huggingface.co/Shengkun/DarwinLM-4.6B), [DarwinLM-8.4B](https://huggingface.co/Shengkun/DarwinLM-8.4B)
|
|
@@ -9,14 +14,13 @@ license: apache-2.0
|
|
| 9 |
---
|
| 10 |
|
| 11 |
This repository contains the weights of DarwinLM, an evolutionary structured pruning methods for large language models, as introduced in our paper. DarwinLM builds upon an evolutionary search process, generating multiple offspring models in each generation through mutation, and selecting the fittest for survival.
|
| 12 |
-
```
|
| 13 |
# Please add trust_remote_code=True as the repo includes custom code to load and run DarwinLM
|
| 14 |
model = AutoModelForCausalLM.from_pretrained("Shengkun/DarwinLM-2.7B", trust_remote_code=True)
|
| 15 |
```
|
| 16 |
|
| 17 |
## Downstream Tasks
|
| 18 |
|
| 19 |
-
|
| 20 |
**2.7B**
|
| 21 |
|
| 22 |
| Method | Param. | SciQ | PIQA | WG | ArcE | ArcC | HS | LogiQA | BoolQ | Avg |
|
|
@@ -52,8 +56,6 @@ model = AutoModelForCausalLM.from_pretrained("Shengkun/DarwinLM-2.7B", trust_rem
|
|
| 52 |
| | **OLMO-0424 (2.05T)** | 7B | 96.1 | 80.1 | 72.1 | 73.8 | 49.2 | 78.0 | 29.3 | 80.8 | 52.1 | 67.9 |
|
| 53 |
| | *DarwinLM (10.0B)* | 8.4B | 89.5 | 78.1 | 70.7 | 79.6 | 57.6 | 74.9 | 33.5 | 73.9 | 57.9 | 68.4 |
|
| 54 |
|
| 55 |
-
|
| 56 |
-
|
| 57 |
## Bibtex
|
| 58 |
```
|
| 59 |
@article{tang2025darwinlm,
|
|
|
|
| 1 |
---
|
| 2 |
license: apache-2.0
|
| 3 |
+
library_name: transformers
|
| 4 |
+
pipeline_tag: text-generation
|
| 5 |
+
tags:
|
| 6 |
+
- code-generation
|
| 7 |
---
|
| 8 |
+
|
| 9 |
**Paper**: [https://arxiv.org/pdf/2502.07780](https://arxiv.org/pdf/2502.07780)
|
| 10 |
**Code**: https://github.com/IST-DASLab/DarwinLM
|
| 11 |
**Models**: [DarwinLM-2.7B](https://huggingface.co/Shengkun/DarwinLM-2.7B), [DarwinLM-4.6B](https://huggingface.co/Shengkun/DarwinLM-4.6B), [DarwinLM-8.4B](https://huggingface.co/Shengkun/DarwinLM-8.4B)
|
|
|
|
| 14 |
---
|
| 15 |
|
| 16 |
This repository contains the weights of DarwinLM, an evolutionary structured pruning methods for large language models, as introduced in our paper. DarwinLM builds upon an evolutionary search process, generating multiple offspring models in each generation through mutation, and selecting the fittest for survival.
|
| 17 |
+
```python
|
| 18 |
# Please add trust_remote_code=True as the repo includes custom code to load and run DarwinLM
|
| 19 |
model = AutoModelForCausalLM.from_pretrained("Shengkun/DarwinLM-2.7B", trust_remote_code=True)
|
| 20 |
```
|
| 21 |
|
| 22 |
## Downstream Tasks
|
| 23 |
|
|
|
|
| 24 |
**2.7B**
|
| 25 |
|
| 26 |
| Method | Param. | SciQ | PIQA | WG | ArcE | ArcC | HS | LogiQA | BoolQ | Avg |
|
|
|
|
| 56 |
| | **OLMO-0424 (2.05T)** | 7B | 96.1 | 80.1 | 72.1 | 73.8 | 49.2 | 78.0 | 29.3 | 80.8 | 52.1 | 67.9 |
|
| 57 |
| | *DarwinLM (10.0B)* | 8.4B | 89.5 | 78.1 | 70.7 | 79.6 | 57.6 | 74.9 | 33.5 | 73.9 | 57.9 | 68.4 |
|
| 58 |
|
|
|
|
|
|
|
| 59 |
## Bibtex
|
| 60 |
```
|
| 61 |
@article{tang2025darwinlm,
|