Add library_name, pipeline_tag and link to code
Browse filesHi! I'm Niels from the community science team at Hugging Face. I've opened this PR to enhance the model card for L2T 1B Shared.
The changes include:
- Adding the `text-generation` pipeline tag for better discoverability.
- Specifying the `transformers` library name to enable automated code snippets.
- Including direct links to the paper and the GitHub repository.
- Providing a brief description of the model based on the associated paper.
Please let me know if you have any questions!
README.md
CHANGED
|
@@ -1,14 +1,21 @@
|
|
| 1 |
---
|
| 2 |
-
license: mit
|
| 3 |
datasets:
|
| 4 |
- HuggingFaceTB/smollm-corpus
|
| 5 |
language:
|
| 6 |
- en
|
|
|
|
|
|
|
|
|
|
| 7 |
---
|
| 8 |
|
| 9 |
# L2T 1B Shared
|
| 10 |
|
|
|
|
| 11 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 12 |
|
| 13 |
## How to Get Started with the Model
|
| 14 |
Use the code below to get started with the model.
|
|
@@ -25,7 +32,7 @@ tokenizer = AutoTokenizer.from_pretrained(
|
|
| 25 |
|
| 26 |
|
| 27 |
## Citation
|
| 28 |
-
```
|
| 29 |
@article{yamaguchi2026enhancinglinguisticcompetencelanguage,
|
| 30 |
title={Enhancing Linguistic Competence of Language Models through Pre-training with Language Learning Tasks},
|
| 31 |
author={Atsuki Yamaguchi and Maggie Mi and Nikolaos Aletras},
|
|
@@ -37,6 +44,4 @@ tokenizer = AutoTokenizer.from_pretrained(
|
|
| 37 |
journal={arXiv},
|
| 38 |
volume={abs/2601.03448}
|
| 39 |
}
|
| 40 |
-
```
|
| 41 |
-
|
| 42 |
-
|
|
|
|
| 1 |
---
|
|
|
|
| 2 |
datasets:
|
| 3 |
- HuggingFaceTB/smollm-corpus
|
| 4 |
language:
|
| 5 |
- en
|
| 6 |
+
license: mit
|
| 7 |
+
library_name: transformers
|
| 8 |
+
pipeline_tag: text-generation
|
| 9 |
---
|
| 10 |
|
| 11 |
# L2T 1B Shared
|
| 12 |
|
| 13 |
+
This model is a 1B parameter language model pre-trained using the **L2T** framework, as presented in [Enhancing Linguistic Competence of Language Models through Pre-training with Language Learning Tasks](https://huggingface.co/papers/2601.03448).
|
| 14 |
|
| 15 |
+
L2T (Language Learning Tasks) is a pre-training framework that integrates specific language learning tasks alongside standard next-token prediction. Inspired by human language acquisition, L2T transforms raw text into structured input-output pairs to provide explicit linguistic stimulation, improving linguistic competence while maintaining competitive performance on general reasoning tasks.
|
| 16 |
+
|
| 17 |
+
- **Paper:** [Enhancing Linguistic Competence of Language Models through Pre-training with Language Learning Tasks](https://huggingface.co/papers/2601.03448)
|
| 18 |
+
- **Code:** [https://github.com/gucci-j/l2t](https://github.com/gucci-j/l2t)
|
| 19 |
|
| 20 |
## How to Get Started with the Model
|
| 21 |
Use the code below to get started with the model.
|
|
|
|
| 32 |
|
| 33 |
|
| 34 |
## Citation
|
| 35 |
+
```bibtex
|
| 36 |
@article{yamaguchi2026enhancinglinguisticcompetencelanguage,
|
| 37 |
title={Enhancing Linguistic Competence of Language Models through Pre-training with Language Learning Tasks},
|
| 38 |
author={Atsuki Yamaguchi and Maggie Mi and Nikolaos Aletras},
|
|
|
|
| 44 |
journal={arXiv},
|
| 45 |
volume={abs/2601.03448}
|
| 46 |
}
|
| 47 |
+
```
|
|
|
|
|
|