Safetensors
English
qwen2
nielsr HF Staff commited on
Commit
e8d0f1e
·
verified ·
1 Parent(s): 671453c

Add library_name, pipeline_tag and link to code

Browse files

Hi! I'm Niels from the community science team at Hugging Face. I've opened this PR to enhance the model card for L2T 1B Shared.

The changes include:
- Adding the `text-generation` pipeline tag for better discoverability.
- Specifying the `transformers` library name to enable automated code snippets.
- Including direct links to the paper and the GitHub repository.
- Providing a brief description of the model based on the associated paper.

Please let me know if you have any questions!

Files changed (1) hide show
  1. README.md +10 -5
README.md CHANGED
@@ -1,14 +1,21 @@
1
  ---
2
- license: mit
3
  datasets:
4
  - HuggingFaceTB/smollm-corpus
5
  language:
6
  - en
 
 
 
7
  ---
8
 
9
  # L2T 1B Shared
10
 
 
11
 
 
 
 
 
12
 
13
  ## How to Get Started with the Model
14
  Use the code below to get started with the model.
@@ -25,7 +32,7 @@ tokenizer = AutoTokenizer.from_pretrained(
25
 
26
 
27
  ## Citation
28
- ```
29
  @article{yamaguchi2026enhancinglinguisticcompetencelanguage,
30
  title={Enhancing Linguistic Competence of Language Models through Pre-training with Language Learning Tasks},
31
  author={Atsuki Yamaguchi and Maggie Mi and Nikolaos Aletras},
@@ -37,6 +44,4 @@ tokenizer = AutoTokenizer.from_pretrained(
37
  journal={arXiv},
38
  volume={abs/2601.03448}
39
  }
40
- ```
41
-
42
-
 
1
  ---
 
2
  datasets:
3
  - HuggingFaceTB/smollm-corpus
4
  language:
5
  - en
6
+ license: mit
7
+ library_name: transformers
8
+ pipeline_tag: text-generation
9
  ---
10
 
11
  # L2T 1B Shared
12
 
13
+ This model is a 1B parameter language model pre-trained using the **L2T** framework, as presented in [Enhancing Linguistic Competence of Language Models through Pre-training with Language Learning Tasks](https://huggingface.co/papers/2601.03448).
14
 
15
+ L2T (Language Learning Tasks) is a pre-training framework that integrates specific language learning tasks alongside standard next-token prediction. Inspired by human language acquisition, L2T transforms raw text into structured input-output pairs to provide explicit linguistic stimulation, improving linguistic competence while maintaining competitive performance on general reasoning tasks.
16
+
17
+ - **Paper:** [Enhancing Linguistic Competence of Language Models through Pre-training with Language Learning Tasks](https://huggingface.co/papers/2601.03448)
18
+ - **Code:** [https://github.com/gucci-j/l2t](https://github.com/gucci-j/l2t)
19
 
20
  ## How to Get Started with the Model
21
  Use the code below to get started with the model.
 
32
 
33
 
34
  ## Citation
35
+ ```bibtex
36
  @article{yamaguchi2026enhancinglinguisticcompetencelanguage,
37
  title={Enhancing Linguistic Competence of Language Models through Pre-training with Language Learning Tasks},
38
  author={Atsuki Yamaguchi and Maggie Mi and Nikolaos Aletras},
 
44
  journal={arXiv},
45
  volume={abs/2601.03448}
46
  }
47
+ ```