Safetensors
English
qwen2

Add pipeline_tag, library_name, and links to paper/code

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +10 -6
README.md CHANGED
@@ -1,14 +1,21 @@
1
  ---
2
- license: mit
3
  datasets:
4
  - HuggingFaceTB/smollm-corpus
5
  language:
6
  - en
 
 
 
7
  ---
8
 
9
  # L2T 500M Shared
10
 
 
11
 
 
 
 
 
12
 
13
  ## How to Get Started with the Model
14
  Use the code below to get started with the model.
@@ -23,9 +30,8 @@ tokenizer = AutoTokenizer.from_pretrained(
23
  )
24
  ```
25
 
26
-
27
  ## Citation
28
- ```
29
  @article{yamaguchi2026enhancinglinguisticcompetencelanguage,
30
  title={Enhancing Linguistic Competence of Language Models through Pre-training with Language Learning Tasks},
31
  author={Atsuki Yamaguchi and Maggie Mi and Nikolaos Aletras},
@@ -37,6 +43,4 @@ tokenizer = AutoTokenizer.from_pretrained(
37
  journal={arXiv},
38
  volume={abs/2601.03448}
39
  }
40
- ```
41
-
42
-
 
1
  ---
 
2
  datasets:
3
  - HuggingFaceTB/smollm-corpus
4
  language:
5
  - en
6
+ license: mit
7
+ pipeline_tag: text-generation
8
+ library_name: transformers
9
  ---
10
 
11
  # L2T 500M Shared
12
 
13
+ This repository contains the **L2T 500M Shared** model, as introduced in the paper [Enhancing Linguistic Competence of Language Models through Pre-training with Language Learning Tasks](https://huggingface.co/papers/2601.03448).
14
 
15
+ **L2T** (Language Learning Tasks) is a pre-training framework that integrates structured linguistic tasks alongside standard next-token prediction. Inspired by human language acquisition, L2T transforms raw text into structured input-output pairs to provide explicit linguistic stimulation, improving linguistic competence while maintaining competitive performance on general reasoning tasks.
16
+
17
+ - **Paper:** [Enhancing Linguistic Competence of Language Models through Pre-training with Language Learning Tasks](https://huggingface.co/papers/2601.03448)
18
+ - **Code:** [GitHub - gucci-j/l2t](https://github.com/gucci-j/l2t)
19
 
20
  ## How to Get Started with the Model
21
  Use the code below to get started with the model.
 
30
  )
31
  ```
32
 
 
33
  ## Citation
34
+ ```bibtex
35
  @article{yamaguchi2026enhancinglinguisticcompetencelanguage,
36
  title={Enhancing Linguistic Competence of Language Models through Pre-training with Language Learning Tasks},
37
  author={Atsuki Yamaguchi and Maggie Mi and Nikolaos Aletras},
 
43
  journal={arXiv},
44
  volume={abs/2601.03448}
45
  }
46
+ ```