Safetensors
English
qwen2
nielsr HF Staff commited on
Commit
a3c877a
·
verified ·
1 Parent(s): 671faa9

Update model card with metadata and links

Browse files

Hi! I've updated the model card to include relevant metadata and links. Specifically:
- Added `library_name: transformers` and `pipeline_tag: text-generation` to the metadata for better discoverability.
- Included links to the official paper and the GitHub repository.
- Maintained the existing sample usage and citation.

These changes will help make the model more discoverable and provide users with the necessary context from your research.

Files changed (1) hide show
  1. README.md +10 -5
README.md CHANGED
@@ -1,14 +1,21 @@
1
  ---
2
- license: mit
3
  datasets:
4
  - HuggingFaceTB/smollm-corpus
5
  language:
6
  - en
 
 
 
7
  ---
8
 
9
  # Raw 1B Shared
10
 
 
11
 
 
 
 
 
12
 
13
  ## How to Get Started with the Model
14
  Use the code below to get started with the model.
@@ -25,7 +32,7 @@ tokenizer = AutoTokenizer.from_pretrained(
25
 
26
 
27
  ## Citation
28
- ```
29
  @article{yamaguchi2026enhancinglinguisticcompetencelanguage,
30
  title={Enhancing Linguistic Competence of Language Models through Pre-training with Language Learning Tasks},
31
  author={Atsuki Yamaguchi and Maggie Mi and Nikolaos Aletras},
@@ -37,6 +44,4 @@ tokenizer = AutoTokenizer.from_pretrained(
37
  journal={arXiv},
38
  volume={abs/2601.03448}
39
  }
40
- ```
41
-
42
-
 
1
  ---
 
2
  datasets:
3
  - HuggingFaceTB/smollm-corpus
4
  language:
5
  - en
6
+ license: mit
7
+ library_name: transformers
8
+ pipeline_tag: text-generation
9
  ---
10
 
11
  # Raw 1B Shared
12
 
13
+ This model is a 1B parameter language model pre-trained as a baseline for the research presented in the paper [Enhancing Linguistic Competence of Language Models through Pre-training with Language Learning Tasks](https://huggingface.co/papers/2601.03448).
14
 
15
+ L2T (Language Learning Tasks) is a pre-training framework that integrates structured linguistic tasks alongside standard next-token prediction to explicitly optimize for linguistic competence in Large Language Models (LLMs). This specific checkpoint is the baseline model trained on raw text.
16
+
17
+ - **Paper:** [Enhancing Linguistic Competence of Language Models through Pre-training with Language Learning Tasks](https://huggingface.co/papers/2601.03448)
18
+ - **Repository:** [gucci-j/l2t](https://github.com/gucci-j/l2t)
19
 
20
  ## How to Get Started with the Model
21
  Use the code below to get started with the model.
 
32
 
33
 
34
  ## Citation
35
+ ```bibtex
36
  @article{yamaguchi2026enhancinglinguisticcompetencelanguage,
37
  title={Enhancing Linguistic Competence of Language Models through Pre-training with Language Learning Tasks},
38
  author={Atsuki Yamaguchi and Maggie Mi and Nikolaos Aletras},
 
44
  journal={arXiv},
45
  volume={abs/2601.03448}
46
  }
47
+ ```