Safetensors
English
qwen2
raw-1b-shared / README.md
nielsr's picture
nielsr HF Staff
Update model card with metadata and links
a3c877a verified
|
raw
history blame
1.72 kB
metadata
datasets:
  - HuggingFaceTB/smollm-corpus
language:
  - en
license: mit
library_name: transformers
pipeline_tag: text-generation

Raw 1B Shared

This model is a 1B parameter language model pre-trained as a baseline for the research presented in the paper Enhancing Linguistic Competence of Language Models through Pre-training with Language Learning Tasks.

L2T (Language Learning Tasks) is a pre-training framework that integrates structured linguistic tasks alongside standard next-token prediction to explicitly optimize for linguistic competence in Large Language Models (LLMs). This specific checkpoint is the baseline model trained on raw text.

How to Get Started with the Model

Use the code below to get started with the model.

from transformers import AutoTokenizer, AutoModelForCausalLM

model = AutoModelForCausalLM.from_pretrained(
    "l2t-project/raw-1b-shared"
)
tokenizer = AutoTokenizer.from_pretrained(
    "l2t-project/raw-1b-shared"
)

Citation

@article{yamaguchi2026enhancinglinguisticcompetencelanguage,
      title={Enhancing Linguistic Competence of Language Models through Pre-training with Language Learning Tasks}, 
      author={Atsuki Yamaguchi and Maggie Mi and Nikolaos Aletras},
      year={2026},
      eprint={2601.03448},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2601.03448},
      journal={arXiv},
      volume={abs/2601.03448}
}