Add model card and metadata

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +17 -5
README.md CHANGED
@@ -1,11 +1,24 @@
 
 
 
 
 
 
1
 
 
2
 
3
- <!-- Autocitation -->
4
 
5
- --------------------
 
 
 
 
 
 
6
 
7
  If you use this project in your research please cite:
8
- ```
9
  @article{patel2026fineinstructions,
10
  title={FineInstructions: Scaling Synthetic Instructions to Pre-Training Scale},
11
  author={Patel, Ajay and Raffel, Colin and Callison-Burch, Chris},
@@ -15,5 +28,4 @@ If you use this project in your research please cite:
15
  primaryClass={cs.CL},
16
  doi={10.48550/arXiv.2601.22146}
17
  }
18
- ```
19
-
 
1
+ ---
2
+ license: cc-by-nc-4.0
3
+ library_name: transformers
4
+ pipeline_tag: text-generation
5
+ ---
6
+ ---
7
 
8
+ # FineInstructions: Scaling Synthetic Instructions to Pre-Training Scale
9
 
10
+ This repository contains model checkpoints for the FineInstructions project, as introduced in the paper [FineInstructions: Scaling Synthetic Instructions to Pre-Training Scale](https://huggingface.co/papers/2601.22146).
11
 
12
+ ## Description
13
+
14
+ FineInstructions is a procedure that transforms internet-scale pre-training documents into billions of synthetic instruction and answer training pairs. The dataset uses ~18M instruction templates created from real user-written queries and prompts. These templates are matched to and instantiated with human-written source documents from unstructured pre-training corpora.
15
+
16
+ With "supervised" synthetic training data generated at this scale, an LLM can be pre-trained from scratch solely with the instruction-tuning objective. This approach is more in-distribution with the expected downstream usage of LLMs (responding to user prompts). Experimental results show that pre-training on FineInstructions outperforms standard pre-training on benchmarks measuring free-form response quality.
17
+
18
+ ## Citation
19
 
20
  If you use this project in your research please cite:
21
+ ```bibtex
22
  @article{patel2026fineinstructions,
23
  title={FineInstructions: Scaling Synthetic Instructions to Pre-Training Scale},
24
  author={Patel, Ajay and Raffel, Colin and Callison-Burch, Chris},
 
28
  primaryClass={cs.CL},
29
  doi={10.48550/arXiv.2601.22146}
30
  }
31
+ ```