QED-75M_artifacts / README.md
levossadtchi's picture
Update README.md
7e3dd98 verified
---
license: mit
tags:
- training-artifacts
- checkpoints
- logs
- causal-lm
---
![Frame 33](https://cdn-uploads.huggingface.co/production/uploads/695b8d7a2114f706bdcee465/U7--7-pw6htcWzk_QPrdv.png)
# QED-75M Artifacts
This repository stores training artifacts for **QED-75M**.
## What is inside
- Training checkpoints (`.pt`)
- Training/evaluation logs
- Auxiliary files used for reproducibility (configs, summaries, intermediate outputs)
## Related model repository
- Main model card and inference-ready model: [levossadtchi/QED-75M](https://huggingface.co/levossadtchi/QED-75M). It contains a checkpoint at step 7400 SFT step.
## Training summary
- Pretraining data volume: **12.6B tokens**
- Multi-stage pipeline: pretraining -> long-context annealing -> SFT
## Notes
- These files are intended for reproducibility, inspection, and research workflows.
- For normal inference, use the main model repository instead of this artifacts repository.
- data/pretokenized contains pretokenized data for stage 1.