File size: 1,326 Bytes
c14a9ef | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 | ---
license: mit
task_categories:
- other
tags:
- ai-models
- training-data
- tokens
- machine-learning
size_categories:
- n<1K
---
# Training Data Scale Registry
A systematic registry of AI model training data size estimates with evidence profiles.
## Dataset Description
This dataset contains structured records of AI models with:
- Token count estimates (min/max/mid)
- Evidence types (E1-E5) and strength (S-High/Medium/Low)
- Uncertainty sources (U1-U5)
- Model metadata (parameters, FLOPs, architecture)
- Raw evidence snippets
## Data Collection
Data is collected from:
- Epoch AI datasets
- Hugging Face model cards
- Technical reports and system cards
- Third-party analyses
## Inference Methods
Token estimates are derived using:
- Chinchilla scaling law
- Hardware back-calculation
- Parameter ratio heuristics
- Textual token clues
- Third-party analyses
## Evidence Profiles
Each model includes an evidence profile indicating:
- **Evidence Types**: How the estimate was derived
- **Evidence Strength**: Confidence in the estimate
- **Uncertainty Sources**: What information is missing
## Usage
```python
from datasets import load_dataset
dataset = load_dataset("midah/odl-training-data")
```
## Citation
If you use this dataset, please cite:
```
Training Data Scale Registry
ODL Research
2025
```
|