llama-isolate-old / README.md
deqing's picture
Add model card
9ff6738 verified
---
library_name: transformers
tags:
- convergent-evolution
- fourier-features
- number-embeddings
license: mit
datasets:
- HuggingFaceFW/fineweb-edu
---
# convergent-llama-300M-muon-isolate
A 300M-parameter language model trained from scratch on **FineWeb-Edu 10BT** (~9.4B tokens, 1 epoch) as part of the *Convergent Evolution* project, which investigates how Fourier features emerge in LLM number embeddings.
## Model details
| | |
|---|---|
| **Architecture** | LLaMA-style Transformer (12 layers, 1024 hidden, 16 heads, GQA) |
| **Parameters** | ~300M |
| **Optimizer** | Muon (for 2D weights) + AdamW (for embeddings/bias/norm) |
| **Data perturbation** | block-diagonal attention mask (numbers cannot attend to context) |
| **Training data** | [FineWeb-Edu](https://huggingface.co/datasets/HuggingFaceFW/fineweb-edu) sample-10BT (~9.4B tokens) |
| **Context length** | 1024 |
| **Tokenizer** | Llama 3 (128K vocab) |
| **Batch size** | 512 sequences |
## Training dynamics
Intermediate checkpoints are saved as branches: `tokens-200M`, `tokens-400M`, ..., `tokens-9.6B`.
```python
from transformers import AutoModelForCausalLM
# Load final checkpoint
model = AutoModelForCausalLM.from_pretrained("deqing/convergent-llama-300M-muon-isolate")
# Load intermediate checkpoint (e.g., at 1B tokens)
model = AutoModelForCausalLM.from_pretrained("deqing/convergent-llama-300M-muon-isolate", revision="tokens-1B")
```
## Citation
Paper forthcoming.