addGPT / README.md
m-bano's picture
Update README.md
6af9e75 verified
---
license: apache-2.0
tags:
- transformer
- education
- arithmetic
- algorithmic-learning
language:
- en
pipeline_tag: text-generation
---
# addGPT: Teaching Transformers to Add
A transformer encoder-decoder that learns integer addition, demonstrating algorithmic learning capabilities.
## Model Description
- **Architecture**: Transformer encoder-decoder
- **Parameters**: ~2.7M (tiny config)
- **Task**: Integer addition (up to 5-digit numbers)
- **Accuracy**: >99% on held-out test set
## Usage
```python
# Download checkpoint
from huggingface_hub import hf_hub_download
checkpoint_path = hf_hub_download(repo_id="m-bano/addGPT", filename="ckpt_step12500.pt")
# Or use the provided script
# python download_checkpoint.py
```
See the [GitHub repository](https://github.com/mbano/addGPT) for complete code and usage instructions.
## Training Details
- **Framework**: PyTorch 2.0+
- **Training Time**: ~10 minutes on GPU
- **Dataset**: 6.4M randomly generated addition problems
- **Optimizer**: AdamW with cosine annealing
## Configuration
```yaml
n_emb: 512
n_heads: 8
n_blocks: 4
max_src_len: 5 # 5-digit operands
```
## License
Apache 2.0