MicroGPT-Go / README.md
Traves's picture
Upload README.md with huggingface_hub
344fa8c verified
---
language: en
license: mit
tags:
- microgpt
- go
- transformer
- local-training
---
# MicroGPT - 1500 Step Checkpoint
This is a local training checkpoint for **MicroGPT (Go Edition)**. It features a compact transformer architecture trained on a specific dataset using the `microgpt` Go runtime.
## Model Details
- **Architecture**: Transformer
- **Layers**: 2
- **Embedding Dimensions**: 80
- **Heads**: 4
- **Block Size**: 128
- **Tokenization**: BPE (cl100k_base)
- **Training Steps**: 1500
- **Validation Loss**: 4.4379
## Usage
This model is designed to be run with the [MicroGPT Go Edition](https://github.com/Traves-Theberge/microgpt-tui-go) engine.
### Loading the Model
1. Download `latest_checkpoint.json` from this repository.
2. Place it in your `go/models/` directory.
3. Use the MicroGPT CLI or TUI to load the model:
```bash
go run . chat-once models/latest_checkpoint.json "Your prompt here"
```
## Training Data
The model was trained using the `databricks-dolly-15k.jsonl` dataset (or variant) as configured in the training hub.
## Attribution
Built on the foundational ideas of Andrej Karpathy.