Update README.md
Browse filesupdated the parameters number
README.md
CHANGED
|
@@ -4,14 +4,14 @@ license: apache-2.0
|
|
| 4 |
---
|
| 5 |
# 🧠 JAT-GPT: Just Another Tiny GPT
|
| 6 |
|
| 7 |
-
Welcome to **JAT-GPT**, the world's most underwhelming large language model — clocking in at a mighty **
|
| 8 |
|
| 9 |
## 📦 Model Details
|
| 10 |
|
| 11 |
- **Model type**: GPT2-based decoder-only transformer
|
| 12 |
- **Architecture**: GPT-2
|
| 13 |
- **Library**: Hugging Face 🤗 Transformers
|
| 14 |
-
- **Parameters**:
|
| 15 |
- **Training Objective**: Learn to predict the next word — and sometimes even the *right* one!
|
| 16 |
- **Pretrained on**: A secret* dataset (*"secret" means the dataset was just some text I could find lying around)
|
| 17 |
- **Training Purpose**: Solely educational. Also for flexing on friends who haven’t trained a language model from scratch.
|
|
|
|
| 4 |
---
|
| 5 |
# 🧠 JAT-GPT: Just Another Tiny GPT
|
| 6 |
|
| 7 |
+
Welcome to **JAT-GPT**, the world's most underwhelming large language model — clocking in at a mighty **17.9 million parameters** (yes, million, not billion — stop laughing).
|
| 8 |
|
| 9 |
## 📦 Model Details
|
| 10 |
|
| 11 |
- **Model type**: GPT2-based decoder-only transformer
|
| 12 |
- **Architecture**: GPT-2
|
| 13 |
- **Library**: Hugging Face 🤗 Transformers
|
| 14 |
+
- **Parameters**: 17.9 million (size isn't everything... right?)
|
| 15 |
- **Training Objective**: Learn to predict the next word — and sometimes even the *right* one!
|
| 16 |
- **Pretrained on**: A secret* dataset (*"secret" means the dataset was just some text I could find lying around)
|
| 17 |
- **Training Purpose**: Solely educational. Also for flexing on friends who haven’t trained a language model from scratch.
|