itsme-nishanth commited on
Commit
0871ce8
·
verified ·
1 Parent(s): ea1cf6d

Update README.md

Browse files

updated the parameters number

Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -4,14 +4,14 @@ license: apache-2.0
4
  ---
5
  # 🧠 JAT-GPT: Just Another Tiny GPT
6
 
7
- Welcome to **JAT-GPT**, the world's most underwhelming large language model — clocking in at a mighty **74 million parameters** (yes, million, not billion — stop laughing).
8
 
9
  ## 📦 Model Details
10
 
11
  - **Model type**: GPT2-based decoder-only transformer
12
  - **Architecture**: GPT-2
13
  - **Library**: Hugging Face 🤗 Transformers
14
- - **Parameters**: 74 million (size isn't everything... right?)
15
  - **Training Objective**: Learn to predict the next word — and sometimes even the *right* one!
16
  - **Pretrained on**: A secret* dataset (*"secret" means the dataset was just some text I could find lying around)
17
  - **Training Purpose**: Solely educational. Also for flexing on friends who haven’t trained a language model from scratch.
 
4
  ---
5
  # 🧠 JAT-GPT: Just Another Tiny GPT
6
 
7
+ Welcome to **JAT-GPT**, the world's most underwhelming large language model — clocking in at a mighty **17.9 million parameters** (yes, million, not billion — stop laughing).
8
 
9
  ## 📦 Model Details
10
 
11
  - **Model type**: GPT2-based decoder-only transformer
12
  - **Architecture**: GPT-2
13
  - **Library**: Hugging Face 🤗 Transformers
14
+ - **Parameters**: 17.9 million (size isn't everything... right?)
15
  - **Training Objective**: Learn to predict the next word — and sometimes even the *right* one!
16
  - **Pretrained on**: A secret* dataset (*"secret" means the dataset was just some text I could find lying around)
17
  - **Training Purpose**: Solely educational. Also for flexing on friends who haven’t trained a language model from scratch.