nindanaoto commited on
Commit
56a6fe8
·
1 Parent(s): 18a89ae

First model version

Browse files
.gitattributes CHANGED
@@ -33,3 +33,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ *.wandb filter=lfs diff=lfs merge=lfs -text
37
+ *.log filter=lfs diff=lfs merge=lfs -text
LICENSE ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ MIT License
2
+
3
+ Copyright (c) 2022 Andrej Karpathy
4
+
5
+ Permission is hereby granted, free of charge, to any person obtaining a copy
6
+ of this software and associated documentation files (the "Software"), to deal
7
+ in the Software without restriction, including without limitation the rights
8
+ to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
9
+ copies of the Software, and to permit persons to whom the Software is
10
+ furnished to do so, subject to the following conditions:
11
+
12
+ The above copyright notice and this permission notice shall be included in all
13
+ copies or substantial portions of the Software.
14
+
15
+ THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16
+ IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17
+ FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
18
+ AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
19
+ LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
20
+ OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
21
+ SOFTWARE.
README.md ADDED
@@ -0,0 +1,230 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # nanoGPT-BitNet158b
2
+
3
+ This model implements BitNet1.58b integrated in nano-GPT model.
4
+
5
+ # nanoGPT
6
+
7
+ ![nanoGPT](assets/nanogpt.jpg)
8
+
9
+ The simplest, fastest repository for training/finetuning medium-sized GPTs. It is a rewrite of [minGPT](https://github.com/karpathy/minGPT) that prioritizes teeth over education. Still under active development, but currently the file `train.py` reproduces GPT-2 (124M) on OpenWebText, running on a single 8XA100 40GB node in about 4 days of training. The code itself is plain and readable: `train.py` is a ~300-line boilerplate training loop and `model.py` a ~300-line GPT model definition, which can optionally load the GPT-2 weights from OpenAI. That's it.
10
+
11
+ ![repro124m](assets/gpt2_124M_loss.png)
12
+
13
+ Because the code is so simple, it is very easy to hack to your needs, train new models from scratch, or finetune pretrained checkpoints (e.g. biggest one currently available as a starting point would be the GPT-2 1.3B model from OpenAI).
14
+
15
+ ## install
16
+
17
+ ```
18
+ pip install torch numpy transformers datasets tiktoken wandb tqdm
19
+ ```
20
+
21
+ Dependencies:
22
+
23
+ - [pytorch](https://pytorch.org) <3
24
+ - [numpy](https://numpy.org/install/) <3
25
+ - `transformers` for huggingface transformers <3 (to load GPT-2 checkpoints)
26
+ - `datasets` for huggingface datasets <3 (if you want to download + preprocess OpenWebText)
27
+ - `tiktoken` for OpenAI's fast BPE code <3
28
+ - `wandb` for optional logging <3
29
+ - `tqdm` for progress bars <3
30
+
31
+ ## quick start
32
+
33
+ If you are not a deep learning professional and you just want to feel the magic and get your feet wet, the fastest way to get started is to train a character-level GPT on the works of Shakespeare. First, we download it as a single (1MB) file and turn it from raw text into one large stream of integers:
34
+
35
+ ```
36
+ $ python data/shakespeare_char/prepare.py
37
+ ```
38
+
39
+ This creates a `train.bin` and `val.bin` in that data directory. Now it is time to train your GPT. The size of it very much depends on the computational resources of your system:
40
+
41
+ **I have a GPU**. Great, we can quickly train a baby GPT with the settings provided in the [config/train_shakespeare_char.py](config/train_shakespeare_char.py) config file:
42
+
43
+ ```
44
+ $ python train.py config/train_shakespeare_char.py
45
+ ```
46
+
47
+ If you peek inside it, you'll see that we're training a GPT with a context size of up to 256 characters, 384 feature channels, and it is a 6-layer Transformer with 6 heads in each layer. On one A100 GPU this training run takes about 3 minutes and the best validation loss is 1.4697. Based on the configuration, the model checkpoints are being written into the `--out_dir` directory `out-shakespeare-char`. So once the training finishes we can sample from the best model by pointing the sampling script at this directory:
48
+
49
+ ```
50
+ $ python sample.py --out_dir=out-shakespeare-char
51
+ ```
52
+
53
+ This generates a few samples, for example:
54
+
55
+ ```
56
+ ANGELO:
57
+ And cowards it be strawn to my bed,
58
+ And thrust the gates of my threats,
59
+ Because he that ale away, and hang'd
60
+ An one with him.
61
+
62
+ DUKE VINCENTIO:
63
+ I thank your eyes against it.
64
+
65
+ DUKE VINCENTIO:
66
+ Then will answer him to save the malm:
67
+ And what have you tyrannous shall do this?
68
+
69
+ DUKE VINCENTIO:
70
+ If you have done evils of all disposition
71
+ To end his power, the day of thrust for a common men
72
+ That I leave, to fight with over-liking
73
+ Hasting in a roseman.
74
+ ```
75
+
76
+ lol `¯\_(ツ)_/¯`. Not bad for a character-level model after 3 minutes of training on a GPU. Better results are quite likely obtainable by instead finetuning a pretrained GPT-2 model on this dataset (see finetuning section later).
77
+
78
+ **I only have a macbook** (or other cheap computer). No worries, we can still train a GPT but we want to dial things down a notch. I recommend getting the bleeding edge PyTorch nightly ([select it here](https://pytorch.org/get-started/locally/) when installing) as it is currently quite likely to make your code more efficient. But even without it, a simple train run could look as follows:
79
+
80
+ ```
81
+ $ python train.py config/train_shakespeare_char.py --device=cpu --compile=False --eval_iters=20 --log_interval=1 --block_size=64 --batch_size=12 --n_layer=4 --n_head=4 --n_embd=128 --max_iters=2000 --lr_decay_iters=2000 --dropout=0.0
82
+ ```
83
+
84
+ Here, since we are running on CPU instead of GPU we must set both `--device=cpu` and also turn off PyTorch 2.0 compile with `--compile=False`. Then when we evaluate we get a bit more noisy but faster estimate (`--eval_iters=20`, down from 200), our context size is only 64 characters instead of 256, and the batch size only 12 examples per iteration, not 64. We'll also use a much smaller Transformer (4 layers, 4 heads, 128 embedding size), and decrease the number of iterations to 2000 (and correspondingly usually decay the learning rate to around max_iters with `--lr_decay_iters`). Because our network is so small we also ease down on regularization (`--dropout=0.0`). This still runs in about ~3 minutes, but gets us a loss of only 1.88 and therefore also worse samples, but it's still good fun:
85
+
86
+ ```
87
+ $ python sample.py --out_dir=out-shakespeare-char --device=cpu
88
+ ```
89
+ Generates samples like this:
90
+
91
+ ```
92
+ GLEORKEN VINGHARD III:
93
+ Whell's the couse, the came light gacks,
94
+ And the for mought you in Aut fries the not high shee
95
+ bot thou the sought bechive in that to doth groan you,
96
+ No relving thee post mose the wear
97
+ ```
98
+
99
+ Not bad for ~3 minutes on a CPU, for a hint of the right character gestalt. If you're willing to wait longer, feel free to tune the hyperparameters, increase the size of the network, the context length (`--block_size`), the length of training, etc.
100
+
101
+ Finally, on Apple Silicon Macbooks and with a recent PyTorch version make sure to add `--device=mps` (short for "Metal Performance Shaders"); PyTorch then uses the on-chip GPU that can *significantly* accelerate training (2-3X) and allow you to use larger networks. See [Issue 28](https://github.com/karpathy/nanoGPT/issues/28) for more.
102
+
103
+ ## reproducing GPT-2
104
+
105
+ A more serious deep learning professional may be more interested in reproducing GPT-2 results. So here we go - we first tokenize the dataset, in this case the [OpenWebText](https://openwebtext2.readthedocs.io/en/latest/), an open reproduction of OpenAI's (private) WebText:
106
+
107
+ ```
108
+ $ python data/openwebtext/prepare.py
109
+ ```
110
+
111
+ This downloads and tokenizes the [OpenWebText](https://huggingface.co/datasets/openwebtext) dataset. It will create a `train.bin` and `val.bin` which holds the GPT2 BPE token ids in one sequence, stored as raw uint16 bytes. Then we're ready to kick off training. To reproduce GPT-2 (124M) you'll want at least an 8X A100 40GB node and run:
112
+
113
+ ```
114
+ $ torchrun --standalone --nproc_per_node=8 train.py config/train_gpt2.py
115
+ ```
116
+
117
+ This will run for about 4 days using PyTorch Distributed Data Parallel (DDP) and go down to loss of ~2.85. Now, a GPT-2 model just evaluated on OWT gets a val loss of about 3.11, but if you finetune it it will come down to ~2.85 territory (due to an apparent domain gap), making the two models ~match.
118
+
119
+ If you're in a cluster environment and you are blessed with multiple GPU nodes you can make GPU go brrrr e.g. across 2 nodes like:
120
+
121
+ ```
122
+ Run on the first (master) node with example IP 123.456.123.456:
123
+ $ torchrun --nproc_per_node=8 --nnodes=2 --node_rank=0 --master_addr=123.456.123.456 --master_port=1234 train.py
124
+ Run on the worker node:
125
+ $ torchrun --nproc_per_node=8 --nnodes=2 --node_rank=1 --master_addr=123.456.123.456 --master_port=1234 train.py
126
+ ```
127
+
128
+ It is a good idea to benchmark your interconnect (e.g. iperf3). In particular, if you don't have Infiniband then also prepend `NCCL_IB_DISABLE=1` to the above launches. Your multinode training will work, but most likely _crawl_. By default checkpoints are periodically written to the `--out_dir`. We can sample from the model by simply `$ python sample.py`.
129
+
130
+ Finally, to train on a single GPU simply run the `$ python train.py` script. Have a look at all of its args, the script tries to be very readable, hackable and transparent. You'll most likely want to tune a number of those variables depending on your needs.
131
+
132
+ ## baselines
133
+
134
+ OpenAI GPT-2 checkpoints allow us to get some baselines in place for openwebtext. We can get the numbers as follows:
135
+
136
+ ```
137
+ $ python train.py eval_gpt2
138
+ $ python train.py eval_gpt2_medium
139
+ $ python train.py eval_gpt2_large
140
+ $ python train.py eval_gpt2_xl
141
+ ```
142
+
143
+ and observe the following losses on train and val:
144
+
145
+ | model | params | train loss | val loss |
146
+ | ------| ------ | ---------- | -------- |
147
+ | gpt2 | 124M | 3.11 | 3.12 |
148
+ | gpt2-medium | 350M | 2.85 | 2.84 |
149
+ | gpt2-large | 774M | 2.66 | 2.67 |
150
+ | gpt2-xl | 1558M | 2.56 | 2.54 |
151
+
152
+ However, we have to note that GPT-2 was trained on (closed, never released) WebText, while OpenWebText is just a best-effort open reproduction of this dataset. This means there is a dataset domain gap. Indeed, taking the GPT-2 (124M) checkpoint and finetuning on OWT directly for a while reaches loss down to ~2.85. This then becomes the more appropriate baseline w.r.t. reproduction.
153
+
154
+ ## finetuning
155
+
156
+ Finetuning is no different than training, we just make sure to initialize from a pretrained model and train with a smaller learning rate. For an example of how to finetune a GPT on new text go to `data/shakespeare` and run `prepare.py` to download the tiny shakespeare dataset and render it into a `train.bin` and `val.bin`, using the OpenAI BPE tokenizer from GPT-2. Unlike OpenWebText this will run in seconds. Finetuning can take very little time, e.g. on a single GPU just a few minutes. Run an example finetuning like:
157
+
158
+ ```
159
+ $ python train.py config/finetune_shakespeare.py
160
+ ```
161
+
162
+ This will load the config parameter overrides in `config/finetune_shakespeare.py` (I didn't tune them much though). Basically, we initialize from a GPT2 checkpoint with `init_from` and train as normal, except shorter and with a small learning rate. If you're running out of memory try decreasing the model size (they are `{'gpt2', 'gpt2-medium', 'gpt2-large', 'gpt2-xl'}`) or possibly decreasing the `block_size` (context length). The best checkpoint (lowest validation loss) will be in the `out_dir` directory, e.g. in `out-shakespeare` by default, per the config file. You can then run the code in `sample.py --out_dir=out-shakespeare`:
163
+
164
+ ```
165
+ THEODORE:
166
+ Thou shalt sell me to the highest bidder: if I die,
167
+ I sell thee to the first; if I go mad,
168
+ I sell thee to the second; if I
169
+ lie, I sell thee to the third; if I slay,
170
+ I sell thee to the fourth: so buy or sell,
171
+ I tell thee again, thou shalt not sell my
172
+ possession.
173
+
174
+ JULIET:
175
+ And if thou steal, thou shalt not sell thyself.
176
+
177
+ THEODORE:
178
+ I do not steal; I sell the stolen goods.
179
+
180
+ THEODORE:
181
+ Thou know'st not what thou sell'st; thou, a woman,
182
+ Thou art ever a victim, a thing of no worth:
183
+ Thou hast no right, no right, but to be sold.
184
+ ```
185
+
186
+ Whoa there, GPT, entering some dark place over there. I didn't really tune the hyperparameters in the config too much, feel free to try!
187
+
188
+ ## sampling / inference
189
+
190
+ Use the script `sample.py` to sample either from pre-trained GPT-2 models released by OpenAI, or from a model you trained yourself. For example, here is a way to sample from the largest available `gpt2-xl` model:
191
+
192
+ ```
193
+ $ python sample.py \
194
+ --init_from=gpt2-xl \
195
+ --start="What is the answer to life, the universe, and everything?" \
196
+ --num_samples=5 --max_new_tokens=100
197
+ ```
198
+
199
+ If you'd like to sample from a model you trained, use the `--out_dir` to point the code appropriately. You can also prompt the model with some text from a file, e.g. `$ python sample.py --start=FILE:prompt.txt`.
200
+
201
+ ## efficiency notes
202
+
203
+ For simple model benchmarking and profiling, `bench.py` might be useful. It's identical to what happens in the meat of the training loop of `train.py`, but omits much of the other complexities.
204
+
205
+ Note that the code by default uses [PyTorch 2.0](https://pytorch.org/get-started/pytorch-2.0/). At the time of writing (Dec 29, 2022) this makes `torch.compile()` available in the nightly release. The improvement from the one line of code is noticeable, e.g. cutting down iteration time from ~250ms / iter to 135ms / iter. Nice work PyTorch team!
206
+
207
+ ## todos
208
+
209
+ - Investigate and add FSDP instead of DDP
210
+ - Eval zero-shot perplexities on standard evals (e.g. LAMBADA? HELM? etc.)
211
+ - Finetune the finetuning script, I think the hyperparams are not great
212
+ - Schedule for linear batch size increase during training
213
+ - Incorporate other embeddings (rotary, alibi)
214
+ - Separate out the optim buffers from model params in checkpoints I think
215
+ - Additional logging around network health (e.g. gradient clip events, magnitudes)
216
+ - Few more investigations around better init etc.
217
+
218
+ ## troubleshooting
219
+
220
+ Note that by default this repo uses PyTorch 2.0 (i.e. `torch.compile`). This is fairly new and experimental, and not yet available on all platforms (e.g. Windows). If you're running into related error messages try to disable this by adding `--compile=False` flag. This will slow down the code but at least it will run.
221
+
222
+ For some context on this repository, GPT, and language modeling it might be helpful to watch my [Zero To Hero series](https://karpathy.ai/zero-to-hero.html). Specifically, the [GPT video](https://www.youtube.com/watch?v=kCc8FmEb1nY) is popular if you have some prior language modeling context.
223
+
224
+ For more questions/discussions feel free to stop by **#nanoGPT** on Discord:
225
+
226
+ [![](https://dcbadge.vercel.app/api/server/3zy8kqD9Cp?compact=true&style=flat)](https://discord.gg/3zy8kqD9Cp)
227
+
228
+ ## acknowledgements
229
+
230
+ All nanoGPT experiments are powered by GPUs on [Lambda labs](https://lambdalabs.com), my favorite Cloud GPU provider. Thank you Lambda labs for sponsoring nanoGPT!
assets/gpt2_124M_loss.png ADDED
assets/nanogpt.jpg ADDED
bench.py ADDED
@@ -0,0 +1,117 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """
2
+ A much shorter version of train.py for benchmarking
3
+ """
4
+ import os
5
+ from contextlib import nullcontext
6
+ import numpy as np
7
+ import time
8
+ import torch
9
+ from model import GPTConfig, GPT
10
+
11
+ # -----------------------------------------------------------------------------
12
+ batch_size = 12
13
+ block_size = 1024
14
+ bias = False
15
+ real_data = True
16
+ seed = 1337
17
+ device = 'cuda' # examples: 'cpu', 'cuda', 'cuda:0', 'cuda:1', etc.
18
+ dtype = 'bfloat16' if torch.cuda.is_available() and torch.cuda.is_bf16_supported() else 'float16' # 'float32' or 'bfloat16' or 'float16'
19
+ compile = True # use PyTorch 2.0 to compile the model to be faster
20
+ profile = False # use pytorch profiler, or just simple benchmarking?
21
+ exec(open('configurator.py').read()) # overrides from command line or config file
22
+ # -----------------------------------------------------------------------------
23
+
24
+ torch.manual_seed(seed)
25
+ torch.cuda.manual_seed(seed)
26
+ torch.backends.cuda.matmul.allow_tf32 = True # allow tf32 on matmul
27
+ torch.backends.cudnn.allow_tf32 = True # allow tf32 on cudnn
28
+ device_type = 'cuda' if 'cuda' in device else 'cpu' # for later use in torch.autocast
29
+ ptdtype = {'float32': torch.float32, 'bfloat16': torch.bfloat16, 'float16': torch.float16}[dtype]
30
+ ctx = nullcontext() if device_type == 'cpu' else torch.amp.autocast(device_type=device_type, dtype=ptdtype)
31
+
32
+ # data loading init
33
+ if real_data:
34
+ dataset = 'openwebtext'
35
+ data_dir = os.path.join('data', dataset)
36
+ train_data = np.memmap(os.path.join(data_dir, 'train.bin'), dtype=np.uint16, mode='r')
37
+ def get_batch(split):
38
+ data = train_data # note ignore split in benchmarking script
39
+ ix = torch.randint(len(data) - block_size, (batch_size,))
40
+ x = torch.stack([torch.from_numpy((data[i:i+block_size]).astype(np.int64)) for i in ix])
41
+ y = torch.stack([torch.from_numpy((data[i+1:i+1+block_size]).astype(np.int64)) for i in ix])
42
+ x, y = x.pin_memory().to(device, non_blocking=True), y.pin_memory().to(device, non_blocking=True)
43
+ return x, y
44
+ else:
45
+ # alternatively, if fixed data is desired to not care about data loading
46
+ x = torch.randint(50304, (batch_size, block_size), device=device)
47
+ y = torch.randint(50304, (batch_size, block_size), device=device)
48
+ get_batch = lambda split: (x, y)
49
+
50
+ # model init
51
+ gptconf = GPTConfig(
52
+ block_size = block_size, # how far back does the model look? i.e. context size
53
+ n_layer = 12, n_head = 12, n_embd = 768, # size of the model
54
+ dropout = 0, # for determinism
55
+ bias = bias,
56
+ )
57
+ model = GPT(gptconf)
58
+ model.to(device)
59
+
60
+ optimizer = model.configure_optimizers(weight_decay=1e-2, learning_rate=1e-4, betas=(0.9, 0.95), device_type=device_type)
61
+
62
+ if compile:
63
+ print("Compiling model...")
64
+ model = torch.compile(model) # pytorch 2.0
65
+
66
+ if profile:
67
+ # useful docs on pytorch profiler:
68
+ # - tutorial https://pytorch.org/tutorials/intermediate/tensorboard_profiler_tutorial.html
69
+ # - api https://pytorch.org/docs/stable/profiler.html#torch.profiler.profile
70
+ wait, warmup, active = 5, 5, 5
71
+ num_steps = wait + warmup + active
72
+ with torch.profiler.profile(
73
+ activities=[torch.profiler.ProfilerActivity.CPU, torch.profiler.ProfilerActivity.CUDA],
74
+ schedule=torch.profiler.schedule(wait=wait, warmup=warmup, active=active, repeat=1),
75
+ on_trace_ready=torch.profiler.tensorboard_trace_handler('./bench_log'),
76
+ record_shapes=False,
77
+ profile_memory=False,
78
+ with_stack=False, # incurs an additional overhead, disable if not needed
79
+ with_flops=True,
80
+ with_modules=False, # only for torchscript models atm
81
+ ) as prof:
82
+
83
+ X, Y = get_batch('train')
84
+ for k in range(num_steps):
85
+ with ctx:
86
+ logits, loss = model(X, Y)
87
+ X, Y = get_batch('train')
88
+ optimizer.zero_grad(set_to_none=True)
89
+ loss.backward()
90
+ optimizer.step()
91
+ lossf = loss.item()
92
+ print(f"{k}/{num_steps} loss: {lossf:.4f}")
93
+
94
+ prof.step() # notify the profiler at end of each step
95
+
96
+ else:
97
+
98
+ # simple benchmarking
99
+ torch.cuda.synchronize()
100
+ for stage, num_steps in enumerate([10, 20]): # burnin, then benchmark
101
+ t0 = time.time()
102
+ X, Y = get_batch('train')
103
+ for k in range(num_steps):
104
+ with ctx:
105
+ logits, loss = model(X, Y)
106
+ X, Y = get_batch('train')
107
+ optimizer.zero_grad(set_to_none=True)
108
+ loss.backward()
109
+ optimizer.step()
110
+ lossf = loss.item()
111
+ print(f"{k}/{num_steps} loss: {lossf:.4f}")
112
+ torch.cuda.synchronize()
113
+ t1 = time.time()
114
+ dt = t1-t0
115
+ mfu = model.estimate_mfu(batch_size * 1 * num_steps, dt)
116
+ if stage == 1:
117
+ print(f"time per iteration: {dt/num_steps*1000:.4f}ms, MFU: {mfu*100:.2f}%")
config/eval_gpt2.py ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ # evaluate the base gpt2
2
+ # n_layer=12, n_head=12, n_embd=768
3
+ # 124M parameters
4
+ batch_size = 8
5
+ eval_iters = 500 # use more iterations to get good estimate
6
+ eval_only = True
7
+ wandb_log = False
8
+ init_from = 'gpt2'
config/eval_gpt2_large.py ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ # evaluate the base gpt2
2
+ # n_layer=36, n_head=20, n_embd=1280
3
+ # 774M parameters
4
+ batch_size = 8
5
+ eval_iters = 500 # use more iterations to get good estimate
6
+ eval_only = True
7
+ wandb_log = False
8
+ init_from = 'gpt2-large'
config/eval_gpt2_medium.py ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ # evaluate the base gpt2
2
+ # n_layer=24, n_head=16, n_embd=1024
3
+ # 350M parameters
4
+ batch_size = 8
5
+ eval_iters = 500 # use more iterations to get good estimate
6
+ eval_only = True
7
+ wandb_log = False
8
+ init_from = 'gpt2-medium'
config/eval_gpt2_xl.py ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ # evaluate the base gpt2
2
+ # n_layer=48, n_head=25, n_embd=1600
3
+ # 1558M parameters
4
+ batch_size = 8
5
+ eval_iters = 500 # use more iterations to get good estimate
6
+ eval_only = True
7
+ wandb_log = False
8
+ init_from = 'gpt2-xl'
config/finetune_shakespeare.py ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import time
2
+
3
+ out_dir = 'out-shakespeare'
4
+ eval_interval = 5
5
+ eval_iters = 40
6
+ wandb_log = False # feel free to turn on
7
+ wandb_project = 'shakespeare'
8
+ wandb_run_name = 'ft-' + str(time.time())
9
+
10
+ dataset = 'shakespeare'
11
+ init_from = 'gpt2-xl' # this is the largest GPT-2 model
12
+
13
+ # only save checkpoints if the validation loss improves
14
+ always_save_checkpoint = False
15
+
16
+ # the number of examples per iter:
17
+ # 1 batch_size * 32 grad_accum * 1024 tokens = 32,768 tokens/iter
18
+ # shakespeare has 301,966 tokens, so 1 epoch ~= 9.2 iters
19
+ batch_size = 1
20
+ gradient_accumulation_steps = 32
21
+ max_iters = 20
22
+
23
+ # finetune at constant LR
24
+ learning_rate = 3e-5
25
+ decay_lr = False
config/train_gpt2.py ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # config for training GPT-2 (124M) down to very nice loss of ~2.85 on 1 node of 8X A100 40GB
2
+ # launch as the following (e.g. in a screen session) and wait ~5 days:
3
+ # $ torchrun --standalone --nproc_per_node=8 train.py config/train_gpt2.py
4
+
5
+ wandb_log = True
6
+ wandb_project = 'owt'
7
+ wandb_run_name='gpt2-124M'
8
+
9
+ # these make the total batch size be ~0.5M
10
+ # 12 batch size * 1024 block size * 5 gradaccum * 8 GPUs = 491,520
11
+ batch_size = 12
12
+ block_size = 1024
13
+ gradient_accumulation_steps = 5 * 8
14
+
15
+ # this makes total number of tokens be 300B
16
+ max_iters = 600000
17
+ lr_decay_iters = 600000
18
+
19
+ # eval stuff
20
+ eval_interval = 1000
21
+ eval_iters = 200
22
+ log_interval = 10
23
+
24
+ # weight decay
25
+ weight_decay = 1e-1
config/train_shakespeare_char.py ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # train a miniature character-level shakespeare model
2
+ # good for debugging and playing on macbooks and such
3
+
4
+ out_dir = 'out-shakespeare-char'
5
+ eval_interval = 250 # keep frequent because we'll overfit
6
+ eval_iters = 200
7
+ log_interval = 10 # don't print too too often
8
+
9
+ # we expect to overfit on this small dataset, so only save when val improves
10
+ always_save_checkpoint = False
11
+
12
+ wandb_log = False # override via command line if you like
13
+ wandb_project = 'shakespeare-char'
14
+ wandb_run_name = 'mini-gpt'
15
+
16
+ dataset = 'shakespeare_char'
17
+ gradient_accumulation_steps = 1
18
+ batch_size = 64
19
+ block_size = 256 # context of up to 256 previous characters
20
+
21
+ # baby GPT model :)
22
+ n_layer = 6
23
+ n_head = 6
24
+ n_embd = 384
25
+ dropout = 0.2
26
+
27
+ learning_rate = 1e-3 # with baby networks can afford to go a bit higher
28
+ max_iters = 5000
29
+ lr_decay_iters = 5000 # make equal to max_iters usually
30
+ min_lr = 1e-4 # learning_rate / 10 usually
31
+ beta2 = 0.99 # make a bit bigger because number of tokens per iter is small
32
+
33
+ warmup_iters = 100 # not super necessary potentially
34
+
35
+ # on macbook also add
36
+ # device = 'cpu' # run on cpu only
37
+ # compile = False # do not torch compile the model
configurator.py ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """
2
+ Poor Man's Configurator. Probably a terrible idea. Example usage:
3
+ $ python train.py config/override_file.py --batch_size=32
4
+ this will first run config/override_file.py, then override batch_size to 32
5
+
6
+ The code in this file will be run as follows from e.g. train.py:
7
+ >>> exec(open('configurator.py').read())
8
+
9
+ So it's not a Python module, it's just shuttling this code away from train.py
10
+ The code in this script then overrides the globals()
11
+
12
+ I know people are not going to love this, I just really dislike configuration
13
+ complexity and having to prepend config. to every single variable. If someone
14
+ comes up with a better simple Python solution I am all ears.
15
+ """
16
+
17
+ import sys
18
+ from ast import literal_eval
19
+
20
+ for arg in sys.argv[1:]:
21
+ if '=' not in arg:
22
+ # assume it's the name of a config file
23
+ assert not arg.startswith('--')
24
+ config_file = arg
25
+ print(f"Overriding config with {config_file}:")
26
+ with open(config_file) as f:
27
+ print(f.read())
28
+ exec(open(config_file).read())
29
+ else:
30
+ # assume it's a --key=value argument
31
+ assert arg.startswith('--')
32
+ key, val = arg.split('=')
33
+ key = key[2:]
34
+ if key in globals():
35
+ try:
36
+ # attempt to eval it it (e.g. if bool, number, or etc)
37
+ attempt = literal_eval(val)
38
+ except (SyntaxError, ValueError):
39
+ # if that goes wrong, just use the string
40
+ attempt = val
41
+ # ensure the types match ok
42
+ assert type(attempt) == type(globals()[key])
43
+ # cross fingers
44
+ print(f"Overriding: {key} = {attempt}")
45
+ globals()[key] = attempt
46
+ else:
47
+ raise ValueError(f"Unknown config key: {key}")
data/openwebtext/prepare.py ADDED
@@ -0,0 +1,81 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # saves the openwebtext dataset to a binary file for training. following was helpful:
2
+ # https://github.com/HazyResearch/flash-attention/blob/main/training/src/datamodules/language_modeling_hf.py
3
+
4
+ import os
5
+ from tqdm import tqdm
6
+ import numpy as np
7
+ import tiktoken
8
+ from datasets import load_dataset # huggingface datasets
9
+
10
+ # number of workers in .map() call
11
+ # good number to use is ~order number of cpu cores // 2
12
+ num_proc = 8
13
+
14
+ # number of workers in load_dataset() call
15
+ # best number might be different from num_proc above as it also depends on NW speed.
16
+ # it is better than 1 usually though
17
+ num_proc_load_dataset = num_proc
18
+
19
+ enc = tiktoken.get_encoding("gpt2")
20
+
21
+ if __name__ == '__main__':
22
+ # takes 54GB in huggingface .cache dir, about 8M documents (8,013,769)
23
+ dataset = load_dataset("openwebtext", num_proc=num_proc_load_dataset)
24
+
25
+ # owt by default only contains the 'train' split, so create a test split
26
+ split_dataset = dataset["train"].train_test_split(test_size=0.0005, seed=2357, shuffle=True)
27
+ split_dataset['val'] = split_dataset.pop('test') # rename the test split to val
28
+
29
+ # this results in:
30
+ # >>> split_dataset
31
+ # DatasetDict({
32
+ # train: Dataset({
33
+ # features: ['text'],
34
+ # num_rows: 8009762
35
+ # })
36
+ # val: Dataset({
37
+ # features: ['text'],
38
+ # num_rows: 4007
39
+ # })
40
+ # })
41
+
42
+ # we now want to tokenize the dataset. first define the encoding function (gpt2 bpe)
43
+ def process(example):
44
+ ids = enc.encode_ordinary(example['text']) # encode_ordinary ignores any special tokens
45
+ ids.append(enc.eot_token) # add the end of text token, e.g. 50256 for gpt2 bpe
46
+ # note: I think eot should be prepended not appended... hmm. it's called "eot" though...
47
+ out = {'ids': ids, 'len': len(ids)}
48
+ return out
49
+
50
+ # tokenize the dataset
51
+ tokenized = split_dataset.map(
52
+ process,
53
+ remove_columns=['text'],
54
+ desc="tokenizing the splits",
55
+ num_proc=num_proc,
56
+ )
57
+
58
+ # concatenate all the ids in each dataset into one large file we can use for training
59
+ for split, dset in tokenized.items():
60
+ arr_len = np.sum(dset['len'], dtype=np.uint64)
61
+ filename = os.path.join(os.path.dirname(__file__), f'{split}.bin')
62
+ dtype = np.uint16 # (can do since enc.max_token_value == 50256 is < 2**16)
63
+ arr = np.memmap(filename, dtype=dtype, mode='w+', shape=(arr_len,))
64
+ total_batches = 1024
65
+
66
+ idx = 0
67
+ for batch_idx in tqdm(range(total_batches), desc=f'writing {filename}'):
68
+ # Batch together samples for faster write
69
+ batch = dset.shard(num_shards=total_batches, index=batch_idx, contiguous=True).with_format('numpy')
70
+ arr_batch = np.concatenate(batch['ids'])
71
+ # Write into mmap
72
+ arr[idx : idx + len(arr_batch)] = arr_batch
73
+ idx += len(arr_batch)
74
+ arr.flush()
75
+
76
+ # train.bin is ~17GB, val.bin ~8.5MB
77
+ # train has ~9B tokens (9,035,582,198)
78
+ # val has ~4M tokens (4,434,897)
79
+
80
+ # to read the bin files later, e.g. with numpy:
81
+ # m = np.memmap('train.bin', dtype=np.uint16, mode='r')
data/openwebtext/readme.md ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ## openwebtext dataset
3
+
4
+ after running `prepare.py` (preprocess) we get:
5
+
6
+ - train.bin is ~17GB, val.bin ~8.5MB
7
+ - train has ~9B tokens (9,035,582,198)
8
+ - val has ~4M tokens (4,434,897)
9
+
10
+ this came from 8,013,769 documents in total.
11
+
12
+ references:
13
+
14
+ - OpenAI's WebText dataset is discussed in [GPT-2 paper](https://d4mucfpksywv.cloudfront.net/better-language-models/language_models_are_unsupervised_multitask_learners.pdf)
15
+ - [OpenWebText](https://skylion007.github.io/OpenWebTextCorpus/) dataset
data/shakespeare/prepare.py ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import os
2
+ import requests
3
+ import tiktoken
4
+ import numpy as np
5
+
6
+ # download the tiny shakespeare dataset
7
+ input_file_path = os.path.join(os.path.dirname(__file__), 'input.txt')
8
+ if not os.path.exists(input_file_path):
9
+ data_url = 'https://raw.githubusercontent.com/karpathy/char-rnn/master/data/tinyshakespeare/input.txt'
10
+ with open(input_file_path, 'w', encoding='utf-8') as f:
11
+ f.write(requests.get(data_url).text)
12
+
13
+ with open(input_file_path, 'r', encoding='utf-8') as f:
14
+ data = f.read()
15
+ n = len(data)
16
+ train_data = data[:int(n*0.9)]
17
+ val_data = data[int(n*0.9):]
18
+
19
+ # encode with tiktoken gpt2 bpe
20
+ enc = tiktoken.get_encoding("gpt2")
21
+ train_ids = enc.encode_ordinary(train_data)
22
+ val_ids = enc.encode_ordinary(val_data)
23
+ print(f"train has {len(train_ids):,} tokens")
24
+ print(f"val has {len(val_ids):,} tokens")
25
+
26
+ # export to bin files
27
+ train_ids = np.array(train_ids, dtype=np.uint16)
28
+ val_ids = np.array(val_ids, dtype=np.uint16)
29
+ train_ids.tofile(os.path.join(os.path.dirname(__file__), 'train.bin'))
30
+ val_ids.tofile(os.path.join(os.path.dirname(__file__), 'val.bin'))
31
+
32
+ # train.bin has 301,966 tokens
33
+ # val.bin has 36,059 tokens
data/shakespeare/readme.md ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+
2
+ # tiny shakespeare
3
+
4
+ Tiny shakespeare, of the good old char-rnn fame :)
5
+
6
+ After running `prepare.py`:
7
+
8
+ - train.bin has 301,966 tokens
9
+ - val.bin has 36,059 tokens
data/shakespeare_char/input.txt ADDED
The diff for this file is too large to render. See raw diff
 
data/shakespeare_char/prepare.py ADDED
@@ -0,0 +1,68 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """
2
+ Prepare the Shakespeare dataset for character-level language modeling.
3
+ So instead of encoding with GPT-2 BPE tokens, we just map characters to ints.
4
+ Will save train.bin, val.bin containing the ids, and meta.pkl containing the
5
+ encoder and decoder and some other related info.
6
+ """
7
+ import os
8
+ import pickle
9
+ import requests
10
+ import numpy as np
11
+
12
+ # download the tiny shakespeare dataset
13
+ input_file_path = os.path.join(os.path.dirname(__file__), 'input.txt')
14
+ if not os.path.exists(input_file_path):
15
+ data_url = 'https://raw.githubusercontent.com/karpathy/char-rnn/master/data/tinyshakespeare/input.txt'
16
+ with open(input_file_path, 'w') as f:
17
+ f.write(requests.get(data_url).text)
18
+
19
+ with open(input_file_path, 'r') as f:
20
+ data = f.read()
21
+ print(f"length of dataset in characters: {len(data):,}")
22
+
23
+ # get all the unique characters that occur in this text
24
+ chars = sorted(list(set(data)))
25
+ vocab_size = len(chars)
26
+ print("all the unique characters:", ''.join(chars))
27
+ print(f"vocab size: {vocab_size:,}")
28
+
29
+ # create a mapping from characters to integers
30
+ stoi = { ch:i for i,ch in enumerate(chars) }
31
+ itos = { i:ch for i,ch in enumerate(chars) }
32
+ def encode(s):
33
+ return [stoi[c] for c in s] # encoder: take a string, output a list of integers
34
+ def decode(l):
35
+ return ''.join([itos[i] for i in l]) # decoder: take a list of integers, output a string
36
+
37
+ # create the train and test splits
38
+ n = len(data)
39
+ train_data = data[:int(n*0.9)]
40
+ val_data = data[int(n*0.9):]
41
+
42
+ # encode both to integers
43
+ train_ids = encode(train_data)
44
+ val_ids = encode(val_data)
45
+ print(f"train has {len(train_ids):,} tokens")
46
+ print(f"val has {len(val_ids):,} tokens")
47
+
48
+ # export to bin files
49
+ train_ids = np.array(train_ids, dtype=np.uint16)
50
+ val_ids = np.array(val_ids, dtype=np.uint16)
51
+ train_ids.tofile(os.path.join(os.path.dirname(__file__), 'train.bin'))
52
+ val_ids.tofile(os.path.join(os.path.dirname(__file__), 'val.bin'))
53
+
54
+ # save the meta information as well, to help us encode/decode later
55
+ meta = {
56
+ 'vocab_size': vocab_size,
57
+ 'itos': itos,
58
+ 'stoi': stoi,
59
+ }
60
+ with open(os.path.join(os.path.dirname(__file__), 'meta.pkl'), 'wb') as f:
61
+ pickle.dump(meta, f)
62
+
63
+ # length of dataset in characters: 1115394
64
+ # all the unique characters:
65
+ # !$&',-.3:;?ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz
66
+ # vocab size: 65
67
+ # train has 1003854 tokens
68
+ # val has 111540 tokens
data/shakespeare_char/readme.md ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+
2
+ # tiny shakespeare, character-level
3
+
4
+ Tiny shakespeare, of the good old char-rnn fame :) Treated on character-level.
5
+
6
+ After running `prepare.py`:
7
+
8
+ - train.bin has 1,003,854 tokens
9
+ - val.bin has 111,540 tokens
model.py ADDED
@@ -0,0 +1,409 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """
2
+ Full definition of a GPT Language Model, all of it in this single file.
3
+ References:
4
+ 1) the official GPT-2 TensorFlow implementation released by OpenAI:
5
+ https://github.com/openai/gpt-2/blob/master/src/model.py
6
+ 2) huggingface/transformers PyTorch implementation:
7
+ https://github.com/huggingface/transformers/blob/main/src/transformers/models/gpt2/modeling_gpt2.py
8
+ """
9
+
10
+ import math
11
+ import inspect
12
+ from dataclasses import dataclass
13
+
14
+ import torch
15
+ import torch.nn as nn
16
+ from torch.nn import functional as F
17
+
18
+ # https://github.com/suito555/bitnet158b
19
+ class BitLinear158b(nn.Linear):
20
+ def __init__(self, in_features, out_features, bias=False, bit_scale = 8):
21
+ super(BitLinear158b, self).__init__(in_features, out_features, bias)
22
+ self.bit_scale = bit_scale
23
+ self.Q_b = 2 ** (self.bit_scale - 1)
24
+ self.eps = 1e-8
25
+
26
+ def quantize_activations(self, input_norm, abs_max_x_value):
27
+ scaled_x = input_norm * self.Q_b / (abs_max_x_value + self.eps)
28
+ quantized_x = torch.round(torch.clamp(
29
+ scaled_x, -self.Q_b + self.eps, self.Q_b - self.eps
30
+ ))
31
+ # Inserting following comment outed line prevents the loss to be NaN if you use torch.compile with torch==2.0.0 or torch==2.2.1
32
+ # If you use torch==2.1.2, it will not be compiled.
33
+ if(torch.isnan(scaled_x).any()):
34
+ print("Nan!")
35
+ #STE
36
+ quantized_x = (quantized_x - scaled_x).detach() + scaled_x
37
+ return quantized_x
38
+
39
+ def ternarize_weights(self,abs_mean_W_value):
40
+ scaled_W = self.weight / (abs_mean_W_value + self.eps)
41
+ quantize_weights = torch.clamp(scaled_W.round(), -1, 1)
42
+ #STE
43
+ quantize_weights = (quantize_weights - self.weight).detach() + self.weight
44
+ return quantize_weights
45
+
46
+ def forward(self, input):
47
+ input_norm = F.layer_norm(input, (self.in_features,))
48
+
49
+ abs_max_x_value = input_norm.abs().max() #gamma
50
+ quant_scaled_input = self.quantize_activations(input_norm,abs_max_x_value)
51
+
52
+ abs_mean_W_value = self.weight.abs().mean() #beta
53
+ ternarized_weights = self.ternarize_weights(abs_mean_W_value)
54
+
55
+ matmal_weight = F.linear(quant_scaled_input, ternarized_weights, self.bias)
56
+
57
+ beta_gamma = abs_mean_W_value * abs_max_x_value
58
+ output = matmal_weight * beta_gamma / self.Q_b
59
+ return output
60
+
61
+ # class BitLinear158bNIQ(nn.Linear):
62
+ # def __init__(self, in_features, out_features, bias=False, bit_scale = 8):
63
+ # super(BitLinear158bNIQ, self).__init__(in_features, out_features, bias)
64
+ # self.bit_scale = bit_scale
65
+ # self.Q_b = 2 ** (self.bit_scale - 1)
66
+ # self.eps = 1e-8
67
+
68
+ # def quantize_activations(self, input_norm, abs_max_x_value):
69
+ # scaled_x = torch.clamp(
70
+ # input_norm * self.Q_b / (abs_max_x_value + self.eps), -self.Q_b + self.eps, self.Q_b - self.eps
71
+ # )
72
+ # return scaled_x
73
+
74
+ # def ternarize_weights(self,abs_mean_W_value):
75
+ # scaled_W = self.weight / (abs_mean_W_value + self.eps)
76
+ # quantize_weights = torch.sign(torch.clamp(scaled_W.round(), -1, 1))
77
+ # #STE
78
+ # quantize_weights = (quantize_weights - self.weight).detach() + self.weight
79
+ # return quantize_weights
80
+
81
+ # def forward(self, input):
82
+ # input_norm = F.layer_norm(input, (self.in_features,))
83
+
84
+ # abs_mean_W_value = self.weight.abs().mean() #beta
85
+ # ternarized_weights = self.ternarize_weights(abs_mean_W_value)
86
+
87
+ # matmal_weight = F.linear(input_norm, ternarized_weights, self.bias)
88
+
89
+ # beta_gamma = abs_mean_W_value
90
+ # output = matmal_weight * beta_gamma / self.Q_b
91
+ # return output
92
+
93
+ class LayerNorm(nn.Module):
94
+ """ LayerNorm but with an optional bias. PyTorch doesn't support simply bias=False """
95
+
96
+ def __init__(self, ndim, bias):
97
+ super().__init__()
98
+ self.weight = nn.Parameter(torch.ones(ndim))
99
+ self.bias = nn.Parameter(torch.zeros(ndim)) if bias else None
100
+
101
+ def forward(self, input):
102
+ return F.layer_norm(input, self.weight.shape, self.weight, self.bias, 1e-5)
103
+
104
+ class CausalSelfAttention(nn.Module):
105
+
106
+ def __init__(self, config):
107
+ super().__init__()
108
+ assert config.n_embd % config.n_head == 0
109
+ # key, query, value projections for all heads, but in a batch
110
+ self.c_attn = BitLinear158b(config.n_embd, 3 * config.n_embd, bias=config.bias)
111
+ # output projection
112
+ self.c_proj = BitLinear158b(config.n_embd, config.n_embd, bias=config.bias)
113
+ # regularization
114
+ self.attn_dropout = nn.Dropout(config.dropout)
115
+ self.resid_dropout = nn.Dropout(config.dropout)
116
+ self.n_head = config.n_head
117
+ self.n_embd = config.n_embd
118
+ self.dropout = config.dropout
119
+ # flash attention make GPU go brrrrr but support is only in PyTorch >= 2.0
120
+ self.flash = hasattr(torch.nn.functional, 'scaled_dot_product_attention')
121
+ if not self.flash:
122
+ print("WARNING: using slow attention. Flash Attention requires PyTorch >= 2.0")
123
+ # causal mask to ensure that attention is only applied to the left in the input sequence
124
+ self.register_buffer("bias", torch.tril(torch.ones(config.block_size, config.block_size))
125
+ .view(1, 1, config.block_size, config.block_size))
126
+
127
+ def forward(self, x):
128
+ B, T, C = x.size() # batch size, sequence length, embedding dimensionality (n_embd)
129
+
130
+ # calculate query, key, values for all heads in batch and move head forward to be the batch dim
131
+ q, k, v = self.c_attn(x).split(self.n_embd, dim=2)
132
+ k = k.view(B, T, self.n_head, C // self.n_head).transpose(1, 2) # (B, nh, T, hs)
133
+ q = q.view(B, T, self.n_head, C // self.n_head).transpose(1, 2) # (B, nh, T, hs)
134
+ v = v.view(B, T, self.n_head, C // self.n_head).transpose(1, 2) # (B, nh, T, hs)
135
+
136
+ # causal self-attention; Self-attend: (B, nh, T, hs) x (B, nh, hs, T) -> (B, nh, T, T)
137
+ if self.flash:
138
+ # efficient attention using Flash Attention CUDA kernels
139
+ y = torch.nn.functional.scaled_dot_product_attention(q, k, v, attn_mask=None, dropout_p=self.dropout if self.training else 0, is_causal=True)
140
+ else:
141
+ # manual implementation of attention
142
+ att = (q @ k.transpose(-2, -1)) * (1.0 / math.sqrt(k.size(-1)))
143
+ att = att.masked_fill(self.bias[:,:,:T,:T] == 0, float('-inf'))
144
+ att = F.softmax(att, dim=-1)
145
+ att = self.attn_dropout(att)
146
+ y = att @ v # (B, nh, T, T) x (B, nh, T, hs) -> (B, nh, T, hs)
147
+ y = y.transpose(1, 2).contiguous().view(B, T, C) # re-assemble all head outputs side by side
148
+
149
+ # output projection
150
+ y = self.resid_dropout(self.c_proj(y))
151
+ return y
152
+
153
+ class MLP(nn.Module):
154
+
155
+ def __init__(self, config):
156
+ super().__init__()
157
+ self.c_fc = BitLinear158b(config.n_embd, 4 * config.n_embd, bias=config.bias)
158
+ self.gelu = nn.GELU()
159
+ self.c_proj = BitLinear158b(4 * config.n_embd, config.n_embd, bias=config.bias)
160
+ self.dropout = nn.Dropout(config.dropout)
161
+
162
+ def forward(self, x):
163
+ x = self.c_fc(x)
164
+ x = self.gelu(x)
165
+ x = self.c_proj(x)
166
+ x = self.dropout(x)
167
+ return x
168
+
169
+ class Block(nn.Module):
170
+
171
+ def __init__(self, config):
172
+ super().__init__()
173
+ self.ln_1 = LayerNorm(config.n_embd, bias=config.bias)
174
+ self.attn = CausalSelfAttention(config)
175
+ self.ln_2 = LayerNorm(config.n_embd, bias=config.bias)
176
+ self.mlp = MLP(config)
177
+
178
+ def forward(self, x):
179
+ x = x + self.attn(self.ln_1(x))
180
+ x = x + self.mlp(self.ln_2(x))
181
+ return x
182
+
183
+ @dataclass
184
+ class GPTConfig:
185
+ block_size: int = 1024
186
+ vocab_size: int = 50304 # GPT-2 vocab_size of 50257, padded up to nearest multiple of 64 for efficiency
187
+ n_layer: int = 12
188
+ n_head: int = 12
189
+ n_embd: int = 768
190
+ dropout: float = 0.0
191
+ bias: bool = True # True: bias in Linears and LayerNorms, like GPT-2. False: a bit better and faster
192
+
193
+ class GPT(nn.Module):
194
+
195
+ def __init__(self, config):
196
+ super().__init__()
197
+ assert config.vocab_size is not None
198
+ assert config.block_size is not None
199
+ self.config = config
200
+
201
+ self.transformer = nn.ModuleDict(dict(
202
+ wte = nn.Embedding(config.vocab_size, config.n_embd),
203
+ wpe = nn.Embedding(config.block_size, config.n_embd),
204
+ drop = nn.Dropout(config.dropout),
205
+ h = nn.ModuleList([Block(config) for _ in range(config.n_layer)]),
206
+ ln_f = LayerNorm(config.n_embd, bias=config.bias),
207
+ ))
208
+ self.lm_head = BitLinear158b(config.n_embd, config.vocab_size, bias=False)
209
+ # with weight tying when using torch.compile() some warnings get generated:
210
+ # "UserWarning: functional_call was passed multiple values for tied weights.
211
+ # This behavior is deprecated and will be an error in future versions"
212
+ # not 100% sure what this is, so far seems to be harmless. TODO investigate
213
+ self.transformer.wte.weight = self.lm_head.weight # https://paperswithcode.com/method/weight-tying
214
+
215
+ # init all weights
216
+ self.apply(self._init_weights)
217
+ # apply special scaled init to the residual projections, per GPT-2 paper
218
+ for pn, p in self.named_parameters():
219
+ if pn.endswith('c_proj.weight'):
220
+ torch.nn.init.normal_(p, mean=0.0, std=0.02/math.sqrt(2 * config.n_layer))
221
+
222
+ # report number of parameters
223
+ print("number of parameters: %.2fM" % (self.get_num_params()/1e6,))
224
+
225
+ def get_num_params(self, non_embedding=True):
226
+ """
227
+ Return the number of parameters in the model.
228
+ For non-embedding count (default), the position embeddings get subtracted.
229
+ The token embeddings would too, except due to the parameter sharing these
230
+ params are actually used as weights in the final layer, so we include them.
231
+ """
232
+ n_params = sum(p.numel() for p in self.parameters())
233
+ if non_embedding:
234
+ n_params -= self.transformer.wpe.weight.numel()
235
+ return n_params
236
+
237
+ def _init_weights(self, module):
238
+ if isinstance(module, nn.Linear):
239
+ torch.nn.init.normal_(module.weight, mean=0.0, std=0.02)
240
+ if module.bias is not None:
241
+ torch.nn.init.zeros_(module.bias)
242
+ elif isinstance(module, BitLinear158b):
243
+ torch.nn.init.normal_(module.weight, mean=0.0, std=0.02)
244
+ if module.bias is not None:
245
+ torch.nn.init.zeros_(module.bias)
246
+ elif isinstance(module, nn.Embedding):
247
+ torch.nn.init.normal_(module.weight, mean=0.0, std=0.02)
248
+
249
+ def forward(self, idx, targets=None):
250
+ device = idx.device
251
+ b, t = idx.size()
252
+ assert t <= self.config.block_size, f"Cannot forward sequence of length {t}, block size is only {self.config.block_size}"
253
+ pos = torch.arange(0, t, dtype=torch.long, device=device) # shape (t)
254
+
255
+ # forward the GPT model itself
256
+ tok_emb = self.transformer.wte(idx) # token embeddings of shape (b, t, n_embd)
257
+ pos_emb = self.transformer.wpe(pos) # position embeddings of shape (t, n_embd)
258
+ x = self.transformer.drop(tok_emb + pos_emb)
259
+ for block in self.transformer.h:
260
+ x = block(x)
261
+ x = self.transformer.ln_f(x)
262
+
263
+ if targets is not None:
264
+ # if we are given some desired targets also calculate the loss
265
+ logits = self.lm_head(x)
266
+ loss = F.cross_entropy(logits.view(-1, logits.size(-1)), targets.view(-1), ignore_index=-1)
267
+ else:
268
+ # inference-time mini-optimization: only forward the lm_head on the very last position
269
+ logits = self.lm_head(x[:, [-1], :]) # note: using list [-1] to preserve the time dim
270
+ loss = None
271
+
272
+ return logits, loss
273
+
274
+ def crop_block_size(self, block_size):
275
+ # model surgery to decrease the block size if necessary
276
+ # e.g. we may load the GPT2 pretrained model checkpoint (block size 1024)
277
+ # but want to use a smaller block size for some smaller, simpler model
278
+ assert block_size <= self.config.block_size
279
+ self.config.block_size = block_size
280
+ self.transformer.wpe.weight = nn.Parameter(self.transformer.wpe.weight[:block_size])
281
+ for block in self.transformer.h:
282
+ if hasattr(block.attn, 'bias'):
283
+ block.attn.bias = block.attn.bias[:,:,:block_size,:block_size]
284
+
285
+ @classmethod
286
+ def from_pretrained(cls, model_type, override_args=None):
287
+ assert model_type in {'gpt2', 'gpt2-medium', 'gpt2-large', 'gpt2-xl'}
288
+ override_args = override_args or {} # default to empty dict
289
+ # only dropout can be overridden see more notes below
290
+ assert all(k == 'dropout' for k in override_args)
291
+ from transformers import GPT2LMHeadModel
292
+ print("loading weights from pretrained gpt: %s" % model_type)
293
+
294
+ # n_layer, n_head and n_embd are determined from model_type
295
+ config_args = {
296
+ 'gpt2': dict(n_layer=12, n_head=12, n_embd=768), # 124M params
297
+ 'gpt2-medium': dict(n_layer=24, n_head=16, n_embd=1024), # 350M params
298
+ 'gpt2-large': dict(n_layer=36, n_head=20, n_embd=1280), # 774M params
299
+ 'gpt2-xl': dict(n_layer=48, n_head=25, n_embd=1600), # 1558M params
300
+ }[model_type]
301
+ print("forcing vocab_size=50257, block_size=1024, bias=True")
302
+ config_args['vocab_size'] = 50257 # always 50257 for GPT model checkpoints
303
+ config_args['block_size'] = 1024 # always 1024 for GPT model checkpoints
304
+ config_args['bias'] = True # always True for GPT model checkpoints
305
+ # we can override the dropout rate, if desired
306
+ if 'dropout' in override_args:
307
+ print(f"overriding dropout rate to {override_args['dropout']}")
308
+ config_args['dropout'] = override_args['dropout']
309
+ # create a from-scratch initialized minGPT model
310
+ config = GPTConfig(**config_args)
311
+ model = GPT(config)
312
+ sd = model.state_dict()
313
+ sd_keys = sd.keys()
314
+ sd_keys = [k for k in sd_keys if not k.endswith('.attn.bias')] # discard this mask / buffer, not a param
315
+
316
+ # init a huggingface/transformers model
317
+ model_hf = GPT2LMHeadModel.from_pretrained(model_type)
318
+ sd_hf = model_hf.state_dict()
319
+
320
+ # copy while ensuring all of the parameters are aligned and match in names and shapes
321
+ sd_keys_hf = sd_hf.keys()
322
+ sd_keys_hf = [k for k in sd_keys_hf if not k.endswith('.attn.masked_bias')] # ignore these, just a buffer
323
+ sd_keys_hf = [k for k in sd_keys_hf if not k.endswith('.attn.bias')] # same, just the mask (buffer)
324
+ transposed = ['attn.c_attn.weight', 'attn.c_proj.weight', 'mlp.c_fc.weight', 'mlp.c_proj.weight']
325
+ # basically the openai checkpoints use a "Conv1D" module, but we only want to use a vanilla Linear
326
+ # this means that we have to transpose these weights when we import them
327
+ assert len(sd_keys_hf) == len(sd_keys), f"mismatched keys: {len(sd_keys_hf)} != {len(sd_keys)}"
328
+ for k in sd_keys_hf:
329
+ if any(k.endswith(w) for w in transposed):
330
+ # special treatment for the Conv1D weights we need to transpose
331
+ assert sd_hf[k].shape[::-1] == sd[k].shape
332
+ with torch.no_grad():
333
+ sd[k].copy_(sd_hf[k].t())
334
+ else:
335
+ # vanilla copy over the other parameters
336
+ assert sd_hf[k].shape == sd[k].shape
337
+ with torch.no_grad():
338
+ sd[k].copy_(sd_hf[k])
339
+
340
+ return model
341
+
342
+ def configure_optimizers(self, weight_decay, learning_rate, betas, device_type):
343
+ # start with all of the candidate parameters
344
+ param_dict = {pn: p for pn, p in self.named_parameters()}
345
+ # filter out those that do not require grad
346
+ param_dict = {pn: p for pn, p in param_dict.items() if p.requires_grad}
347
+ # create optim groups. Any parameters that is 2D will be weight decayed, otherwise no.
348
+ # i.e. all weight tensors in matmuls + embeddings decay, all biases and layernorms don't.
349
+ decay_params = [p for n, p in param_dict.items() if p.dim() >= 2]
350
+ nodecay_params = [p for n, p in param_dict.items() if p.dim() < 2]
351
+ optim_groups = [
352
+ {'params': decay_params, 'weight_decay': weight_decay},
353
+ {'params': nodecay_params, 'weight_decay': 0.0}
354
+ ]
355
+ num_decay_params = sum(p.numel() for p in decay_params)
356
+ num_nodecay_params = sum(p.numel() for p in nodecay_params)
357
+ print(f"num decayed parameter tensors: {len(decay_params)}, with {num_decay_params:,} parameters")
358
+ print(f"num non-decayed parameter tensors: {len(nodecay_params)}, with {num_nodecay_params:,} parameters")
359
+ # Create AdamW optimizer and use the fused version if it is available
360
+ fused_available = 'fused' in inspect.signature(torch.optim.AdamW).parameters
361
+ use_fused = fused_available and device_type == 'cuda'
362
+ extra_args = dict(fused=True) if use_fused else dict()
363
+ optimizer = torch.optim.AdamW(optim_groups, lr=learning_rate, betas=betas, **extra_args)
364
+ print(f"using fused AdamW: {use_fused}")
365
+
366
+ return optimizer
367
+
368
+ def estimate_mfu(self, fwdbwd_per_iter, dt):
369
+ """ estimate model flops utilization (MFU) in units of A100 bfloat16 peak FLOPS """
370
+ # first estimate the number of flops we do per iteration.
371
+ # see PaLM paper Appendix B as ref: https://arxiv.org/abs/2204.02311
372
+ N = self.get_num_params()
373
+ cfg = self.config
374
+ L, H, Q, T = cfg.n_layer, cfg.n_head, cfg.n_embd//cfg.n_head, cfg.block_size
375
+ flops_per_token = 6*N + 12*L*H*Q*T
376
+ flops_per_fwdbwd = flops_per_token * T
377
+ flops_per_iter = flops_per_fwdbwd * fwdbwd_per_iter
378
+ # express our flops throughput as ratio of A100 bfloat16 peak flops
379
+ flops_achieved = flops_per_iter * (1.0/dt) # per second
380
+ flops_promised = 312e12 # A100 GPU bfloat16 peak flops is 312 TFLOPS
381
+ mfu = flops_achieved / flops_promised
382
+ return mfu
383
+
384
+ @torch.no_grad()
385
+ def generate(self, idx, max_new_tokens, temperature=1.0, top_k=None):
386
+ """
387
+ Take a conditioning sequence of indices idx (LongTensor of shape (b,t)) and complete
388
+ the sequence max_new_tokens times, feeding the predictions back into the model each time.
389
+ Most likely you'll want to make sure to be in model.eval() mode of operation for this.
390
+ """
391
+ for _ in range(max_new_tokens):
392
+ # if the sequence context is growing too long we must crop it at block_size
393
+ idx_cond = idx if idx.size(1) <= self.config.block_size else idx[:, -self.config.block_size:]
394
+ # forward the model to get the logits for the index in the sequence
395
+ logits, _ = self(idx_cond)
396
+ # pluck the logits at the final step and scale by desired temperature
397
+ logits = logits[:, -1, :] / temperature
398
+ # optionally crop the logits to only the top k options
399
+ if top_k is not None:
400
+ v, _ = torch.topk(logits, min(top_k, logits.size(-1)))
401
+ logits[logits < v[:, [-1]]] = -float('Inf')
402
+ # apply softmax to convert logits to (normalized) probabilities
403
+ probs = F.softmax(logits, dim=-1)
404
+ # sample from the distribution
405
+ idx_next = torch.multinomial(probs, num_samples=1)
406
+ # append sampled index to the running sequence and continue
407
+ idx = torch.cat((idx, idx_next), dim=1)
408
+
409
+ return idx
out-shakespeare-char/ckpt.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee4d1c19709b37b859cd0d0d43460c8defb17b3280e903c8e4bd6887c457aad1
3
+ size 128986325
out/ckpt.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e2e32b8589b2a59914f136c93b8eece1213a3a05bfe6a0eaa4620ba36452bab9
3
+ size 1492568069
sample.py ADDED
@@ -0,0 +1,89 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """
2
+ Sample from a trained model
3
+ """
4
+ import os
5
+ import pickle
6
+ from contextlib import nullcontext
7
+ import torch
8
+ import tiktoken
9
+ from model import GPTConfig, GPT
10
+
11
+ # -----------------------------------------------------------------------------
12
+ init_from = 'resume' # either 'resume' (from an out_dir) or a gpt2 variant (e.g. 'gpt2-xl')
13
+ out_dir = 'out' # ignored if init_from is not 'resume'
14
+ start = "\n" # or "<|endoftext|>" or etc. Can also specify a file, use as: "FILE:prompt.txt"
15
+ num_samples = 10 # number of samples to draw
16
+ max_new_tokens = 500 # number of tokens generated in each sample
17
+ temperature = 0.8 # 1.0 = no change, < 1.0 = less random, > 1.0 = more random, in predictions
18
+ top_k = 200 # retain only the top_k most likely tokens, clamp others to have 0 probability
19
+ seed = 1337
20
+ device = 'cuda' # examples: 'cpu', 'cuda', 'cuda:0', 'cuda:1', etc.
21
+ dtype = 'bfloat16' if torch.cuda.is_available() and torch.cuda.is_bf16_supported() else 'float16' # 'float32' or 'bfloat16' or 'float16'
22
+ compile = False # use PyTorch 2.0 to compile the model to be faster
23
+ exec(open('configurator.py').read()) # overrides from command line or config file
24
+ # -----------------------------------------------------------------------------
25
+
26
+ torch.manual_seed(seed)
27
+ torch.cuda.manual_seed(seed)
28
+ torch.backends.cuda.matmul.allow_tf32 = True # allow tf32 on matmul
29
+ torch.backends.cudnn.allow_tf32 = True # allow tf32 on cudnn
30
+ device_type = 'cuda' if 'cuda' in device else 'cpu' # for later use in torch.autocast
31
+ ptdtype = {'float32': torch.float32, 'bfloat16': torch.bfloat16, 'float16': torch.float16}[dtype]
32
+ ctx = nullcontext() if device_type == 'cpu' else torch.amp.autocast(device_type=device_type, dtype=ptdtype)
33
+
34
+ # model
35
+ if init_from == 'resume':
36
+ # init from a model saved in a specific directory
37
+ ckpt_path = os.path.join(out_dir, 'ckpt.pt')
38
+ checkpoint = torch.load(ckpt_path, map_location=device)
39
+ gptconf = GPTConfig(**checkpoint['model_args'])
40
+ model = GPT(gptconf)
41
+ state_dict = checkpoint['model']
42
+ unwanted_prefix = '_orig_mod.'
43
+ for k,v in list(state_dict.items()):
44
+ if k.startswith(unwanted_prefix):
45
+ state_dict[k[len(unwanted_prefix):]] = state_dict.pop(k)
46
+ model.load_state_dict(state_dict)
47
+ elif init_from.startswith('gpt2'):
48
+ # init from a given GPT-2 model
49
+ model = GPT.from_pretrained(init_from, dict(dropout=0.0))
50
+
51
+ model.eval()
52
+ model.to(device)
53
+ if compile:
54
+ model = torch.compile(model) # requires PyTorch 2.0 (optional)
55
+
56
+ # look for the meta pickle in case it is available in the dataset folder
57
+ load_meta = False
58
+ if init_from == 'resume' and 'config' in checkpoint and 'dataset' in checkpoint['config']: # older checkpoints might not have these...
59
+ meta_path = os.path.join('data', checkpoint['config']['dataset'], 'meta.pkl')
60
+ load_meta = os.path.exists(meta_path)
61
+ if load_meta:
62
+ print(f"Loading meta from {meta_path}...")
63
+ with open(meta_path, 'rb') as f:
64
+ meta = pickle.load(f)
65
+ # TODO want to make this more general to arbitrary encoder/decoder schemes
66
+ stoi, itos = meta['stoi'], meta['itos']
67
+ encode = lambda s: [stoi[c] for c in s]
68
+ decode = lambda l: ''.join([itos[i] for i in l])
69
+ else:
70
+ # ok let's assume gpt-2 encodings by default
71
+ print("No meta.pkl found, assuming GPT-2 encodings...")
72
+ enc = tiktoken.get_encoding("gpt2")
73
+ encode = lambda s: enc.encode(s, allowed_special={"<|endoftext|>"})
74
+ decode = lambda l: enc.decode(l)
75
+
76
+ # encode the beginning of the prompt
77
+ if start.startswith('FILE:'):
78
+ with open(start[5:], 'r', encoding='utf-8') as f:
79
+ start = f.read()
80
+ start_ids = encode(start)
81
+ x = (torch.tensor(start_ids, dtype=torch.long, device=device)[None, ...])
82
+
83
+ # run generation
84
+ with torch.no_grad():
85
+ with ctx:
86
+ for k in range(num_samples):
87
+ y = model.generate(x, max_new_tokens, temperature=temperature, top_k=top_k)
88
+ print(decode(y[0].tolist()))
89
+ print('---------------')
scaling_laws.ipynb ADDED
The diff for this file is too large to render. See raw diff
 
train.py ADDED
@@ -0,0 +1,336 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """
2
+ This training script can be run both on a single gpu in debug mode,
3
+ and also in a larger training run with distributed data parallel (ddp).
4
+
5
+ To run on a single GPU, example:
6
+ $ python train.py --batch_size=32 --compile=False
7
+
8
+ To run with DDP on 4 gpus on 1 node, example:
9
+ $ torchrun --standalone --nproc_per_node=4 train.py
10
+
11
+ To run with DDP on 4 gpus across 2 nodes, example:
12
+ - Run on the first (master) node with example IP 123.456.123.456:
13
+ $ torchrun --nproc_per_node=8 --nnodes=2 --node_rank=0 --master_addr=123.456.123.456 --master_port=1234 train.py
14
+ - Run on the worker node:
15
+ $ torchrun --nproc_per_node=8 --nnodes=2 --node_rank=1 --master_addr=123.456.123.456 --master_port=1234 train.py
16
+ (If your cluster does not have Infiniband interconnect prepend NCCL_IB_DISABLE=1)
17
+ """
18
+
19
+ import os
20
+ import time
21
+ import math
22
+ import pickle
23
+ from contextlib import nullcontext
24
+
25
+ import numpy as np
26
+ import torch
27
+ from torch.nn.parallel import DistributedDataParallel as DDP
28
+ from torch.distributed import init_process_group, destroy_process_group
29
+
30
+ from model import GPTConfig, GPT
31
+
32
+ # -----------------------------------------------------------------------------
33
+ # default config values designed to train a gpt2 (124M) on OpenWebText
34
+ # I/O
35
+ out_dir = 'out'
36
+ eval_interval = 2000
37
+ log_interval = 1
38
+ eval_iters = 200
39
+ eval_only = False # if True, script exits right after the first eval
40
+ always_save_checkpoint = True # if True, always save a checkpoint after each eval
41
+ init_from = 'scratch' # 'scratch' or 'resume' or 'gpt2*'
42
+ # wandb logging
43
+ wandb_log = False # disabled by default
44
+ wandb_project = 'owt'
45
+ wandb_run_name = 'gpt2' # 'run' + str(time.time())
46
+ # data
47
+ dataset = 'openwebtext'
48
+ gradient_accumulation_steps = 5 * 8 # used to simulate larger batch sizes
49
+ batch_size = 12 # if gradient_accumulation_steps > 1, this is the micro-batch size
50
+ block_size = 1024
51
+ # model
52
+ n_layer = 12
53
+ n_head = 12
54
+ n_embd = 768
55
+ dropout = 0.0 # for pretraining 0 is good, for finetuning try 0.1+
56
+ bias = False # do we use bias inside LayerNorm and Linear layers?
57
+ # adamw optimizer
58
+ learning_rate = 6e-4 # max learning rate
59
+ max_iters = 600000 # total number of training iterations
60
+ weight_decay = 1e-1
61
+ beta1 = 0.9
62
+ beta2 = 0.95
63
+ grad_clip = 1.0 # clip gradients at this value, or disable if == 0.0
64
+ # learning rate decay settings
65
+ decay_lr = True # whether to decay the learning rate
66
+ warmup_iters = 2000 # how many steps to warm up for
67
+ lr_decay_iters = 600000 # should be ~= max_iters per Chinchilla
68
+ min_lr = 6e-5 # minimum learning rate, should be ~= learning_rate/10 per Chinchilla
69
+ # DDP settings
70
+ backend = 'nccl' # 'nccl', 'gloo', etc.
71
+ # system
72
+ device = 'cuda' # examples: 'cpu', 'cuda', 'cuda:0', 'cuda:1' etc., or try 'mps' on macbooks
73
+ dtype = 'bfloat16' if torch.cuda.is_available() and torch.cuda.is_bf16_supported() else 'float16' # 'float32', 'bfloat16', or 'float16', the latter will auto implement a GradScaler
74
+ compile = False # use PyTorch 2.0 to compile the model to be faster, default off to unpredicted behaviors
75
+ # -----------------------------------------------------------------------------
76
+ config_keys = [k for k,v in globals().items() if not k.startswith('_') and isinstance(v, (int, float, bool, str))]
77
+ exec(open('configurator.py').read()) # overrides from command line or config file
78
+ config = {k: globals()[k] for k in config_keys} # will be useful for logging
79
+ # -----------------------------------------------------------------------------
80
+
81
+ # various inits, derived attributes, I/O setup
82
+ ddp = int(os.environ.get('RANK', -1)) != -1 # is this a ddp run?
83
+ if ddp:
84
+ init_process_group(backend=backend)
85
+ ddp_rank = int(os.environ['RANK'])
86
+ ddp_local_rank = int(os.environ['LOCAL_RANK'])
87
+ ddp_world_size = int(os.environ['WORLD_SIZE'])
88
+ device = f'cuda:{ddp_local_rank}'
89
+ torch.cuda.set_device(device)
90
+ master_process = ddp_rank == 0 # this process will do logging, checkpointing etc.
91
+ seed_offset = ddp_rank # each process gets a different seed
92
+ # world_size number of processes will be training simultaneously, so we can scale
93
+ # down the desired gradient accumulation iterations per process proportionally
94
+ assert gradient_accumulation_steps % ddp_world_size == 0
95
+ gradient_accumulation_steps //= ddp_world_size
96
+ else:
97
+ # if not ddp, we are running on a single gpu, and one process
98
+ master_process = True
99
+ seed_offset = 0
100
+ ddp_world_size = 1
101
+ tokens_per_iter = gradient_accumulation_steps * ddp_world_size * batch_size * block_size
102
+ print(f"tokens per iteration will be: {tokens_per_iter:,}")
103
+
104
+ if master_process:
105
+ os.makedirs(out_dir, exist_ok=True)
106
+ torch.manual_seed(1337 + seed_offset)
107
+ torch.backends.cuda.matmul.allow_tf32 = True # allow tf32 on matmul
108
+ torch.backends.cudnn.allow_tf32 = True # allow tf32 on cudnn
109
+ device_type = 'cuda' if 'cuda' in device else 'cpu' # for later use in torch.autocast
110
+ # note: float16 data type will automatically use a GradScaler
111
+ ptdtype = {'float32': torch.float32, 'bfloat16': torch.bfloat16, 'float16': torch.float16}[dtype]
112
+ ctx = nullcontext() if device_type == 'cpu' else torch.amp.autocast(device_type=device_type, dtype=ptdtype)
113
+
114
+ # poor man's data loader
115
+ data_dir = os.path.join('data', dataset)
116
+ def get_batch(split):
117
+ # We recreate np.memmap every batch to avoid a memory leak, as per
118
+ # https://stackoverflow.com/questions/45132940/numpy-memmap-memory-usage-want-to-iterate-once/61472122#61472122
119
+ if split == 'train':
120
+ data = np.memmap(os.path.join(data_dir, 'train.bin'), dtype=np.uint16, mode='r')
121
+ else:
122
+ data = np.memmap(os.path.join(data_dir, 'val.bin'), dtype=np.uint16, mode='r')
123
+ ix = torch.randint(len(data) - block_size, (batch_size,))
124
+ x = torch.stack([torch.from_numpy((data[i:i+block_size]).astype(np.int64)) for i in ix])
125
+ y = torch.stack([torch.from_numpy((data[i+1:i+1+block_size]).astype(np.int64)) for i in ix])
126
+ if device_type == 'cuda':
127
+ # pin arrays x,y, which allows us to move them to GPU asynchronously (non_blocking=True)
128
+ x, y = x.pin_memory().to(device, non_blocking=True), y.pin_memory().to(device, non_blocking=True)
129
+ else:
130
+ x, y = x.to(device), y.to(device)
131
+ return x, y
132
+
133
+ # init these up here, can override if init_from='resume' (i.e. from a checkpoint)
134
+ iter_num = 0
135
+ best_val_loss = 1e9
136
+
137
+ # attempt to derive vocab_size from the dataset
138
+ meta_path = os.path.join(data_dir, 'meta.pkl')
139
+ meta_vocab_size = None
140
+ if os.path.exists(meta_path):
141
+ with open(meta_path, 'rb') as f:
142
+ meta = pickle.load(f)
143
+ meta_vocab_size = meta['vocab_size']
144
+ print(f"found vocab_size = {meta_vocab_size} (inside {meta_path})")
145
+
146
+ # model init
147
+ model_args = dict(n_layer=n_layer, n_head=n_head, n_embd=n_embd, block_size=block_size,
148
+ bias=bias, vocab_size=None, dropout=dropout) # start with model_args from command line
149
+ if init_from == 'scratch':
150
+ # init a new model from scratch
151
+ print("Initializing a new model from scratch")
152
+ # determine the vocab size we'll use for from-scratch training
153
+ if meta_vocab_size is None:
154
+ print("defaulting to vocab_size of GPT-2 to 50304 (50257 rounded up for efficiency)")
155
+ model_args['vocab_size'] = meta_vocab_size if meta_vocab_size is not None else 50304
156
+ gptconf = GPTConfig(**model_args)
157
+ model = GPT(gptconf)
158
+ elif init_from == 'resume':
159
+ print(f"Resuming training from {out_dir}")
160
+ # resume training from a checkpoint.
161
+ ckpt_path = os.path.join(out_dir, 'ckpt.pt')
162
+ checkpoint = torch.load(ckpt_path, map_location=device)
163
+ checkpoint_model_args = checkpoint['model_args']
164
+ # force these config attributes to be equal otherwise we can't even resume training
165
+ # the rest of the attributes (e.g. dropout) can stay as desired from command line
166
+ for k in ['n_layer', 'n_head', 'n_embd', 'block_size', 'bias', 'vocab_size']:
167
+ model_args[k] = checkpoint_model_args[k]
168
+ # create the model
169
+ gptconf = GPTConfig(**model_args)
170
+ model = GPT(gptconf)
171
+ state_dict = checkpoint['model']
172
+ # fix the keys of the state dictionary :(
173
+ # honestly no idea how checkpoints sometimes get this prefix, have to debug more
174
+ unwanted_prefix = '_orig_mod.'
175
+ for k,v in list(state_dict.items()):
176
+ if k.startswith(unwanted_prefix):
177
+ state_dict[k[len(unwanted_prefix):]] = state_dict.pop(k)
178
+ model.load_state_dict(state_dict)
179
+ iter_num = checkpoint['iter_num']
180
+ best_val_loss = checkpoint['best_val_loss']
181
+ elif init_from.startswith('gpt2'):
182
+ print(f"Initializing from OpenAI GPT-2 weights: {init_from}")
183
+ # initialize from OpenAI GPT-2 weights
184
+ override_args = dict(dropout=dropout)
185
+ model = GPT.from_pretrained(init_from, override_args)
186
+ # read off the created config params, so we can store them into checkpoint correctly
187
+ for k in ['n_layer', 'n_head', 'n_embd', 'block_size', 'bias', 'vocab_size']:
188
+ model_args[k] = getattr(model.config, k)
189
+ # crop down the model block size if desired, using model surgery
190
+ if block_size < model.config.block_size:
191
+ model.crop_block_size(block_size)
192
+ model_args['block_size'] = block_size # so that the checkpoint will have the right value
193
+ model.to(device)
194
+
195
+ # initialize a GradScaler. If enabled=False scaler is a no-op
196
+ scaler = torch.cuda.amp.GradScaler(enabled=(dtype == 'float16'))
197
+
198
+ # optimizer
199
+ optimizer = model.configure_optimizers(weight_decay, learning_rate, (beta1, beta2), device_type)
200
+ if init_from == 'resume':
201
+ optimizer.load_state_dict(checkpoint['optimizer'])
202
+ checkpoint = None # free up memory
203
+
204
+ # compile the model
205
+ if compile:
206
+ print("compiling the model... (takes a ~minute)")
207
+ unoptimized_model = model
208
+ model = torch.compile(model) # requires PyTorch 2.0
209
+
210
+ # wrap model into DDP container
211
+ if ddp:
212
+ model = DDP(model, device_ids=[ddp_local_rank])
213
+
214
+ # helps estimate an arbitrarily accurate loss over either split using many batches
215
+ @torch.no_grad()
216
+ def estimate_loss():
217
+ out = {}
218
+ model.eval()
219
+ for split in ['train', 'val']:
220
+ losses = torch.zeros(eval_iters)
221
+ for k in range(eval_iters):
222
+ X, Y = get_batch(split)
223
+ with ctx:
224
+ logits, loss = model(X, Y)
225
+ losses[k] = loss.item()
226
+ out[split] = losses.mean()
227
+ model.train()
228
+ return out
229
+
230
+ # learning rate decay scheduler (cosine with warmup)
231
+ def get_lr(it):
232
+ # 1) linear warmup for warmup_iters steps
233
+ if it < warmup_iters:
234
+ return learning_rate * it / warmup_iters
235
+ # 2) if it > lr_decay_iters, return min learning rate
236
+ if it > lr_decay_iters:
237
+ return min_lr
238
+ # 3) in between, use cosine decay down to min learning rate
239
+ decay_ratio = (it - warmup_iters) / (lr_decay_iters - warmup_iters)
240
+ assert 0 <= decay_ratio <= 1
241
+ coeff = 0.5 * (1.0 + math.cos(math.pi * decay_ratio)) # coeff ranges 0..1
242
+ return min_lr + coeff * (learning_rate - min_lr)
243
+
244
+ # logging
245
+ if wandb_log and master_process:
246
+ import wandb
247
+ wandb.init(project=wandb_project, name=wandb_run_name, config=config)
248
+
249
+ # training loop
250
+ X, Y = get_batch('train') # fetch the very first batch
251
+ t0 = time.time()
252
+ local_iter_num = 0 # number of iterations in the lifetime of this process
253
+ raw_model = model.module if ddp else model # unwrap DDP container if needed
254
+ running_mfu = -1.0
255
+ while True:
256
+
257
+ # determine and set the learning rate for this iteration
258
+ lr = get_lr(iter_num) if decay_lr else learning_rate
259
+ for param_group in optimizer.param_groups:
260
+ param_group['lr'] = lr
261
+
262
+ # evaluate the loss on train/val sets and write checkpoints
263
+ if iter_num % eval_interval == 0 and master_process:
264
+ losses = estimate_loss()
265
+ print(f"step {iter_num}: train loss {losses['train']:.4f}, val loss {losses['val']:.4f}")
266
+ if wandb_log:
267
+ wandb.log({
268
+ "iter": iter_num,
269
+ "train/loss": losses['train'],
270
+ "val/loss": losses['val'],
271
+ "lr": lr,
272
+ "mfu": running_mfu*100, # convert to percentage
273
+ })
274
+ if losses['val'] < best_val_loss or always_save_checkpoint:
275
+ best_val_loss = losses['val']
276
+ if iter_num > 0:
277
+ checkpoint = {
278
+ 'model': raw_model.state_dict(),
279
+ 'optimizer': optimizer.state_dict(),
280
+ 'model_args': model_args,
281
+ 'iter_num': iter_num,
282
+ 'best_val_loss': best_val_loss,
283
+ 'config': config,
284
+ }
285
+ print(f"saving checkpoint to {out_dir}")
286
+ torch.save(checkpoint, os.path.join(out_dir, 'ckpt.pt'))
287
+ if iter_num == 0 and eval_only:
288
+ break
289
+
290
+ # forward backward update, with optional gradient accumulation to simulate larger batch size
291
+ # and using the GradScaler if data type is float16
292
+ for micro_step in range(gradient_accumulation_steps):
293
+ if ddp:
294
+ # in DDP training we only need to sync gradients at the last micro step.
295
+ # the official way to do this is with model.no_sync() context manager, but
296
+ # I really dislike that this bloats the code and forces us to repeat code
297
+ # looking at the source of that context manager, it just toggles this variable
298
+ model.require_backward_grad_sync = (micro_step == gradient_accumulation_steps - 1)
299
+ with ctx:
300
+ logits, loss = model(X, Y)
301
+ loss = loss / gradient_accumulation_steps # scale the loss to account for gradient accumulation
302
+ # immediately async prefetch next batch while model is doing the forward pass on the GPU
303
+ X, Y = get_batch('train')
304
+ # backward pass, with gradient scaling if training in fp16
305
+ scaler.scale(loss).backward()
306
+ # clip the gradient
307
+ if grad_clip != 0.0:
308
+ scaler.unscale_(optimizer)
309
+ torch.nn.utils.clip_grad_norm_(model.parameters(), grad_clip)
310
+ # step the optimizer and scaler if training in fp16
311
+ scaler.step(optimizer)
312
+ scaler.update()
313
+ # flush the gradients as soon as we can, no need for this memory anymore
314
+ optimizer.zero_grad(set_to_none=True)
315
+
316
+ # timing and logging
317
+ t1 = time.time()
318
+ dt = t1 - t0
319
+ t0 = t1
320
+ if iter_num % log_interval == 0 and master_process:
321
+ # get loss as float. note: this is a CPU-GPU sync point
322
+ # scale up to undo the division above, approximating the true total loss (exact would have been a sum)
323
+ lossf = loss.item() * gradient_accumulation_steps
324
+ if local_iter_num >= 5: # let the training loop settle a bit
325
+ mfu = raw_model.estimate_mfu(batch_size * gradient_accumulation_steps, dt)
326
+ running_mfu = mfu if running_mfu == -1.0 else 0.9*running_mfu + 0.1*mfu
327
+ print(f"iter {iter_num}: loss {lossf:.4f}, time {dt*1000:.2f}ms, mfu {running_mfu*100:.2f}%")
328
+ iter_num += 1
329
+ local_iter_num += 1
330
+
331
+ # termination conditions
332
+ if iter_num > max_iters:
333
+ break
334
+
335
+ if ddp:
336
+ destroy_process_group()
transformer_sizing.ipynb ADDED
@@ -0,0 +1,402 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cells": [
3
+ {
4
+ "attachments": {},
5
+ "cell_type": "markdown",
6
+ "metadata": {},
7
+ "source": [
8
+ "### Transformer Theoretical Model\n",
9
+ "\n",
10
+ "This notebook stores a bunch of analysis about a Transformer, e.g. estimates the number of FLOPs, parameters, peak memory footprint, checkpoint size, etc."
11
+ ]
12
+ },
13
+ {
14
+ "cell_type": "code",
15
+ "execution_count": 1,
16
+ "metadata": {},
17
+ "outputs": [],
18
+ "source": [
19
+ "from collections import OrderedDict"
20
+ ]
21
+ },
22
+ {
23
+ "cell_type": "code",
24
+ "execution_count": 2,
25
+ "metadata": {},
26
+ "outputs": [],
27
+ "source": [
28
+ "# config_args = {\n",
29
+ "# 'gpt2': dict(n_layer=12, n_head=12, n_embd=768), # 124M params\n",
30
+ "# 'gpt2-medium': dict(n_layer=24, n_head=16, n_embd=1024), # 350M params\n",
31
+ "# 'gpt2-large': dict(n_layer=36, n_head=20, n_embd=1280), # 774M params\n",
32
+ "# 'gpt2-xl': dict(n_layer=48, n_head=25, n_embd=1600), # 1558M params\n",
33
+ "# }[model_type]\n",
34
+ "\n",
35
+ "block_size = 1024\n",
36
+ "vocab_size = 50257\n",
37
+ "n_layer = 12\n",
38
+ "n_head = 12\n",
39
+ "n_embd = 768\n",
40
+ "bias = False\n",
41
+ "assert not bias, \"this notebook assumes bias=False just for simplicity\""
42
+ ]
43
+ },
44
+ {
45
+ "cell_type": "code",
46
+ "execution_count": 3,
47
+ "metadata": {},
48
+ "outputs": [
49
+ {
50
+ "name": "stdout",
51
+ "output_type": "stream",
52
+ "text": [
53
+ "we see: 124337664, expected: 124337664, match: True\n",
54
+ "name params ratio (%) \n",
55
+ "emebedding/position 786432 0.6325\n",
56
+ "embedding/token 38597376 31.0424\n",
57
+ "embedding 39383808 31.6749\n",
58
+ "attention/ln 768 0.0006\n",
59
+ "attention/kqv 1769472 1.4231\n",
60
+ "attention/proj 589824 0.4744\n",
61
+ "attention 2360064 1.8981\n",
62
+ "mlp/ln 768 0.0006\n",
63
+ "mlp/ffw 2359296 1.8975\n",
64
+ "mlp/proj 2359296 1.8975\n",
65
+ "mlp 4719360 3.7956\n",
66
+ "block 7079424 5.6937\n",
67
+ "transformer 84953088 68.3245\n",
68
+ "ln_f 768 0.0006\n",
69
+ "dense 0 0.0000\n",
70
+ "total 124337664 100.0000\n"
71
+ ]
72
+ }
73
+ ],
74
+ "source": [
75
+ "def params():\n",
76
+ " \"\"\" estimates the number of parameters in the model\"\"\"\n",
77
+ " out = OrderedDict()\n",
78
+ "\n",
79
+ " # token and position embeddings\n",
80
+ " out['emebedding/position'] = n_embd * block_size\n",
81
+ " out['embedding/token'] = n_embd * vocab_size\n",
82
+ " out['embedding'] = out['emebedding/position'] + out['embedding/token']\n",
83
+ "\n",
84
+ " # attention blocks\n",
85
+ " out['attention/ln'] = n_embd # note, bias=False in our LN\n",
86
+ " out['attention/kqv'] = n_embd * 3*n_embd\n",
87
+ " out['attention/proj'] = n_embd**2\n",
88
+ " out['attention'] = out['attention/ln'] + out['attention/kqv'] + out['attention/proj']\n",
89
+ "\n",
90
+ " # MLP blocks\n",
91
+ " ffw_size = 4*n_embd # feed forward size\n",
92
+ " out['mlp/ln'] = n_embd\n",
93
+ " out['mlp/ffw'] = n_embd * ffw_size\n",
94
+ " out['mlp/proj'] = ffw_size * n_embd\n",
95
+ " out['mlp'] = out['mlp/ln'] + out['mlp/ffw'] + out['mlp/proj']\n",
96
+ " \n",
97
+ " # the transformer and the rest of it\n",
98
+ " out['block'] = out['attention'] + out['mlp']\n",
99
+ " out['transformer'] = n_layer * out['block']\n",
100
+ " out['ln_f'] = n_embd # final layernorm\n",
101
+ " out['dense'] = 0 # 0 because of parameter sharing. This layer uses the weights from the embedding layer\n",
102
+ "\n",
103
+ " # total\n",
104
+ " out['total'] = out['embedding'] + out['transformer'] + out['ln_f'] + out['dense']\n",
105
+ "\n",
106
+ " return out\n",
107
+ "\n",
108
+ "# compare our param count to that reported by PyTorch\n",
109
+ "p = params()\n",
110
+ "params_total = p['total']\n",
111
+ "print(f\"we see: {params_total}, expected: {124337664}, match: {params_total == 124337664}\")\n",
112
+ "# create a header\n",
113
+ "print(f\"{'name':20s} {'params':10s} {'ratio (%)':10s}\")\n",
114
+ "for k,v in p.items():\n",
115
+ " print(f\"{k:20s} {v:10d} {v/params_total*100:10.4f}\")\n",
116
+ " "
117
+ ]
118
+ },
119
+ {
120
+ "cell_type": "code",
121
+ "execution_count": 4,
122
+ "metadata": {},
123
+ "outputs": [
124
+ {
125
+ "name": "stdout",
126
+ "output_type": "stream",
127
+ "text": [
128
+ "est checkpoint size: 1.49 GB\n",
129
+ "measured with wc -c ckpt.pt: 1542470366\n",
130
+ "fluff ratio: 103.38%\n"
131
+ ]
132
+ }
133
+ ],
134
+ "source": [
135
+ "# we can now calculate the size of each checkpoint\n",
136
+ "# params are stored in fp32, and the AdamW optimizer has 2 additional buffers per param for statistics\n",
137
+ "params_bytes = params_total*4\n",
138
+ "params_and_buffers_bytes = params_bytes + 2*params_bytes\n",
139
+ "print(f\"est checkpoint size: {params_and_buffers_bytes/1e9:.2f} GB\")\n",
140
+ "measured_bytes = 1542470366 # from wc -c ckpt.pt\n",
141
+ "print(f\"measured with wc -c ckpt.pt: {measured_bytes}\")\n",
142
+ "print(f\"fluff ratio: {measured_bytes/params_and_buffers_bytes*100:.2f}%\")"
143
+ ]
144
+ },
145
+ {
146
+ "attachments": {},
147
+ "cell_type": "markdown",
148
+ "metadata": {},
149
+ "source": [
150
+ "We can also estimate the ratio of our GPU memory that will be taken up just by the weights and the buffers inside the AdamW optimizer"
151
+ ]
152
+ },
153
+ {
154
+ "cell_type": "code",
155
+ "execution_count": 5,
156
+ "metadata": {},
157
+ "outputs": [
158
+ {
159
+ "name": "stdout",
160
+ "output_type": "stream",
161
+ "text": [
162
+ "memory ratio taken up just for parameters: 3.73%\n"
163
+ ]
164
+ }
165
+ ],
166
+ "source": [
167
+ "gpu_memory = 40e9 # 40 GB A100 GPU, roughly\n",
168
+ "print(f\"memory ratio taken up just for parameters: {params_and_buffers_bytes / gpu_memory * 100:.2f}%\")"
169
+ ]
170
+ },
171
+ {
172
+ "attachments": {},
173
+ "cell_type": "markdown",
174
+ "metadata": {},
175
+ "source": [
176
+ "i.e. not that much of the memory for this tiny model, most of the memory is activations (forward and backward). This of course changes dramatically for larger and larger models."
177
+ ]
178
+ },
179
+ {
180
+ "attachments": {},
181
+ "cell_type": "markdown",
182
+ "metadata": {},
183
+ "source": [
184
+ "Let's estimate FLOPs for a single forward pass."
185
+ ]
186
+ },
187
+ {
188
+ "cell_type": "code",
189
+ "execution_count": 6,
190
+ "metadata": {},
191
+ "outputs": [
192
+ {
193
+ "name": "stdout",
194
+ "output_type": "stream",
195
+ "text": [
196
+ "name flops ratio (%) \n",
197
+ "attention/kqv 3623878656 1.2426\n",
198
+ "attention/scores 1610612736 0.5522\n",
199
+ "attention/reduce 1610612736 0.5522\n",
200
+ "attention/proj 1207959552 0.4142\n",
201
+ "attention 8053063680 2.7612\n",
202
+ "mlp/ffw1 4831838208 1.6567\n",
203
+ "mlp/ffw2 4831838208 1.6567\n",
204
+ "mlp 9663676416 3.3135\n",
205
+ "block 17716740096 6.0747\n",
206
+ "transformer 212600881152 72.8963\n",
207
+ "dense 79047426048 27.1037\n",
208
+ "forward_total 291648307200 100.0000\n",
209
+ "backward_total 583296614400 200.0000\n",
210
+ "total 874944921600 300.0000\n"
211
+ ]
212
+ }
213
+ ],
214
+ "source": [
215
+ "def flops():\n",
216
+ " # we only count Weight FLOPs, all other layers (LayerNorm, Softmax, etc) are effectively irrelevant\n",
217
+ " # we count actual FLOPs, not MACs. Hence 2* all over the place\n",
218
+ " # basically for any matrix multiply A (BxC) @ B (CxD) -> (BxD) flops are 2*B*C*D\n",
219
+ "\n",
220
+ " out = OrderedDict()\n",
221
+ " head_size = n_embd // n_head\n",
222
+ "\n",
223
+ " # attention blocks\n",
224
+ " # 1) the projection to key, query, values\n",
225
+ " out['attention/kqv'] = 2 * block_size * (n_embd * 3*n_embd)\n",
226
+ " # 2) calculating the attention scores\n",
227
+ " out['attention/scores'] = 2 * block_size * block_size * n_embd\n",
228
+ " # 3) the reduction of the values (B, nh, T, T) x (B, nh, T, hs) -> (B, nh, T, hs)\n",
229
+ " out['attention/reduce'] = 2 * n_head * (block_size * block_size * head_size)\n",
230
+ " # 4) the final linear projection\n",
231
+ " out['attention/proj'] = 2 * block_size * (n_embd * n_embd)\n",
232
+ " out['attention'] = sum(out['attention/'+k] for k in ['kqv', 'scores', 'reduce', 'proj'])\n",
233
+ "\n",
234
+ " # MLP blocks\n",
235
+ " ffw_size = 4*n_embd # feed forward size\n",
236
+ " out['mlp/ffw1'] = 2 * block_size * (n_embd * ffw_size)\n",
237
+ " out['mlp/ffw2'] = 2 * block_size * (ffw_size * n_embd)\n",
238
+ " out['mlp'] = out['mlp/ffw1'] + out['mlp/ffw2']\n",
239
+ "\n",
240
+ " # the transformer and the rest of it\n",
241
+ " out['block'] = out['attention'] + out['mlp']\n",
242
+ " out['transformer'] = n_layer * out['block']\n",
243
+ " out['dense'] = 2 * block_size * (n_embd * vocab_size)\n",
244
+ "\n",
245
+ " # forward,backward,total\n",
246
+ " out['forward_total'] = out['transformer'] + out['dense']\n",
247
+ " out['backward_total'] = 2 * out['forward_total'] # use common estimate of bwd = 2*fwd\n",
248
+ " out['total'] = out['forward_total'] + out['backward_total']\n",
249
+ "\n",
250
+ " return out\n",
251
+ " \n",
252
+ "# compare our param count to that reported by PyTorch\n",
253
+ "f = flops()\n",
254
+ "flops_total = f['forward_total']\n",
255
+ "print(f\"{'name':20s} {'flops':14s} {'ratio (%)':10s}\")\n",
256
+ "for k,v in f.items():\n",
257
+ " print(f\"{k:20s} {v:14d} {v/flops_total*100:10.4f}\")\n",
258
+ " "
259
+ ]
260
+ },
261
+ {
262
+ "cell_type": "code",
263
+ "execution_count": 7,
264
+ "metadata": {},
265
+ "outputs": [
266
+ {
267
+ "name": "stdout",
268
+ "output_type": "stream",
269
+ "text": [
270
+ "palm_flops: 875062886400, flops: 874944921600, ratio: 1.0001\n"
271
+ ]
272
+ }
273
+ ],
274
+ "source": [
275
+ "# now here is an estimate copy pasted from the PaLM paper\n",
276
+ "# this formula is often used to calculate MFU (model flops utilization)\n",
277
+ "def palm_flops():\n",
278
+ " \"\"\"estimate of the model flops following PaLM paper formula\"\"\"\n",
279
+ " # non-embedding model parameters. note that we do not subtract the\n",
280
+ " # embedding/token params because those are tied and get used in the last layer.\n",
281
+ " N = params()['total'] - params()['emebedding/position']\n",
282
+ " L, H, Q, T = n_layer, n_head, n_embd//n_head, block_size\n",
283
+ " mf_per_token = 6*N + 12*L*H*Q*T\n",
284
+ " mf = mf_per_token * block_size\n",
285
+ " return mf\n",
286
+ "\n",
287
+ "print(f\"palm_flops: {palm_flops():d}, flops: {flops()['total']:d}, ratio: {palm_flops()/flops()['total']:.4f}\")"
288
+ ]
289
+ },
290
+ {
291
+ "attachments": {},
292
+ "cell_type": "markdown",
293
+ "metadata": {},
294
+ "source": [
295
+ "Ok they are quite similar, giving some confidence that my math in flops() function was ~ok. Now, A100 is cited at 312TFLOPS bfloat16 on tensor cores. So what is our model flops utilization (MFU)? I trained the model above with a batch_size of 20 and grad_accum of 5, which runs in about 755ms on a single A100 GPU. We get:"
296
+ ]
297
+ },
298
+ {
299
+ "cell_type": "code",
300
+ "execution_count": 8,
301
+ "metadata": {},
302
+ "outputs": [
303
+ {
304
+ "name": "stdout",
305
+ "output_type": "stream",
306
+ "text": [
307
+ "fraction of A100 used: 37.14%\n"
308
+ ]
309
+ }
310
+ ],
311
+ "source": [
312
+ "# here is what we currently roughly measure\n",
313
+ "batch_size = 20 * 5 # 5 is grad_accum, so total batch size is 100\n",
314
+ "measured_time = 0.755 # in seconds per iteration\n",
315
+ "measured_throughput = batch_size / measured_time\n",
316
+ "flops_achieved = f['total'] * measured_throughput\n",
317
+ "\n",
318
+ "# A100 is cited to be 312 TFLOPS of bloat16 running on tensor cores\n",
319
+ "a100_flops_promised = 312e12\n",
320
+ "\n",
321
+ "# the fraction of the A100 that we are using:\n",
322
+ "print(f\"fraction of A100 used: {flops_achieved / a100_flops_promised * 100:.2f}%\")"
323
+ ]
324
+ },
325
+ {
326
+ "attachments": {},
327
+ "cell_type": "markdown",
328
+ "metadata": {},
329
+ "source": [
330
+ "For reference, we'd prefer to be somewhere around 50%+, and not just for a single GPU but for an entire DDP run. So we still have some work to do, but at least we're within a factor of ~2X of what is achievable with this GPU."
331
+ ]
332
+ },
333
+ {
334
+ "cell_type": "code",
335
+ "execution_count": 9,
336
+ "metadata": {},
337
+ "outputs": [
338
+ {
339
+ "name": "stdout",
340
+ "output_type": "stream",
341
+ "text": [
342
+ "time needed to train the model: 3.46 days\n"
343
+ ]
344
+ }
345
+ ],
346
+ "source": [
347
+ "# Finally let's check out the 6ND approximation as total cost of training in FLOPs\n",
348
+ "model_size = params()['total'] # this is number of parameters, N\n",
349
+ "tokens_num = 300e9 # 300B tokens, this is dataset size in tokens, D\n",
350
+ "a100_flops = 312e12 # 312 TFLOPS\n",
351
+ "assumed_mfu = 0.3 # assume this model flops utilization (take the current 37% from above and add some DDP overhead)\n",
352
+ "flops_throughput = a100_flops * 8 * assumed_mfu # assume an 8XA100 node at 30% utilization\n",
353
+ "flops_needed = 6 * model_size * tokens_num # 6ND\n",
354
+ "time_needed_s = flops_needed / flops_throughput # in seconds\n",
355
+ "print(f\"time needed to train the model: {time_needed_s/3600/24:.2f} days\")"
356
+ ]
357
+ },
358
+ {
359
+ "attachments": {},
360
+ "cell_type": "markdown",
361
+ "metadata": {},
362
+ "source": [
363
+ "This is not a bad estimate at all. I trained this model and it converged in roughly 4 days. Btw as a good reference for where 6ND comes from and some intuition around it I recommend [Dzmitry's post](https://medium.com/@dzmitrybahdanau/the-flops-calculus-of-language-model-training-3b19c1f025e4)."
364
+ ]
365
+ },
366
+ {
367
+ "attachments": {},
368
+ "cell_type": "markdown",
369
+ "metadata": {},
370
+ "source": [
371
+ "Now, FLOPs are just one constraint, the other that we have to keep a close track of is the memory bandwidth. TODO estimate LOAD/STORE costs of our model later."
372
+ ]
373
+ }
374
+ ],
375
+ "metadata": {
376
+ "kernelspec": {
377
+ "display_name": "pytorch2",
378
+ "language": "python",
379
+ "name": "python3"
380
+ },
381
+ "language_info": {
382
+ "codemirror_mode": {
383
+ "name": "ipython",
384
+ "version": 3
385
+ },
386
+ "file_extension": ".py",
387
+ "mimetype": "text/x-python",
388
+ "name": "python",
389
+ "nbconvert_exporter": "python",
390
+ "pygments_lexer": "ipython3",
391
+ "version": "3.10.8"
392
+ },
393
+ "orig_nbformat": 4,
394
+ "vscode": {
395
+ "interpreter": {
396
+ "hash": "7f5833218766b48e6e35e4452ee875aac0e2188d05bbe5298f2c62b79f08b222"
397
+ }
398
+ }
399
+ },
400
+ "nbformat": 4,
401
+ "nbformat_minor": 2
402
+ }
wandb/debug-internal.log ADDED
@@ -0,0 +1 @@
 
 
1
+ run-20240305_145418-2p2hzuh2/logs/debug-internal.log
wandb/debug.log ADDED
@@ -0,0 +1 @@
 
 
1
+ run-20240305_145418-2p2hzuh2/logs/debug.log
wandb/latest-run ADDED
@@ -0,0 +1 @@
 
 
1
+ run-20240305_145418-2p2hzuh2
wandb/run-20240305_145418-2p2hzuh2/files/config.yaml ADDED
@@ -0,0 +1,131 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ cli_version: 0.14.0
7
+ framework: huggingface
8
+ huggingface_version: 4.36.2
9
+ is_jupyter_run: false
10
+ is_kaggle_kernel: false
11
+ python_version: 3.10.12
12
+ start_time: 1709650458.652402
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 11
17
+ - 49
18
+ - 55
19
+ 2:
20
+ - 1
21
+ - 11
22
+ - 49
23
+ - 55
24
+ 3:
25
+ - 13
26
+ - 16
27
+ - 23
28
+ 4: 3.10.12
29
+ 5: 0.14.0
30
+ 6: 4.36.2
31
+ 8:
32
+ - 5
33
+ always_save_checkpoint:
34
+ desc: null
35
+ value: true
36
+ backend:
37
+ desc: null
38
+ value: nccl
39
+ batch_size:
40
+ desc: null
41
+ value: 12
42
+ beta1:
43
+ desc: null
44
+ value: 0.9
45
+ beta2:
46
+ desc: null
47
+ value: 0.95
48
+ bias:
49
+ desc: null
50
+ value: false
51
+ block_size:
52
+ desc: null
53
+ value: 1024
54
+ compile:
55
+ desc: null
56
+ value: false
57
+ dataset:
58
+ desc: null
59
+ value: openwebtext
60
+ decay_lr:
61
+ desc: null
62
+ value: true
63
+ device:
64
+ desc: null
65
+ value: cuda
66
+ dropout:
67
+ desc: null
68
+ value: 0.0
69
+ dtype:
70
+ desc: null
71
+ value: bfloat16
72
+ eval_interval:
73
+ desc: null
74
+ value: 1000
75
+ eval_iters:
76
+ desc: null
77
+ value: 200
78
+ eval_only:
79
+ desc: null
80
+ value: false
81
+ grad_clip:
82
+ desc: null
83
+ value: 1.0
84
+ gradient_accumulation_steps:
85
+ desc: null
86
+ value: 40
87
+ init_from:
88
+ desc: null
89
+ value: scratch
90
+ learning_rate:
91
+ desc: null
92
+ value: 0.0006
93
+ log_interval:
94
+ desc: null
95
+ value: 10
96
+ lr_decay_iters:
97
+ desc: null
98
+ value: 600000
99
+ max_iters:
100
+ desc: null
101
+ value: 600000
102
+ min_lr:
103
+ desc: null
104
+ value: 6.0e-05
105
+ n_embd:
106
+ desc: null
107
+ value: 768
108
+ n_head:
109
+ desc: null
110
+ value: 12
111
+ n_layer:
112
+ desc: null
113
+ value: 12
114
+ out_dir:
115
+ desc: null
116
+ value: out
117
+ wandb_log:
118
+ desc: null
119
+ value: true
120
+ wandb_project:
121
+ desc: null
122
+ value: owt
123
+ wandb_run_name:
124
+ desc: null
125
+ value: gpt2-124M
126
+ warmup_iters:
127
+ desc: null
128
+ value: 2000
129
+ weight_decay:
130
+ desc: null
131
+ value: 0.1
wandb/run-20240305_145418-2p2hzuh2/files/output.log ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0b6c5aee22a0850652c64e911a704560dfd1ed84edae9f8ff5ddf10671a7c2b
3
+ size 968720
wandb/run-20240305_145418-2p2hzuh2/files/requirements.txt ADDED
@@ -0,0 +1,329 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ absl-py==1.4.0
2
+ accelerate==0.26.1
3
+ aiodns==3.0.0
4
+ aiofiles==23.1.0
5
+ aiohttp-retry==2.8.3
6
+ aiohttp==3.8.4
7
+ aiosignal==1.3.1
8
+ aleph-alpha-client==2.16.0
9
+ altair==4.2.2
10
+ antlr4-python3-runtime==4.8
11
+ anyio==3.6.2
12
+ appdirs==1.4.4
13
+ argparse==1.4.0
14
+ astunparse==1.6.3
15
+ async-timeout==4.0.2
16
+ attr==0.3.2
17
+ attrs==22.2.0
18
+ automat==20.2.0
19
+ awscli==1.32.54
20
+ babel==2.8.0
21
+ bcrypt==3.2.0
22
+ beartype==0.17.1
23
+ bitarray==2.9.2
24
+ bitsandbytes==0.42.0
25
+ blinker==1.4
26
+ boto3==1.34.54
27
+ botocore==1.34.54
28
+ cachecontrol==0.12.10
29
+ cachetools==5.3.0
30
+ cachy==0.3.0
31
+ certifi==2022.12.7
32
+ cffi==1.15.1
33
+ chardet==4.0.0
34
+ charset-normalizer==2.1.1
35
+ cleo==0.8.1
36
+ click==8.0.3
37
+ clikit==0.6.2
38
+ cloud-init==23.4.4
39
+ cmake==3.25.0
40
+ colorama==0.4.4
41
+ colt5-attention==0.10.19
42
+ command-not-found==0.3
43
+ configobj==5.0.6
44
+ constantly==15.1.0
45
+ contourpy==1.0.7
46
+ crashtest==0.3.1
47
+ cryptography==3.4.8
48
+ cycler==0.11.0
49
+ cython==3.0.8
50
+ dataclasses-json==0.5.7
51
+ datasets==2.17.1
52
+ dbus-python==1.2.18
53
+ deeplake==3.2.11
54
+ dill==0.3.6
55
+ distlib==0.3.6
56
+ distro-info==1.0
57
+ distro==1.7.0
58
+ dm-tree==0.1.8
59
+ docker-pycreds==0.4.0
60
+ docutils==0.16
61
+ einops-exts==0.0.4
62
+ einops==0.7.0
63
+ einx==0.1.3
64
+ entrypoints==0.4
65
+ evaluate==0.4.1
66
+ exceptiongroup==1.2.0
67
+ fairscale==0.4.13
68
+ fairseq==0.12.2
69
+ fastapi==0.95.0
70
+ ffmpy==0.3.0
71
+ filelock==3.9.0
72
+ fire==0.5.0
73
+ flake8==7.0.0
74
+ flatbuffers==23.5.26
75
+ flexgen==0.1.5
76
+ fonttools==4.39.2
77
+ frozendict==2.4.0
78
+ frozenlist==1.3.3
79
+ fsspec==2023.10.0
80
+ gast==0.5.4
81
+ gensim==4.3.1
82
+ gitdb==4.0.10
83
+ gitpython==3.1.31
84
+ google-auth-oauthlib==1.0.0
85
+ google-auth==2.16.2
86
+ google-pasta==0.2.0
87
+ gradio==3.22.1
88
+ greenlet==2.0.2
89
+ grpcio==1.51.3
90
+ h11==0.14.0
91
+ h5py==3.10.0
92
+ html5lib==1.1
93
+ httpcore==0.16.3
94
+ httplib2==0.20.2
95
+ httpx==0.23.3
96
+ hub==3.0.1
97
+ huggingface-hub==0.21.3
98
+ humbug==0.2.8
99
+ hydra-core==1.0.7
100
+ hyperlink==21.0.0
101
+ idna==3.4
102
+ importlib-metadata==4.6.4
103
+ incremental==21.3.0
104
+ iniconfig==2.0.0
105
+ jax==0.4.25
106
+ jaxlib==0.4.25
107
+ jeepney==0.7.1
108
+ jieba==0.42.1
109
+ jinja2==3.1.2
110
+ jmespath==1.0.1
111
+ joblib==1.2.0
112
+ jsonpatch==1.32
113
+ jsonpointer==2.0
114
+ jsonschema==3.2.0
115
+ keras==2.14.0
116
+ keyring==23.5.0
117
+ kiwisolver==1.4.4
118
+ langchain==0.0.95
119
+ launchpadlib==1.10.16
120
+ lazr.restfulclient==0.14.4
121
+ lazr.uri==1.0.6
122
+ libclang==16.0.6
123
+ libcst==1.1.0
124
+ linkify-it-py==2.0.0
125
+ lion-pytorch==0.1.2
126
+ lit==15.0.7
127
+ llama==0.0.0
128
+ local-attention==1.9.0
129
+ lockfile==0.12.2
130
+ loguru==0.6.0
131
+ loralib==0.1.1
132
+ lxml==5.1.0
133
+ markdown-it-py==2.2.0
134
+ markdown==3.4.1
135
+ markupsafe==2.1.2
136
+ marshmallow-enum==1.5.1
137
+ marshmallow==3.19.0
138
+ matplotlib==3.7.1
139
+ mccabe==0.7.0
140
+ mdit-py-plugins==0.3.3
141
+ mdurl==0.1.2
142
+ ml-dtypes==0.2.0
143
+ more-itertools==8.10.0
144
+ mpmath==1.2.1
145
+ msgpack==1.0.3
146
+ multidict==6.0.4
147
+ multiprocess==0.70.14
148
+ mypy-extensions==1.0.0
149
+ namex==0.0.7
150
+ netifaces==0.11.0
151
+ networkx==3.0
152
+ ninja==1.11.1
153
+ numcodecs==0.11.0
154
+ numexpr==2.9.0
155
+ numpy==1.24.1
156
+ nvidia-cublas-cu11==11.11.3.6
157
+ nvidia-cublas-cu12==12.1.3.1
158
+ nvidia-cuda-cupti-cu11==11.8.87
159
+ nvidia-cuda-cupti-cu12==12.1.105
160
+ nvidia-cuda-nvcc-cu11==11.8.89
161
+ nvidia-cuda-nvcc-cu12==12.3.107
162
+ nvidia-cuda-nvrtc-cu11==11.7.99
163
+ nvidia-cuda-nvrtc-cu12==12.1.105
164
+ nvidia-cuda-runtime-cu11==11.8.89
165
+ nvidia-cuda-runtime-cu12==12.1.105
166
+ nvidia-cudnn-cu11==8.7.0.84
167
+ nvidia-cudnn-cu12==8.9.2.26
168
+ nvidia-cufft-cu11==10.9.0.58
169
+ nvidia-cufft-cu12==11.0.2.54
170
+ nvidia-curand-cu11==10.3.0.86
171
+ nvidia-curand-cu12==10.3.2.106
172
+ nvidia-cusolver-cu11==11.4.1.48
173
+ nvidia-cusolver-cu12==11.4.5.107
174
+ nvidia-cusparse-cu11==11.7.5.86
175
+ nvidia-cusparse-cu12==12.1.0.106
176
+ nvidia-ml-py==11.525.84
177
+ nvidia-nccl-cu11==2.16.5
178
+ nvidia-nccl-cu12==2.19.3
179
+ nvidia-nvjitlink-cu12==12.3.101
180
+ nvidia-nvtx-cu12==12.1.105
181
+ nvitop==1.0.0
182
+ oauthlib==3.2.0
183
+ omegaconf==2.0.6
184
+ opt-einsum==3.3.0
185
+ orjson==3.8.7
186
+ packaging==23.0
187
+ pandas==1.5.3
188
+ pastel==0.2.1
189
+ pathos==0.3.0
190
+ pathtools==0.1.2
191
+ peft==0.3.0.dev0
192
+ pendulum==3.0.0
193
+ pexpect==4.8.0
194
+ pillow==9.3.0
195
+ pip==24.0
196
+ pipenv==2023.2.18
197
+ pkginfo==1.8.2
198
+ platformdirs==3.0.0
199
+ pluggy==1.4.0
200
+ poetry-core==1.0.7
201
+ poetry==1.1.12
202
+ portalocker==2.8.2
203
+ pox==0.3.2
204
+ ppft==1.7.6.6
205
+ prettytable==3.10.0
206
+ protobuf==4.22.1
207
+ psutil==5.9.4
208
+ ptyprocess==0.7.0
209
+ pulp==2.7.0
210
+ pyarrow-hotfix==0.6
211
+ pyarrow==15.0.0
212
+ pyasn1-modules==0.2.1
213
+ pyasn1==0.4.8
214
+ pycairo==1.20.1
215
+ pycares==4.3.0
216
+ pycodestyle==2.11.1
217
+ pycparser==2.21
218
+ pydantic==1.10.5
219
+ pydub==0.25.1
220
+ pyflakes==3.2.0
221
+ pygments==2.17.2
222
+ pygobject==3.42.1
223
+ pyhamcrest==2.0.2
224
+ pyjwt==2.3.0
225
+ pylev==1.2.0
226
+ pyopenssl==21.0.0
227
+ pyparsing==2.4.7
228
+ pyrsistent==0.18.1
229
+ pyserial==3.5
230
+ pytest==7.4.2
231
+ python-apt==2.4.0+ubuntu3
232
+ python-dateutil==2.8.2
233
+ python-dotenv==1.0.1
234
+ python-magic==0.4.24
235
+ python-multipart==0.0.6
236
+ pytz==2022.1
237
+ pyyaml==6.0
238
+ quant-cuda==0.0.0
239
+ regex==2022.10.31
240
+ requests-oauthlib==1.3.1
241
+ requests-toolbelt==0.9.1
242
+ requests==2.28.1
243
+ responses==0.18.0
244
+ rfc3986==1.5.0
245
+ rich==13.7.0
246
+ rouge==1.0.1
247
+ rsa==4.7.2
248
+ s3transfer==0.10.0
249
+ sacrebleu==2.4.0
250
+ sacremoses==0.0.53
251
+ safetensors==0.4.2
252
+ scikit-learn==1.2.2
253
+ scipy==1.9.3
254
+ secretstorage==3.3.1
255
+ sentencepiece==0.1.99
256
+ sentry-sdk==1.17.0
257
+ service-identity==18.1.0
258
+ setproctitle==1.3.2
259
+ setuptools==67.4.0
260
+ shellingham==1.4.0
261
+ six==1.16.0
262
+ skypilot==0.4.1
263
+ smart-open==6.3.0
264
+ smmap==5.0.0
265
+ sniffio==1.3.0
266
+ sos==4.5.6
267
+ sqlalchemy==1.4.46
268
+ ssh-import-id==5.11
269
+ starlette==0.26.1
270
+ sympy==1.11.1
271
+ systemd-python==234
272
+ tabulate==0.9.0
273
+ tenacity==8.2.2
274
+ tensorboard-data-server==0.7.0
275
+ tensorboard-plugin-wit==1.8.1
276
+ tensorboard==2.14.1
277
+ tensorflow-estimator==2.14.0
278
+ tensorflow-io-gcs-filesystem==0.36.0
279
+ tensorflow==2.14.1
280
+ tensorrt-bindings==8.6.1
281
+ tensorrt-libs==8.6.1
282
+ tensorrt==8.6.1.post1
283
+ termcolor==2.2.0
284
+ text2vec==1.1.7
285
+ textgen==0.1.7
286
+ threadpoolctl==3.1.0
287
+ tiktoken==0.5.2
288
+ time-machine==2.13.0
289
+ timm==0.9.12
290
+ tokenizers==0.15.2
291
+ tokenmonster==1.1.12
292
+ tomli==2.0.1
293
+ tomlkit==0.9.2
294
+ toolz==0.12.0
295
+ torch==2.2.1
296
+ torchaudio==2.2.1
297
+ torchdiffeq==0.2.3
298
+ torchfix==0.4.0
299
+ torchvision==0.17.1
300
+ tqdm==4.66.1
301
+ transformers==4.36.2
302
+ triton==2.2.0
303
+ twisted==22.1.0
304
+ typing-extensions==4.10.0
305
+ typing-inspect==0.8.0
306
+ typing==3.7.4.3
307
+ tzdata==2024.1
308
+ ubuntu-advantage-tools==8001
309
+ uc-micro-py==1.0.1
310
+ ufw==0.36.1
311
+ unknown==0.0.0
312
+ urllib3==1.26.13
313
+ uvicorn==0.21.1
314
+ vector-quantize-pytorch==1.12.16
315
+ virtualenv-clone==0.5.7
316
+ virtualenv==20.20.0
317
+ wadllib==1.3.6
318
+ wandb==0.14.0
319
+ wcwidth==0.2.13
320
+ webencodings==0.5.1
321
+ websockets==10.4
322
+ werkzeug==2.2.3
323
+ wheel==0.37.1
324
+ wrapt==1.14.1
325
+ xdot==1.2
326
+ xxhash==3.2.0
327
+ yarl==1.8.2
328
+ zipp==1.0.0
329
+ zope.interface==5.4.0
wandb/run-20240305_145418-2p2hzuh2/files/wandb-metadata.json ADDED
@@ -0,0 +1,371 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.17.0-1024-oem-x86_64-with-glibc2.35",
3
+ "python": "3.10.12",
4
+ "heartbeatAt": "2024-03-05T14:54:19.885381",
5
+ "startedAt": "2024-03-05T14:54:18.638787",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [
9
+ "config/train_gpt2.py"
10
+ ],
11
+ "state": "running",
12
+ "program": "/home/ubuntu/sources/nanoGPT-BitNet158b/train.py",
13
+ "codePath": "train.py",
14
+ "git": {
15
+ "remote": "https://github.com/nindanaoto/nanoGPT-BitNet158b.git",
16
+ "commit": "706ef44f9dc1e7023ce6e342a72d237d43b09d7c"
17
+ },
18
+ "email": "matsuoka.kotaro@gmail.com",
19
+ "root": "/home/ubuntu/sources/nanoGPT-BitNet158b",
20
+ "host": "kmatsuoka-dev",
21
+ "username": "ubuntu",
22
+ "executable": "/usr/bin/python3",
23
+ "cpu_count": 32,
24
+ "cpu_count_logical": 64,
25
+ "cpu_freq": {
26
+ "current": 1067.1877656249999,
27
+ "min": 800.0,
28
+ "max": 3200.0
29
+ },
30
+ "cpu_freq_per_core": [
31
+ {
32
+ "current": 800.0,
33
+ "min": 800.0,
34
+ "max": 3200.0
35
+ },
36
+ {
37
+ "current": 800.0,
38
+ "min": 800.0,
39
+ "max": 3200.0
40
+ },
41
+ {
42
+ "current": 800.0,
43
+ "min": 800.0,
44
+ "max": 3200.0
45
+ },
46
+ {
47
+ "current": 800.0,
48
+ "min": 800.0,
49
+ "max": 3200.0
50
+ },
51
+ {
52
+ "current": 800.0,
53
+ "min": 800.0,
54
+ "max": 3200.0
55
+ },
56
+ {
57
+ "current": 800.0,
58
+ "min": 800.0,
59
+ "max": 3200.0
60
+ },
61
+ {
62
+ "current": 800.0,
63
+ "min": 800.0,
64
+ "max": 3200.0
65
+ },
66
+ {
67
+ "current": 800.0,
68
+ "min": 800.0,
69
+ "max": 3200.0
70
+ },
71
+ {
72
+ "current": 800.0,
73
+ "min": 800.0,
74
+ "max": 3200.0
75
+ },
76
+ {
77
+ "current": 800.0,
78
+ "min": 800.0,
79
+ "max": 3200.0
80
+ },
81
+ {
82
+ "current": 800.0,
83
+ "min": 800.0,
84
+ "max": 3200.0
85
+ },
86
+ {
87
+ "current": 800.0,
88
+ "min": 800.0,
89
+ "max": 3200.0
90
+ },
91
+ {
92
+ "current": 800.0,
93
+ "min": 800.0,
94
+ "max": 3200.0
95
+ },
96
+ {
97
+ "current": 800.0,
98
+ "min": 800.0,
99
+ "max": 3200.0
100
+ },
101
+ {
102
+ "current": 3200.0,
103
+ "min": 800.0,
104
+ "max": 3200.0
105
+ },
106
+ {
107
+ "current": 800.0,
108
+ "min": 800.0,
109
+ "max": 3200.0
110
+ },
111
+ {
112
+ "current": 800.0,
113
+ "min": 800.0,
114
+ "max": 3200.0
115
+ },
116
+ {
117
+ "current": 800.0,
118
+ "min": 800.0,
119
+ "max": 3200.0
120
+ },
121
+ {
122
+ "current": 800.0,
123
+ "min": 800.0,
124
+ "max": 3200.0
125
+ },
126
+ {
127
+ "current": 800.0,
128
+ "min": 800.0,
129
+ "max": 3200.0
130
+ },
131
+ {
132
+ "current": 800.0,
133
+ "min": 800.0,
134
+ "max": 3200.0
135
+ },
136
+ {
137
+ "current": 800.0,
138
+ "min": 800.0,
139
+ "max": 3200.0
140
+ },
141
+ {
142
+ "current": 800.01,
143
+ "min": 800.0,
144
+ "max": 3200.0
145
+ },
146
+ {
147
+ "current": 800.0,
148
+ "min": 800.0,
149
+ "max": 3200.0
150
+ },
151
+ {
152
+ "current": 800.0,
153
+ "min": 800.0,
154
+ "max": 3200.0
155
+ },
156
+ {
157
+ "current": 800.0,
158
+ "min": 800.0,
159
+ "max": 3200.0
160
+ },
161
+ {
162
+ "current": 3200.0,
163
+ "min": 800.0,
164
+ "max": 3200.0
165
+ },
166
+ {
167
+ "current": 2900.0,
168
+ "min": 800.0,
169
+ "max": 3200.0
170
+ },
171
+ {
172
+ "current": 800.0,
173
+ "min": 800.0,
174
+ "max": 3200.0
175
+ },
176
+ {
177
+ "current": 800.0,
178
+ "min": 800.0,
179
+ "max": 3200.0
180
+ },
181
+ {
182
+ "current": 800.0,
183
+ "min": 800.0,
184
+ "max": 3200.0
185
+ },
186
+ {
187
+ "current": 800.0,
188
+ "min": 800.0,
189
+ "max": 3200.0
190
+ },
191
+ {
192
+ "current": 800.0,
193
+ "min": 800.0,
194
+ "max": 3200.0
195
+ },
196
+ {
197
+ "current": 3135.471,
198
+ "min": 800.0,
199
+ "max": 3200.0
200
+ },
201
+ {
202
+ "current": 800.0,
203
+ "min": 800.0,
204
+ "max": 3200.0
205
+ },
206
+ {
207
+ "current": 800.0,
208
+ "min": 800.0,
209
+ "max": 3200.0
210
+ },
211
+ {
212
+ "current": 800.0,
213
+ "min": 800.0,
214
+ "max": 3200.0
215
+ },
216
+ {
217
+ "current": 800.0,
218
+ "min": 800.0,
219
+ "max": 3200.0
220
+ },
221
+ {
222
+ "current": 800.0,
223
+ "min": 800.0,
224
+ "max": 3200.0
225
+ },
226
+ {
227
+ "current": 3200.0,
228
+ "min": 800.0,
229
+ "max": 3200.0
230
+ },
231
+ {
232
+ "current": 800.0,
233
+ "min": 800.0,
234
+ "max": 3200.0
235
+ },
236
+ {
237
+ "current": 800.0,
238
+ "min": 800.0,
239
+ "max": 3200.0
240
+ },
241
+ {
242
+ "current": 800.0,
243
+ "min": 800.0,
244
+ "max": 3200.0
245
+ },
246
+ {
247
+ "current": 800.0,
248
+ "min": 800.0,
249
+ "max": 3200.0
250
+ },
251
+ {
252
+ "current": 800.0,
253
+ "min": 800.0,
254
+ "max": 3200.0
255
+ },
256
+ {
257
+ "current": 800.0,
258
+ "min": 800.0,
259
+ "max": 3200.0
260
+ },
261
+ {
262
+ "current": 800.0,
263
+ "min": 800.0,
264
+ "max": 3200.0
265
+ },
266
+ {
267
+ "current": 800.0,
268
+ "min": 800.0,
269
+ "max": 3200.0
270
+ },
271
+ {
272
+ "current": 800.0,
273
+ "min": 800.0,
274
+ "max": 3200.0
275
+ },
276
+ {
277
+ "current": 1000.0,
278
+ "min": 800.0,
279
+ "max": 3200.0
280
+ },
281
+ {
282
+ "current": 800.0,
283
+ "min": 800.0,
284
+ "max": 3200.0
285
+ },
286
+ {
287
+ "current": 800.0,
288
+ "min": 800.0,
289
+ "max": 3200.0
290
+ },
291
+ {
292
+ "current": 800.0,
293
+ "min": 800.0,
294
+ "max": 3200.0
295
+ },
296
+ {
297
+ "current": 800.0,
298
+ "min": 800.0,
299
+ "max": 3200.0
300
+ },
301
+ {
302
+ "current": 800.0,
303
+ "min": 800.0,
304
+ "max": 3200.0
305
+ },
306
+ {
307
+ "current": 800.0,
308
+ "min": 800.0,
309
+ "max": 3200.0
310
+ },
311
+ {
312
+ "current": 800.0,
313
+ "min": 800.0,
314
+ "max": 3200.0
315
+ },
316
+ {
317
+ "current": 800.0,
318
+ "min": 800.0,
319
+ "max": 3200.0
320
+ },
321
+ {
322
+ "current": 2100.0,
323
+ "min": 800.0,
324
+ "max": 3200.0
325
+ },
326
+ {
327
+ "current": 800.0,
328
+ "min": 800.0,
329
+ "max": 3200.0
330
+ },
331
+ {
332
+ "current": 3200.0,
333
+ "min": 800.0,
334
+ "max": 3200.0
335
+ },
336
+ {
337
+ "current": 800.0,
338
+ "min": 800.0,
339
+ "max": 3200.0
340
+ },
341
+ {
342
+ "current": 800.0,
343
+ "min": 800.0,
344
+ "max": 3200.0
345
+ },
346
+ {
347
+ "current": 800.0,
348
+ "min": 800.0,
349
+ "max": 3200.0
350
+ }
351
+ ],
352
+ "disk": {
353
+ "total": 489.8067626953125,
354
+ "used": 437.306884765625
355
+ },
356
+ "gpu": "NVIDIA A100-PCIE-40GB",
357
+ "gpu_count": 2,
358
+ "gpu_devices": [
359
+ {
360
+ "name": "NVIDIA A100-PCIE-40GB",
361
+ "memory_total": 42949672960
362
+ },
363
+ {
364
+ "name": "NVIDIA A100-PCIE-40GB",
365
+ "memory_total": 42949672960
366
+ }
367
+ ],
368
+ "memory": {
369
+ "total": 125.55450057983398
370
+ }
371
+ }
wandb/run-20240305_145418-2p2hzuh2/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"iter": 181000, "train/loss": 3.46370792388916, "val/loss": 3.463677406311035, "lr": 0.0004891606857702766, "mfu": 11.183942385348718, "_timestamp": 1710748872.7244864, "_runtime": 1098414.0720844269, "_step": 181, "_wandb": {"runtime": 1100396}}
wandb/run-20240305_145418-2p2hzuh2/logs/debug-internal.log ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22b2f669c19a47e7a4b129b8d4a4cff9a026cfe00bfcec3ce7cd471be9bc1e23
3
+ size 55714008
wandb/run-20240305_145418-2p2hzuh2/logs/debug.log ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e65cb3cae2d5b736b25eec25f8498eb23acf6258c7685d33598087017854683
3
+ size 3988
wandb/run-20240305_145418-2p2hzuh2/run-2p2hzuh2.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea004efe6ba51903e0b0ce054e474144cf91b6c8bf1bf90b517224cf34c73655
3
+ size 91548136