adjust readme
Browse files
README.md
CHANGED
|
@@ -4,7 +4,7 @@ license: mit
|
|
| 4 |
|
| 5 |
# einygpt
|
| 6 |
|
| 7 |
-
Here's the models I've trained
|
| 8 |
|
| 9 |
- [a multihead attention model](./model_weights_mha.pth) replicating the model discussed in the [TinyStories paper](https://arxiv.org/abs/2305.07759) using the GPT2Tokenizer
|
| 10 |
- [a multiquery attention model](model_weights_mqa.pth) using the GPT2Tokenizer
|
|
|
|
| 4 |
|
| 5 |
# einygpt
|
| 6 |
|
| 7 |
+
Here's the models I've trained using the transformer I wrote in [einygpt](https://github.com/clankur/einygpt). For reference they are:
|
| 8 |
|
| 9 |
- [a multihead attention model](./model_weights_mha.pth) replicating the model discussed in the [TinyStories paper](https://arxiv.org/abs/2305.07759) using the GPT2Tokenizer
|
| 10 |
- [a multiquery attention model](model_weights_mqa.pth) using the GPT2Tokenizer
|