Update README.md
Browse files
README.md
CHANGED
|
@@ -2,4 +2,6 @@
|
|
| 2 |
tags:
|
| 3 |
- llama
|
| 4 |
---
|
| 5 |
-
252M custom transformer architecture with two transformations from the llama2-70b embeddings to 1024-dimensional from 8192-d and back from 1024-d to 8192-d for the llama2-70b language modelling head.
|
|
|
|
|
|
|
|
|
| 2 |
tags:
|
| 3 |
- llama
|
| 4 |
---
|
| 5 |
+
A randomly initialized checkpoint of a 252M custom transformer architecture with two linear transformations from the llama2-70b embeddings to 1024-dimensional space from 8192-d and then back from 1024-d to 8192-d for the llama2-70b language modelling head.
|
| 6 |
+
|
| 7 |
+
To be trained
|