Commit
·
5ef77ef
1
Parent(s):
c41323c
config change
Browse files
__pycache__/configuration_minitransformer.cpython-312.pyc
CHANGED
|
Binary files a/__pycache__/configuration_minitransformer.cpython-312.pyc and b/__pycache__/configuration_minitransformer.cpython-312.pyc differ
|
|
|
config.json
CHANGED
|
@@ -2,7 +2,7 @@
|
|
| 2 |
"model_type": "minitransformer",
|
| 3 |
"_name_or_path": "Transformer_500M",
|
| 4 |
"architectures": ["MiniTransformer"],
|
| 5 |
-
"n_embd":
|
| 6 |
"n_heads": 8,
|
| 7 |
"n_layers": 25,
|
| 8 |
"seq_len": 8192,
|
|
|
|
| 2 |
"model_type": "minitransformer",
|
| 3 |
"_name_or_path": "Transformer_500M",
|
| 4 |
"architectures": ["MiniTransformer"],
|
| 5 |
+
"n_embd": 768,
|
| 6 |
"n_heads": 8,
|
| 7 |
"n_layers": 25,
|
| 8 |
"seq_len": 8192,
|
configuration_minitransformer.py
CHANGED
|
@@ -7,7 +7,7 @@ class MiniTransformerConfig(PretrainedConfig):
|
|
| 7 |
def __init__(
|
| 8 |
self,
|
| 9 |
bsz: int = 1,
|
| 10 |
-
n_embd: int =
|
| 11 |
n_heads: int = 8,
|
| 12 |
n_layers: int = 25,
|
| 13 |
seq_len: int = 8192,
|
|
|
|
| 7 |
def __init__(
|
| 8 |
self,
|
| 9 |
bsz: int = 1,
|
| 10 |
+
n_embd: int = 768,
|
| 11 |
n_heads: int = 8,
|
| 12 |
n_layers: int = 25,
|
| 13 |
seq_len: int = 8192,
|