Upload 11 files
Browse files- config.json +1 -2
config.json
CHANGED
|
@@ -9,7 +9,6 @@
|
|
| 9 |
"hidden_size": 2048,
|
| 10 |
"initializer_range": 0.02,
|
| 11 |
"intermediate_size": 11008,
|
| 12 |
-
"is_decoder": true,
|
| 13 |
"layer_types": [
|
| 14 |
"full_attention",
|
| 15 |
"full_attention",
|
|
@@ -59,7 +58,7 @@
|
|
| 59 |
"rope_theta": 1000000.0,
|
| 60 |
"sliding_window": null,
|
| 61 |
"tie_word_embeddings": true,
|
| 62 |
-
"torch_dtype": "
|
| 63 |
"transformers_version": "4.55.4",
|
| 64 |
"use_cache": true,
|
| 65 |
"use_sliding_window": false,
|
|
|
|
| 9 |
"hidden_size": 2048,
|
| 10 |
"initializer_range": 0.02,
|
| 11 |
"intermediate_size": 11008,
|
|
|
|
| 12 |
"layer_types": [
|
| 13 |
"full_attention",
|
| 14 |
"full_attention",
|
|
|
|
| 58 |
"rope_theta": 1000000.0,
|
| 59 |
"sliding_window": null,
|
| 60 |
"tie_word_embeddings": true,
|
| 61 |
+
"torch_dtype": "float32",
|
| 62 |
"transformers_version": "4.55.4",
|
| 63 |
"use_cache": true,
|
| 64 |
"use_sliding_window": false,
|