Upload Qwen2ForSequenceClassification
Browse files- config.json +2 -2
- model.safetensors +2 -2
config.json
CHANGED
|
@@ -7,7 +7,7 @@
|
|
| 7 |
"dtype": "bfloat16",
|
| 8 |
"eos_token_id": 151645,
|
| 9 |
"hidden_act": "silu",
|
| 10 |
-
"hidden_size":
|
| 11 |
"id2label": {
|
| 12 |
"0": "LABEL_0"
|
| 13 |
},
|
|
@@ -85,7 +85,7 @@
|
|
| 85 |
"max_position_embeddings": 32768,
|
| 86 |
"max_window_layers": 70,
|
| 87 |
"model_type": "qwen2",
|
| 88 |
-
"num_attention_heads":
|
| 89 |
"num_hidden_layers": 2,
|
| 90 |
"num_key_value_heads": 8,
|
| 91 |
"rms_norm_eps": 1e-06,
|
|
|
|
| 7 |
"dtype": "bfloat16",
|
| 8 |
"eos_token_id": 151645,
|
| 9 |
"hidden_act": "silu",
|
| 10 |
+
"hidden_size": 16,
|
| 11 |
"id2label": {
|
| 12 |
"0": "LABEL_0"
|
| 13 |
},
|
|
|
|
| 85 |
"max_position_embeddings": 32768,
|
| 86 |
"max_window_layers": 70,
|
| 87 |
"model_type": "qwen2",
|
| 88 |
+
"num_attention_heads": 4,
|
| 89 |
"num_hidden_layers": 2,
|
| 90 |
"num_key_value_heads": 8,
|
| 91 |
"rms_norm_eps": 1e-06,
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:62b519ab581e46e03c0f935e7dc595fb513dbcabcf6e55c06895dd6d25821367
|
| 3 |
+
size 10183992
|