qgallouedec HF Staff commited on
Commit
484d22f
·
verified ·
1 Parent(s): ac355da

Upload Qwen3ForSequenceClassification

Browse files
Files changed (2) hide show
  1. config.json +47 -11
  2. model.safetensors +2 -2
config.json CHANGED
@@ -4,35 +4,71 @@
4
  ],
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
 
 
 
7
  "head_dim": 128,
8
  "hidden_act": "silu",
9
- "hidden_size": 8,
10
  "id2label": {
11
  "0": "LABEL_0"
12
  },
13
  "initializer_range": 0.02,
14
- "intermediate_size": 32,
15
  "label2id": {
16
  "LABEL_0": 0
17
  },
18
  "layer_types": [
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
19
  "full_attention",
20
  "full_attention"
21
  ],
22
- "max_position_embeddings": 32768,
23
- "max_window_layers": 28,
24
  "model_type": "qwen3",
25
- "num_attention_heads": 4,
26
  "num_hidden_layers": 2,
27
- "num_key_value_heads": 2,
28
  "rms_norm_eps": 1e-06,
29
  "rope_scaling": null,
30
- "rope_theta": 10000.0,
31
  "sliding_window": null,
32
- "tie_word_embeddings": false,
33
- "torch_dtype": "float32",
34
- "transformers_version": "4.55.0.dev0",
35
  "use_cache": true,
36
  "use_sliding_window": false,
37
- "vocab_size": 151669
38
  }
 
4
  ],
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
7
+ "bos_token_id": 151643,
8
+ "dtype": "bfloat16",
9
+ "eos_token_id": 151645,
10
  "head_dim": 128,
11
  "hidden_act": "silu",
12
+ "hidden_size": 2560,
13
  "id2label": {
14
  "0": "LABEL_0"
15
  },
16
  "initializer_range": 0.02,
17
+ "intermediate_size": 9728,
18
  "label2id": {
19
  "LABEL_0": 0
20
  },
21
  "layer_types": [
22
+ "full_attention",
23
+ "full_attention",
24
+ "full_attention",
25
+ "full_attention",
26
+ "full_attention",
27
+ "full_attention",
28
+ "full_attention",
29
+ "full_attention",
30
+ "full_attention",
31
+ "full_attention",
32
+ "full_attention",
33
+ "full_attention",
34
+ "full_attention",
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention",
43
+ "full_attention",
44
+ "full_attention",
45
+ "full_attention",
46
+ "full_attention",
47
+ "full_attention",
48
+ "full_attention",
49
+ "full_attention",
50
+ "full_attention",
51
+ "full_attention",
52
+ "full_attention",
53
+ "full_attention",
54
+ "full_attention",
55
+ "full_attention",
56
  "full_attention",
57
  "full_attention"
58
  ],
59
+ "max_position_embeddings": 40960,
60
+ "max_window_layers": 36,
61
  "model_type": "qwen3",
62
+ "num_attention_heads": 32,
63
  "num_hidden_layers": 2,
64
+ "num_key_value_heads": 8,
65
  "rms_norm_eps": 1e-06,
66
  "rope_scaling": null,
67
+ "rope_theta": 1000000,
68
  "sliding_window": null,
69
+ "tie_word_embeddings": true,
70
+ "transformers_version": "4.57.0.dev0",
 
71
  "use_cache": true,
72
  "use_sliding_window": false,
73
+ "vocab_size": 151936
74
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3a0ecdf32244a8a38b46dded58f9572d07fa3cce579d2fae5051611a9de48114
3
- size 4962680
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:caa8db406072b0d0b67115adb7fa4a1e46a4153d8caea9849eb413c3ce4e09e0
3
+ size 1181653752