m4faisal commited on
Commit
268de79
·
verified ·
1 Parent(s): 9c67ca0

Training in progress, epoch 1

Browse files
Files changed (3) hide show
  1. config.json +5 -5
  2. model.safetensors +2 -2
  3. training_args.bin +2 -2
config.json CHANGED
@@ -1,14 +1,14 @@
1
  {
2
  "_name_or_path": "microsoft/deberta-v3-xsmall",
3
  "architectures": [
4
- "DebertaForSequenceClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
  "hidden_act": "gelu",
8
  "hidden_dropout_prob": 0.1,
9
  "hidden_size": 384,
10
  "id2label": {
11
- "0": "NEGATIVE",
12
  "1": "INSIGNIFICANT SIMILARITY",
13
  "2": "LITTLE SIMILARITY",
14
  "3": "MEDIUM SIMILARITY",
@@ -21,14 +21,14 @@
21
  "INSIGNIFICANT SIMILARITY": 1,
22
  "LITTLE SIMILARITY": 2,
23
  "MEDIUM SIMILARITY": 3,
24
- "NEGATIVE": 0,
25
  "POSITIVE": 5,
26
  "VERY SIMILAR": 4
27
  },
28
  "layer_norm_eps": 1e-07,
29
  "max_position_embeddings": 512,
30
  "max_relative_positions": -1,
31
- "model_type": "deberta",
32
  "norm_rel_ebd": "layer_norm",
33
  "num_attention_heads": 6,
34
  "num_hidden_layers": 12,
@@ -45,7 +45,7 @@
45
  "relative_attention": true,
46
  "share_att_key": true,
47
  "torch_dtype": "float32",
48
- "transformers_version": "4.35.2",
49
  "type_vocab_size": 0,
50
  "vocab_size": 128100
51
  }
 
1
  {
2
  "_name_or_path": "microsoft/deberta-v3-xsmall",
3
  "architectures": [
4
+ "DebertaV2ForSequenceClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
  "hidden_act": "gelu",
8
  "hidden_dropout_prob": 0.1,
9
  "hidden_size": 384,
10
  "id2label": {
11
+ "0": "NO SIMILARITY",
12
  "1": "INSIGNIFICANT SIMILARITY",
13
  "2": "LITTLE SIMILARITY",
14
  "3": "MEDIUM SIMILARITY",
 
21
  "INSIGNIFICANT SIMILARITY": 1,
22
  "LITTLE SIMILARITY": 2,
23
  "MEDIUM SIMILARITY": 3,
24
+ "NO SIMILARITY": 0,
25
  "POSITIVE": 5,
26
  "VERY SIMILAR": 4
27
  },
28
  "layer_norm_eps": 1e-07,
29
  "max_position_embeddings": 512,
30
  "max_relative_positions": -1,
31
+ "model_type": "deberta-v2",
32
  "norm_rel_ebd": "layer_norm",
33
  "num_attention_heads": 6,
34
  "num_hidden_layers": 12,
 
45
  "relative_attention": true,
46
  "share_att_key": true,
47
  "torch_dtype": "float32",
48
+ "transformers_version": "4.37.2",
49
  "type_vocab_size": 0,
50
  "vocab_size": 128100
51
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:07e56fbd175c6243450e320705d041cf9b461853f85c960446d9821629ae1c61
3
- size 298292320
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0c06369378cd2f70a286813135374a53bb5702cce063f7029710670b291fdcc
3
+ size 283353600
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3b956960426f662b27e2ee745af6825a1dc305971cb2907dc3780f733231ba25
3
- size 4536
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb081b39a3482f148898c756a67b9bbe6398c1d9c2f5bef87d0098c5eca7f7f2
3
+ size 4664