Sayan01 commited on
Commit
c5b2510
·
1 Parent(s): d841fc0

Training in progress, epoch 1

Browse files
config.json CHANGED
@@ -9,14 +9,16 @@
9
  "hidden_dropout_prob": 0.1,
10
  "hidden_size": 128,
11
  "id2label": {
12
- "0": "negative",
13
- "1": "positive"
 
14
  },
15
  "initializer_range": 0.02,
16
  "intermediate_size": 512,
17
  "label2id": {
18
- "negative": "0",
19
- "positive": "1"
 
20
  },
21
  "layer_norm_eps": 1e-12,
22
  "max_position_embeddings": 512,
@@ -27,7 +29,7 @@
27
  "position_embedding_type": "absolute",
28
  "problem_type": "single_label_classification",
29
  "torch_dtype": "float32",
30
- "transformers_version": "4.20.0",
31
  "type_vocab_size": 2,
32
  "use_cache": true,
33
  "vocab_size": 30522
 
9
  "hidden_dropout_prob": 0.1,
10
  "hidden_size": 128,
11
  "id2label": {
12
+ "0": "entailment",
13
+ "1": "neutral",
14
+ "2": "contradiction"
15
  },
16
  "initializer_range": 0.02,
17
  "intermediate_size": 512,
18
  "label2id": {
19
+ "contradiction": "2",
20
+ "entailment": "0",
21
+ "neutral": "1"
22
  },
23
  "layer_norm_eps": 1e-12,
24
  "max_position_embeddings": 512,
 
29
  "position_embedding_type": "absolute",
30
  "problem_type": "single_label_classification",
31
  "torch_dtype": "float32",
32
+ "transformers_version": "4.20.1",
33
  "type_vocab_size": 2,
34
  "use_cache": true,
35
  "vocab_size": 30522
logs/1657126391.824885/events.out.tfevents.1657126391.9181eea70df4.83.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eb3fb53312f1106dd6d08d2c41eda34ecb26a3704c1816a9f39157a9d92297ec
3
+ size 5347
logs/events.out.tfevents.1657126391.9181eea70df4.83.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43366632709cf4ccf2756aebb70b03745096400d8e42fc8cf837d12b7f5c6f4f
3
+ size 4209
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1390aafac65df57c228d10a286e1e48a51f0637e3a2108a108fa2856fe14285a
3
- size 17561831
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:292ecb16953563d02a27f4c89213d2279fdd1704c71a3a9f78550cd0abcd88d8
3
+ size 17562343
tokenizer.json CHANGED
@@ -1,6 +1,11 @@
1
  {
2
  "version": "1.0",
3
- "truncation": null,
 
 
 
 
 
4
  "padding": null,
5
  "added_tokens": [
6
  {
 
1
  {
2
  "version": "1.0",
3
+ "truncation": {
4
+ "direction": "Right",
5
+ "max_length": 512,
6
+ "strategy": "LongestFirst",
7
+ "stride": 0
8
+ },
9
  "padding": null,
10
  "added_tokens": [
11
  {
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:67219588024bbeb74892c6bd1b2d4a156621886bb498a8b4e08d390482dedd8b
3
  size 3375
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3f8ad35752d3afa4198c7fc22b6ff1c93d132a3d7725389b484dab7f212445a
3
  size 3375