Sayan01 commited on
Commit
390d0ef
·
1 Parent(s): 262ad1f

Training in progress, epoch 1

Browse files
config.json CHANGED
@@ -9,14 +9,14 @@
9
  "hidden_dropout_prob": 0.1,
10
  "hidden_size": 128,
11
  "id2label": {
12
- "0": "unacceptable",
13
- "1": "acceptable"
14
  },
15
  "initializer_range": 0.02,
16
  "intermediate_size": 512,
17
  "label2id": {
18
- "acceptable": "1",
19
- "unacceptable": "0"
20
  },
21
  "layer_norm_eps": 1e-12,
22
  "max_position_embeddings": 512,
 
9
  "hidden_dropout_prob": 0.1,
10
  "hidden_size": 128,
11
  "id2label": {
12
+ "0": "not_equivalent",
13
+ "1": "equivalent"
14
  },
15
  "initializer_range": 0.02,
16
  "intermediate_size": 512,
17
  "label2id": {
18
+ "equivalent": "1",
19
+ "not_equivalent": "0"
20
  },
21
  "layer_norm_eps": 1e-12,
22
  "max_position_embeddings": 512,
logs/1655981347.9312732/events.out.tfevents.1655981347.9fd42316f55c.73.27 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df437517363a4a87f0844dc84e802123e5b95adf2e6484e522a3fc5efc342af1
3
+ size 5347
logs/events.out.tfevents.1655980824.9fd42316f55c.73.24 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d6ee5c14d2c589691f92f997de100c906eb3f788d4bc4efec840225f21e2fba7
3
- size 7998
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b9666aaeedd05abd985e84ff2122002346df8253672f12cd4d0f232242abd4c
3
+ size 8832
logs/events.out.tfevents.1655981347.9fd42316f55c.73.26 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:315cf54eb1b261bbde121b4cdc48cb3003e7cb5401c6543b5ee0b67f485af6ef
3
+ size 4162
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a5c28f7c02ff537952aa17e411b908fe6b0227a4314ac68ca209b713d6654643
3
  size 17561831
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:290791aaed67eca7f54b57da745f473f94d04c6084ab0cbb6128fe71e548ae1c
3
  size 17561831
tokenizer.json CHANGED
@@ -1,6 +1,11 @@
1
  {
2
  "version": "1.0",
3
- "truncation": null,
 
 
 
 
 
4
  "padding": null,
5
  "added_tokens": [
6
  {
 
1
  {
2
  "version": "1.0",
3
+ "truncation": {
4
+ "direction": "Right",
5
+ "max_length": 512,
6
+ "strategy": "LongestFirst",
7
+ "stride": 0
8
+ },
9
  "padding": null,
10
  "added_tokens": [
11
  {