tijesu24 commited on
Commit
c44bc7f
·
verified ·
1 Parent(s): 041a154

Training in progress, epoch 1

Browse files
config.json CHANGED
@@ -7,24 +7,20 @@
7
  "hidden_dropout_prob": 0.1,
8
  "hidden_size": 768,
9
  "id2label": {
10
- "0": "O",
11
- "1": "B-HEADER",
12
- "2": "I-HEADER",
13
- "3": "B-QUESTION",
14
- "4": "I-QUESTION",
15
- "5": "B-ANSWER",
16
- "6": "I-ANSWER"
17
  },
18
  "initializer_range": 0.02,
19
  "intermediate_size": 3072,
20
  "label2id": {
21
- "B-ANSWER": 5,
22
- "B-HEADER": 1,
23
- "B-QUESTION": 3,
24
- "I-ANSWER": 6,
25
- "I-HEADER": 2,
26
- "I-QUESTION": 4,
27
- "O": 0
28
  },
29
  "layer_norm_eps": 1e-12,
30
  "max_2d_position_embeddings": 1024,
 
7
  "hidden_dropout_prob": 0.1,
8
  "hidden_size": 768,
9
  "id2label": {
10
+ "0": "S-TOTAL",
11
+ "1": "S-DATE",
12
+ "2": "S-ADDRESS",
13
+ "3": "S-COMPANY",
14
+ "4": "O"
 
 
15
  },
16
  "initializer_range": 0.02,
17
  "intermediate_size": 3072,
18
  "label2id": {
19
+ "O": 4,
20
+ "S-ADDRESS": 2,
21
+ "S-COMPANY": 3,
22
+ "S-DATE": 1,
23
+ "S-TOTAL": 0
 
 
24
  },
25
  "layer_norm_eps": 1e-12,
26
  "max_2d_position_embeddings": 1024,
logs/events.out.tfevents.1749837810.5b536cfb03a6.35.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d7ad13c6b43340cc692244ccf25e4796ff472a49e76a38907c3371950de9c6c1
3
+ size 5811
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7257e65d40fc0b3e8220effddb5a708751420552ec19178f9a7718daef5d8d6f
3
- size 450558212
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ad54b8a72db534be23a5f245860de3a70d7d30f79f3307366b3790f58eadfd53
3
+ size 450552060