danielsaggau commited on
Commit
2def44b
·
1 Parent(s): e798ef1

Training in progress, epoch 1

Browse files
Files changed (6) hide show
  1. config.json +51 -13
  2. pytorch_model.bin +2 -2
  3. tokenizer.json +0 -0
  4. tokenizer_config.json +2 -2
  5. training_args.bin +1 -1
  6. vocab.txt +0 -0
config.json CHANGED
@@ -1,15 +1,32 @@
1
  {
2
- "_name_or_path": "nlpaueb/legal-bert-small-uncased",
3
  "architectures": [
4
- "BertForSequenceClassification"
5
  ],
 
6
  "attention_probs_dropout_prob": 0.1,
7
- "bos_token_id": 0,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8
  "classifier_dropout": null,
9
- "eos_token_ids": 0,
 
 
10
  "hidden_act": "gelu",
11
  "hidden_dropout_prob": 0.1,
12
- "hidden_size": 512,
13
  "id2label": {
14
  "0": "LABEL_0",
15
  "1": "LABEL_1",
@@ -20,13 +37,32 @@
20
  "6": "LABEL_6",
21
  "7": "LABEL_7",
22
  "8": "LABEL_8",
23
- "9": "LABEL_9"
 
 
 
 
 
 
 
 
 
24
  },
 
25
  "initializer_range": 0.02,
26
- "intermediate_size": 2048,
27
  "label2id": {
28
  "LABEL_0": 0,
29
  "LABEL_1": 1,
 
 
 
 
 
 
 
 
 
30
  "LABEL_2": 2,
31
  "LABEL_3": 3,
32
  "LABEL_4": 4,
@@ -36,15 +72,17 @@
36
  "LABEL_8": 8,
37
  "LABEL_9": 9
38
  },
39
- "layer_norm_eps": 1e-12,
40
- "max_position_embeddings": 512,
41
- "model_type": "bert",
42
- "num_attention_heads": 8,
43
- "num_hidden_layers": 6,
44
- "output_past": true,
 
45
  "pad_token_id": 0,
46
  "position_embedding_type": "absolute",
47
  "problem_type": "multi_label_classification",
 
48
  "torch_dtype": "float32",
49
  "transformers_version": "4.25.1",
50
  "type_vocab_size": 2,
 
1
  {
2
+ "_name_or_path": "/content/drive/MyDrive/bregman_mimic_40ksteps/40000",
3
  "architectures": [
4
+ "LongformerForSequenceClassification"
5
  ],
6
+ "attention_mode": "longformer",
7
  "attention_probs_dropout_prob": 0.1,
8
+ "attention_window": [
9
+ 128,
10
+ 128,
11
+ 128,
12
+ 128,
13
+ 128,
14
+ 128,
15
+ 128,
16
+ 128,
17
+ 128,
18
+ 128,
19
+ 128,
20
+ 128
21
+ ],
22
+ "bos_token_id": null,
23
  "classifier_dropout": null,
24
+ "cls_token_id": 101,
25
+ "eos_token_id": null,
26
+ "gradient_checkpointing": false,
27
  "hidden_act": "gelu",
28
  "hidden_dropout_prob": 0.1,
29
+ "hidden_size": 768,
30
  "id2label": {
31
  "0": "LABEL_0",
32
  "1": "LABEL_1",
 
37
  "6": "LABEL_6",
38
  "7": "LABEL_7",
39
  "8": "LABEL_8",
40
+ "9": "LABEL_9",
41
+ "10": "LABEL_10",
42
+ "11": "LABEL_11",
43
+ "12": "LABEL_12",
44
+ "13": "LABEL_13",
45
+ "14": "LABEL_14",
46
+ "15": "LABEL_15",
47
+ "16": "LABEL_16",
48
+ "17": "LABEL_17",
49
+ "18": "LABEL_18"
50
  },
51
+ "ignore_attention_mask": false,
52
  "initializer_range": 0.02,
53
+ "intermediate_size": 3072,
54
  "label2id": {
55
  "LABEL_0": 0,
56
  "LABEL_1": 1,
57
+ "LABEL_10": 10,
58
+ "LABEL_11": 11,
59
+ "LABEL_12": 12,
60
+ "LABEL_13": 13,
61
+ "LABEL_14": 14,
62
+ "LABEL_15": 15,
63
+ "LABEL_16": 16,
64
+ "LABEL_17": 17,
65
+ "LABEL_18": 18,
66
  "LABEL_2": 2,
67
  "LABEL_3": 3,
68
  "LABEL_4": 4,
 
72
  "LABEL_8": 8,
73
  "LABEL_9": 9
74
  },
75
+ "layer_norm_eps": 1e-05,
76
+ "max_position_embeddings": 4098,
77
+ "model_max_length": 4096,
78
+ "model_type": "longformer",
79
+ "num_attention_heads": 12,
80
+ "num_hidden_layers": 12,
81
+ "onnx_export": false,
82
  "pad_token_id": 0,
83
  "position_embedding_type": "absolute",
84
  "problem_type": "multi_label_classification",
85
+ "sep_token_id": 102,
86
  "torch_dtype": "float32",
87
  "transformers_version": "4.25.1",
88
  "type_vocab_size": 2,
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f4856891ee67c3e24b802f84aa683b6bdb2842c575911df58d034103cd880f7c
3
- size 140334005
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca163dfd7352c0d7a72223cf73e0dd091a8d737b31fb01eb946ad090ee2c920b
3
+ size 536507161
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -3,8 +3,8 @@
3
  "do_basic_tokenize": true,
4
  "do_lower_case": true,
5
  "mask_token": "[MASK]",
6
- "model_max_length": 512,
7
- "name_or_path": "nlpaueb/legal-bert-small-uncased",
8
  "never_split": null,
9
  "pad_token": "[PAD]",
10
  "sep_token": "[SEP]",
 
3
  "do_basic_tokenize": true,
4
  "do_lower_case": true,
5
  "mask_token": "[MASK]",
6
+ "model_max_length": 4096,
7
+ "name_or_path": "/content/drive/MyDrive/bregman_mimic_40ksteps/40000",
8
  "never_split": null,
9
  "pad_token": "[PAD]",
10
  "sep_token": "[SEP]",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ce7afc4e786f73c06e083fd2171b30cf6e23df894efb7f258d7f54a68b10cf67
3
  size 3451
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e703b12fd68fe2c571ef89b6bf1c1dbc86cb30bb6cf0298b0e32b7ee8b31ec66
3
  size 3451
vocab.txt CHANGED
The diff for this file is too large to render. See raw diff