lekhnathrijal commited on
Commit
b506157
·
verified ·
1 Parent(s): fc939d3

Training in progress, step 100

Browse files
Files changed (4) hide show
  1. config.json +6 -24
  2. model.safetensors +2 -2
  3. tokenizer.json +1 -10
  4. training_args.bin +2 -2
config.json CHANGED
@@ -1,32 +1,14 @@
1
  {
2
- "_name_or_path": "google-bert/bert-large-uncased",
3
  "architectures": [
4
  "BertForSequenceClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
  "classifier_dropout": null,
8
- "custom_pipelines": {
9
- "question-classifier": {
10
- "default": {
11
- "model": {
12
- "pt": [
13
- "ai-research-lab/bert-question-classifier",
14
- "main"
15
- ]
16
- }
17
- },
18
- "impl": "classifier_pipeline.MultiTaskClassifierPipeline",
19
- "pt": [
20
- "AutoModelForSequenceClassification"
21
- ],
22
- "tf": [],
23
- "type": "text"
24
- }
25
- },
26
  "gradient_checkpointing": false,
27
  "hidden_act": "gelu",
28
  "hidden_dropout_prob": 0.1,
29
- "hidden_size": 1024,
30
  "id2label": {
31
  "0": "d",
32
  "1": "y",
@@ -64,7 +46,7 @@
64
  "33": "positive_perception"
65
  },
66
  "initializer_range": 0.02,
67
- "intermediate_size": 4096,
68
  "label2id": {
69
  "accumulated wealth": 6,
70
  "business": 22,
@@ -104,8 +86,8 @@
104
  "layer_norm_eps": 1e-12,
105
  "max_position_embeddings": 512,
106
  "model_type": "bert",
107
- "num_attention_heads": 16,
108
- "num_hidden_layers": 24,
109
  "pad_token_id": 0,
110
  "position_embedding_type": "absolute",
111
  "task_specific_params": {
@@ -286,7 +268,7 @@
286
  }
287
  },
288
  "torch_dtype": "float32",
289
- "transformers_version": "4.48.1",
290
  "type_vocab_size": 2,
291
  "use_cache": true,
292
  "vocab_size": 30522
 
1
  {
2
+ "_name_or_path": "google-bert/bert-base-uncased",
3
  "architectures": [
4
  "BertForSequenceClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
  "classifier_dropout": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8
  "gradient_checkpointing": false,
9
  "hidden_act": "gelu",
10
  "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
  "id2label": {
13
  "0": "d",
14
  "1": "y",
 
46
  "33": "positive_perception"
47
  },
48
  "initializer_range": 0.02,
49
+ "intermediate_size": 3072,
50
  "label2id": {
51
  "accumulated wealth": 6,
52
  "business": 22,
 
86
  "layer_norm_eps": 1e-12,
87
  "max_position_embeddings": 512,
88
  "model_type": "bert",
89
+ "num_attention_heads": 12,
90
+ "num_hidden_layers": 12,
91
  "pad_token_id": 0,
92
  "position_embedding_type": "absolute",
93
  "task_specific_params": {
 
268
  }
269
  },
270
  "torch_dtype": "float32",
271
+ "transformers_version": "4.49.0",
272
  "type_vocab_size": 2,
273
  "use_cache": true,
274
  "vocab_size": 30522
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:38841421c2a4c9d481c6a1c0af0e8323026294db894c62d366e5832b2bc8b19e
3
- size 1340753968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e01f75989b155d28668fcf4ab110ad6780b7b2783c7f4c5775859de89ab21fa
3
+ size 438057080
tokenizer.json CHANGED
@@ -6,16 +6,7 @@
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
9
- "padding": {
10
- "strategy": {
11
- "Fixed": 512
12
- },
13
- "direction": "Right",
14
- "pad_to_multiple_of": null,
15
- "pad_id": 0,
16
- "pad_type_id": 0,
17
- "pad_token": "[PAD]"
18
- },
19
  "added_tokens": [
20
  {
21
  "id": 0,
 
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
9
+ "padding": null,
 
 
 
 
 
 
 
 
 
10
  "added_tokens": [
11
  {
12
  "id": 0,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:24441a5c5eac7a327d27fb0efacbb46534247f26598002cfcbf4bfb03aefbcdd
3
- size 5368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b00eb55177b98f1dc327e038c02b62a70d3904e2de165d2c342edbb856051d73
3
+ size 5432