Ellie5757575757 commited on
Commit
d73663e
·
verified ·
1 Parent(s): 1d2b5c4

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +97 -64
config.json CHANGED
@@ -1,71 +1,104 @@
1
  {
2
- "architectures": [
3
- "BertForSequenceClassification"
4
- ],
5
- "attention_probs_dropout_prob": 0.1,
6
- "classifier_dropout": null,
7
- "gradient_checkpointing": false,
8
- "hidden_act": "gelu",
9
- "hidden_dropout_prob": 0.1,
10
- "hidden_size": 768,
11
- "initializer_range": 0.02,
12
- "intermediate_size": 3072,
13
- "layer_norm_eps": 1e-12,
14
- "max_position_embeddings": 512,
15
- "model_type": "bert",
16
- "num_attention_heads": 12,
17
- "num_hidden_layers": 12,
18
- "pad_token_id": 0,
19
- "position_embedding_type": "absolute",
20
- "transformers_version": "4.30.0",
21
- "type_vocab_size": 2,
22
- "use_cache": true,
23
- "vocab_size": 28996,
24
- "torch_dtype": "float32",
25
  "id2label": {
26
- "0": "Normal",
27
- "1": "Mild Aphasia",
28
- "2": "Moderate Aphasia",
29
- "3": "Severe Aphasia"
 
 
 
 
 
30
  },
 
31
  "label2id": {
32
- "Normal": 0,
33
- "Mild Aphasia": 1,
34
- "Moderate Aphasia": 2,
35
- "Severe Aphasia": 3
 
 
 
 
 
36
  },
37
- "num_labels": 4,
38
- "problem_type": "single_label_classification",
39
- "finetuning_task": "aphasia_classification",
40
- "_name_or_path": "dmis-lab/biobert-base-cased-v1.1",
41
- "base_model": "dmis-lab/biobert-base-cased-v1.1",
42
- "task_specific_params": {
43
- "aphasia_classification": {
44
- "max_length": 512,
45
- "do_lower_case": false,
46
- "early_stopping": true,
47
- "num_beams": 1
48
- }
 
 
 
49
  },
50
- "custom_config": {
51
- "model_version": "1.0",
52
- "training_data": "aphasia_corpus",
53
- "evaluation_metrics": {
54
- "accuracy": 0.85,
55
- "f1_score": 0.83,
56
- "precision": 0.84,
57
- "recall": 0.82
58
- },
59
- "feature_extraction": {
60
- "linguistic_features": true,
61
- "syntactic_complexity": true,
62
- "semantic_coherence": true,
63
- "discourse_markers": true
64
- },
65
- "preprocessing": {
66
- "cha_format": true,
67
- "json_structured": true,
68
- "text_normalization": true
69
- }
70
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
71
  }
 
1
  {
2
+ "model_name": "microsoft/BiomedNLP-PubMedBERT-base-uncased-abstract-fulltext",
3
+ "model_type": "StableAphasiaClassifier",
4
+ "num_labels": 9,
5
+ "architectures": ["StableAphasiaClassifier"],
6
+
7
+ "_name_or_path": "microsoft/BiomedNLP-PubMedBERT-base-uncased-abstract-fulltext",
8
+ "problem_type": "single_label_classification",
9
+
10
+ "aphasia_types_mapping": {
11
+ "BROCA": 0,
12
+ "TRANSMOTOR": 1,
13
+ "NOTAPHASICBYWAB": 2,
14
+ "CONDUCTION": 3,
15
+ "WERNICKE": 4,
16
+ "ANOMIC": 5,
17
+ "GLOBAL": 6,
18
+ "ISOLATION": 7,
19
+ "TRANSSENSORY": 8
20
+ },
21
+
 
 
 
22
  "id2label": {
23
+ "0": "BROCA",
24
+ "1": "TRANSMOTOR",
25
+ "2": "NOTAPHASICBYWAB",
26
+ "3": "CONDUCTION",
27
+ "4": "WERNICKE",
28
+ "5": "ANOMIC",
29
+ "6": "GLOBAL",
30
+ "7": "ISOLATION",
31
+ "8": "TRANSSENSORY"
32
  },
33
+
34
  "label2id": {
35
+ "BROCA": 0,
36
+ "TRANSMOTOR": 1,
37
+ "NOTAPHASICBYWAB": 2,
38
+ "CONDUCTION": 3,
39
+ "WERNICKE": 4,
40
+ "ANOMIC": 5,
41
+ "GLOBAL": 6,
42
+ "ISOLATION": 7,
43
+ "TRANSSENSORY": 8
44
  },
45
+
46
+ "model_config": {
47
+ "max_length": 512,
48
+ "hidden_size": 768,
49
+ "pos_vocab_size": 150,
50
+ "pos_emb_dim": 64,
51
+ "grammar_dim": 3,
52
+ "grammar_hidden_dim": 64,
53
+ "duration_hidden_dim": 128,
54
+ "prosody_dim": 32,
55
+ "num_attention_heads": 8,
56
+ "attention_dropout": 0.3,
57
+ "classifier_hidden_dims": [512, 256],
58
+ "dropout_rate": 0.3,
59
+ "activation_fn": "tanh"
60
  },
61
+
62
+ "training_config": {
63
+ "learning_rate": 0.0005,
64
+ "weight_decay": 0.01,
65
+ "warmup_ratio": 0.1,
66
+ "batch_size": 10,
67
+ "gradient_accumulation_steps": 4,
68
+ "use_focal_loss": true,
69
+ "focal_alpha": 1.0,
70
+ "focal_gamma": 2.0,
71
+ "use_mixup": false,
72
+ "mixup_alpha": 0.2,
73
+ "use_label_smoothing": true,
74
+ "label_smoothing": 0.1
75
+ },
76
+
77
+ "adaptive_lr_config": {
78
+ "adaptive_lr": true,
79
+ "lr_patience": 3,
80
+ "lr_factor": 0.8,
81
+ "lr_increase_factor": 1.2,
82
+ "min_lr": 1e-06,
83
+ "max_lr": 0.001,
84
+ "oscillation_amplitude": 0.1
85
+ },
86
+
87
+ "tokenizer_config": {
88
+ "pad_token": "[PAD]",
89
+ "unk_token": "[UNK]",
90
+ "sep_token": "[SEP]",
91
+ "cls_token": "[CLS]",
92
+ "mask_token": "[MASK]",
93
+ "additional_special_tokens": [
94
+ "[DIALOGUE]",
95
+ "[TURN]",
96
+ "[PAUSE]",
97
+ "[REPEAT]",
98
+ "[HESITATION]"
99
+ ]
100
+ },
101
+
102
+ "torch_dtype": "float32",
103
+ "transformers_version": "4.36.0"
104
  }