Commit ·
c5da19e
1
Parent(s): daa8956
retrained using class weights to address imbalance
Browse files- README.md +2 -2
- config.json +1 -1
- pytorch_model.bin +1 -1
- runs/version_0/{events.out.tfevents.1649251721.panacea.34025.0 → events.out.tfevents.1650244759.panacea.16205.0} +2 -2
- runs/version_0/{events.out.tfevents.1649252717.panacea.34025.1 → events.out.tfevents.1650245505.panacea.16205.1} +2 -2
- runs/version_0/hparams.yaml +3 -0
README.md
CHANGED
|
@@ -30,6 +30,6 @@ Stats:
|
|
| 30 |
- **Validation set**: 141 dialogues, using manual donation intention labels from Persuasion For Good 'AnnSet'
|
| 31 |
- **Test set**: 143 dialogues, using manual donation intention labels from Persuasion For Good 'AnnSet'
|
| 32 |
|
| 33 |
-
**Test Macro F1**: 0.
|
| 34 |
|
| 35 |
-
**Test Accuracy**: 0.
|
|
|
|
| 30 |
- **Validation set**: 141 dialogues, using manual donation intention labels from Persuasion For Good 'AnnSet'
|
| 31 |
- **Test set**: 143 dialogues, using manual donation intention labels from Persuasion For Good 'AnnSet'
|
| 32 |
|
| 33 |
+
**Test Macro F1**: 0.893
|
| 34 |
|
| 35 |
+
**Test Accuracy**: 0.902
|
config.json
CHANGED
|
@@ -5,7 +5,7 @@
|
|
| 5 |
],
|
| 6 |
"attention_probs_dropout_prob": 0.1,
|
| 7 |
"bos_token_id": 0,
|
| 8 |
-
"classifier_dropout":
|
| 9 |
"eos_token_id": 2,
|
| 10 |
"hidden_act": "gelu",
|
| 11 |
"hidden_dropout_prob": 0.1,
|
|
|
|
| 5 |
],
|
| 6 |
"attention_probs_dropout_prob": 0.1,
|
| 7 |
"bos_token_id": 0,
|
| 8 |
+
"classifier_dropout": 0.1,
|
| 9 |
"eos_token_id": 2,
|
| 10 |
"hidden_act": "gelu",
|
| 11 |
"hidden_dropout_prob": 0.1,
|
pytorch_model.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1421605805
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:44a5dd064a761d0defd5a4f2a4acbea6e43baf1b8c0c790fce81f5db17ee630d
|
| 3 |
size 1421605805
|
runs/version_0/{events.out.tfevents.1649251721.panacea.34025.0 → events.out.tfevents.1650244759.panacea.16205.0}
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7d94c767b259402a5f887831a025bb47c5c0d2bdfa0d84f9838afae2e272f945
|
| 3 |
+
size 17391
|
runs/version_0/{events.out.tfevents.1649252717.panacea.34025.1 → events.out.tfevents.1650245505.panacea.16205.1}
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:669b84e00fd0b5135e7dbdd2fabf0948449cd135c392cb38049f6c10b78072ce
|
| 3 |
+
size 4482
|
runs/version_0/hparams.yaml
CHANGED
|
@@ -13,6 +13,7 @@ check_val_every_n_epoch: 1
|
|
| 13 |
checkpoint_callback: null
|
| 14 |
checkpoint_save_top_k: 1
|
| 15 |
checkpoint_save_weights_only: true
|
|
|
|
| 16 |
data_random_state: 42
|
| 17 |
datasetpath: ../data/persuasion_for_good/AnnotatedData
|
| 18 |
default_root_dir: null
|
|
@@ -45,6 +46,7 @@ min_steps: null
|
|
| 45 |
model_random_state: 42
|
| 46 |
move_metrics_to_cpu: false
|
| 47 |
multiple_trainloader_mode: max_size_cycle
|
|
|
|
| 48 |
num_nodes: 1
|
| 49 |
num_processes: 1
|
| 50 |
num_sanity_val_steps: 2
|
|
@@ -70,5 +72,6 @@ val_check_interval: 0.25
|
|
| 70 |
val_split: 0.1
|
| 71 |
warmup_steps: 0
|
| 72 |
weight_decay: 0.0
|
|
|
|
| 73 |
weights_save_path: null
|
| 74 |
weights_summary: top
|
|
|
|
| 13 |
checkpoint_callback: null
|
| 14 |
checkpoint_save_top_k: 1
|
| 15 |
checkpoint_save_weights_only: true
|
| 16 |
+
classifier_dropout: 0.1
|
| 17 |
data_random_state: 42
|
| 18 |
datasetpath: ../data/persuasion_for_good/AnnotatedData
|
| 19 |
default_root_dir: null
|
|
|
|
| 46 |
model_random_state: 42
|
| 47 |
move_metrics_to_cpu: false
|
| 48 |
multiple_trainloader_mode: max_size_cycle
|
| 49 |
+
num_classes: 2
|
| 50 |
num_nodes: 1
|
| 51 |
num_processes: 1
|
| 52 |
num_sanity_val_steps: 2
|
|
|
|
| 72 |
val_split: 0.1
|
| 73 |
warmup_steps: 0
|
| 74 |
weight_decay: 0.0
|
| 75 |
+
weighted_loss: true
|
| 76 |
weights_save_path: null
|
| 77 |
weights_summary: top
|