Upload JengaAI trained multi-task NLP model
Browse files- checkpoints/best/checkpoint_meta.json +34 -0
- checkpoints/best/model.pt +3 -0
- checkpoints/epoch_6/checkpoint_meta.json +34 -0
- checkpoints/epoch_6/model.pt +3 -0
- checkpoints/epoch_7/checkpoint_meta.json +34 -0
- checkpoints/epoch_7/model.pt +3 -0
- encoder_config/config.json +27 -0
- experiment_config.yaml +131 -0
- logs/security_3task_fusion/events.out.tfevents.1771237843.71520f106676.2965.0 +3 -0
- metadata.json +16 -0
- model.pt +3 -0
checkpoints/best/checkpoint_meta.json
ADDED
|
@@ -0,0 +1,34 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"epoch": 2,
|
| 3 |
+
"metrics": {
|
| 4 |
+
"fraud_detection_fraud_accuracy": 0.5,
|
| 5 |
+
"fraud_detection_fraud_precision": 0.8333333333333334,
|
| 6 |
+
"fraud_detection_fraud_recall": 0.5,
|
| 7 |
+
"fraud_detection_fraud_f1": 0.5,
|
| 8 |
+
"sentiment_analysis_sentiment_accuracy": 0.6666666666666666,
|
| 9 |
+
"sentiment_analysis_sentiment_precision": 0.8333333333333334,
|
| 10 |
+
"sentiment_analysis_sentiment_recall": 0.6666666666666666,
|
| 11 |
+
"sentiment_analysis_sentiment_f1": 0.6666666666666666,
|
| 12 |
+
"call_quality_opening_precision": 0.966,
|
| 13 |
+
"call_quality_opening_recall": 0.966,
|
| 14 |
+
"call_quality_opening_f1": 0.966,
|
| 15 |
+
"call_quality_listening_precision": 0.8960259935016246,
|
| 16 |
+
"call_quality_listening_recall": 0.9421813403416557,
|
| 17 |
+
"call_quality_listening_f1": 0.9185242121445042,
|
| 18 |
+
"call_quality_proactiveness_precision": 0.7589473684210526,
|
| 19 |
+
"call_quality_proactiveness_recall": 0.828735632183908,
|
| 20 |
+
"call_quality_proactiveness_f1": 0.7923076923076923,
|
| 21 |
+
"call_quality_resolution_precision": 0.9283206553540082,
|
| 22 |
+
"call_quality_resolution_recall": 0.8880492583263364,
|
| 23 |
+
"call_quality_resolution_f1": 0.9077385209555142,
|
| 24 |
+
"call_quality_hold_precision": 0.7696629213483146,
|
| 25 |
+
"call_quality_hold_recall": 0.6650485436893204,
|
| 26 |
+
"call_quality_hold_f1": 0.7135416666666666,
|
| 27 |
+
"call_quality_closing_precision": 0.868,
|
| 28 |
+
"call_quality_closing_recall": 0.868,
|
| 29 |
+
"call_quality_closing_f1": 0.868,
|
| 30 |
+
"eval_loss": 1.9685399062990203,
|
| 31 |
+
"train_loss_avg": 1.9783029844836584
|
| 32 |
+
},
|
| 33 |
+
"is_best": true
|
| 34 |
+
}
|
checkpoints/best/model.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c8b696a7fa15b75f2ac27d2c8f26442591698f223e5dce2e38de5190ba7db982
|
| 3 |
+
size 503468023
|
checkpoints/epoch_6/checkpoint_meta.json
ADDED
|
@@ -0,0 +1,34 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"epoch": 6,
|
| 3 |
+
"metrics": {
|
| 4 |
+
"fraud_detection_fraud_accuracy": 1.0,
|
| 5 |
+
"fraud_detection_fraud_precision": 1.0,
|
| 6 |
+
"fraud_detection_fraud_recall": 1.0,
|
| 7 |
+
"fraud_detection_fraud_f1": 1.0,
|
| 8 |
+
"sentiment_analysis_sentiment_accuracy": 0.6666666666666666,
|
| 9 |
+
"sentiment_analysis_sentiment_precision": 0.8333333333333334,
|
| 10 |
+
"sentiment_analysis_sentiment_recall": 0.6666666666666666,
|
| 11 |
+
"sentiment_analysis_sentiment_f1": 0.6666666666666666,
|
| 12 |
+
"call_quality_opening_precision": 0.955,
|
| 13 |
+
"call_quality_opening_recall": 0.955,
|
| 14 |
+
"call_quality_opening_f1": 0.955,
|
| 15 |
+
"call_quality_listening_precision": 0.8877526174823472,
|
| 16 |
+
"call_quality_listening_recall": 0.9582128777923784,
|
| 17 |
+
"call_quality_listening_f1": 0.9216380182002022,
|
| 18 |
+
"call_quality_proactiveness_precision": 0.7404233870967742,
|
| 19 |
+
"call_quality_proactiveness_recall": 0.8442528735632184,
|
| 20 |
+
"call_quality_proactiveness_f1": 0.7889366272824919,
|
| 21 |
+
"call_quality_resolution_precision": 0.8931277056277056,
|
| 22 |
+
"call_quality_resolution_recall": 0.9238734956619088,
|
| 23 |
+
"call_quality_resolution_f1": 0.9082404732425368,
|
| 24 |
+
"call_quality_hold_precision": 0.7391304347826086,
|
| 25 |
+
"call_quality_hold_recall": 0.6601941747572816,
|
| 26 |
+
"call_quality_hold_f1": 0.6974358974358974,
|
| 27 |
+
"call_quality_closing_precision": 0.857,
|
| 28 |
+
"call_quality_closing_recall": 0.857,
|
| 29 |
+
"call_quality_closing_f1": 0.857,
|
| 30 |
+
"eval_loss": 2.2391653188511613,
|
| 31 |
+
"train_loss_avg": 1.4627334099144689
|
| 32 |
+
},
|
| 33 |
+
"is_best": false
|
| 34 |
+
}
|
checkpoints/epoch_6/model.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d69f2d1a5baa2f1f073d43ca46218f420a9f73473bd6375ecd45053b2001b834
|
| 3 |
+
size 503468023
|
checkpoints/epoch_7/checkpoint_meta.json
ADDED
|
@@ -0,0 +1,34 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"epoch": 7,
|
| 3 |
+
"metrics": {
|
| 4 |
+
"fraud_detection_fraud_accuracy": 1.0,
|
| 5 |
+
"fraud_detection_fraud_precision": 1.0,
|
| 6 |
+
"fraud_detection_fraud_recall": 1.0,
|
| 7 |
+
"fraud_detection_fraud_f1": 1.0,
|
| 8 |
+
"sentiment_analysis_sentiment_accuracy": 0.3333333333333333,
|
| 9 |
+
"sentiment_analysis_sentiment_precision": 0.1111111111111111,
|
| 10 |
+
"sentiment_analysis_sentiment_recall": 0.3333333333333333,
|
| 11 |
+
"sentiment_analysis_sentiment_f1": 0.16666666666666666,
|
| 12 |
+
"call_quality_opening_precision": 0.955,
|
| 13 |
+
"call_quality_opening_recall": 0.955,
|
| 14 |
+
"call_quality_opening_f1": 0.955,
|
| 15 |
+
"call_quality_listening_precision": 0.8899975663178389,
|
| 16 |
+
"call_quality_listening_recall": 0.9611038107752956,
|
| 17 |
+
"call_quality_listening_f1": 0.9241849886277483,
|
| 18 |
+
"call_quality_proactiveness_precision": 0.7540134645261523,
|
| 19 |
+
"call_quality_proactiveness_recall": 0.8367816091954023,
|
| 20 |
+
"call_quality_proactiveness_f1": 0.7932443475892127,
|
| 21 |
+
"call_quality_resolution_precision": 0.9047486809219661,
|
| 22 |
+
"call_quality_resolution_recall": 0.9118387909319899,
|
| 23 |
+
"call_quality_resolution_f1": 0.9082798996375802,
|
| 24 |
+
"call_quality_hold_precision": 0.7282051282051282,
|
| 25 |
+
"call_quality_hold_recall": 0.6893203883495146,
|
| 26 |
+
"call_quality_hold_f1": 0.7082294264339152,
|
| 27 |
+
"call_quality_closing_precision": 0.853,
|
| 28 |
+
"call_quality_closing_recall": 0.853,
|
| 29 |
+
"call_quality_closing_f1": 0.853,
|
| 30 |
+
"eval_loss": 2.245173363469717,
|
| 31 |
+
"train_loss_avg": 1.345002403695299
|
| 32 |
+
},
|
| 33 |
+
"is_best": false
|
| 34 |
+
}
|
checkpoints/epoch_7/model.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9a65dcb9159567e935bd33163026780135a1e88e77f39e6141193cc50a2aacc3
|
| 3 |
+
size 503468023
|
encoder_config/config.json
ADDED
|
@@ -0,0 +1,27 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"add_cross_attention": false,
|
| 3 |
+
"architectures": [
|
| 4 |
+
"RobertaForMaskedLM"
|
| 5 |
+
],
|
| 6 |
+
"attention_probs_dropout_prob": 0.1,
|
| 7 |
+
"bos_token_id": 0,
|
| 8 |
+
"classifier_dropout": null,
|
| 9 |
+
"eos_token_id": 2,
|
| 10 |
+
"hidden_act": "gelu",
|
| 11 |
+
"hidden_dropout_prob": 0.1,
|
| 12 |
+
"hidden_size": 768,
|
| 13 |
+
"initializer_range": 0.02,
|
| 14 |
+
"intermediate_size": 3072,
|
| 15 |
+
"is_decoder": false,
|
| 16 |
+
"layer_norm_eps": 1e-05,
|
| 17 |
+
"max_position_embeddings": 514,
|
| 18 |
+
"model_type": "roberta",
|
| 19 |
+
"num_attention_heads": 12,
|
| 20 |
+
"num_hidden_layers": 12,
|
| 21 |
+
"pad_token_id": 1,
|
| 22 |
+
"tie_word_embeddings": true,
|
| 23 |
+
"transformers_version": "5.0.0",
|
| 24 |
+
"type_vocab_size": 1,
|
| 25 |
+
"use_cache": true,
|
| 26 |
+
"vocab_size": 50265
|
| 27 |
+
}
|
experiment_config.yaml
ADDED
|
@@ -0,0 +1,131 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
project_name: security_nlp_3task
|
| 2 |
+
tasks:
|
| 3 |
+
- name: fraud_detection
|
| 4 |
+
type: single_label_classification
|
| 5 |
+
data_path: fixtures/sample_classification.jsonl
|
| 6 |
+
heads:
|
| 7 |
+
- name: fraud
|
| 8 |
+
num_labels: 2
|
| 9 |
+
weight: 1.0
|
| 10 |
+
dropout: 0.1
|
| 11 |
+
text_column: text
|
| 12 |
+
label_column: label
|
| 13 |
+
label_maps: null
|
| 14 |
+
- name: sentiment_analysis
|
| 15 |
+
type: single_label_classification
|
| 16 |
+
data_path: fixtures/sample_sentiment.jsonl
|
| 17 |
+
heads:
|
| 18 |
+
- name: sentiment
|
| 19 |
+
num_labels: 3
|
| 20 |
+
weight: 1.0
|
| 21 |
+
dropout: 0.1
|
| 22 |
+
text_column: text
|
| 23 |
+
label_column: label
|
| 24 |
+
label_maps: null
|
| 25 |
+
- name: call_quality
|
| 26 |
+
type: question_answering
|
| 27 |
+
data_path: /content/JengaAI/fixtures/synthetic_qa_metrics_data_v01x.json
|
| 28 |
+
heads:
|
| 29 |
+
- name: opening
|
| 30 |
+
num_labels: 1
|
| 31 |
+
weight: 1.0
|
| 32 |
+
dropout: 0.1
|
| 33 |
+
- name: listening
|
| 34 |
+
num_labels: 5
|
| 35 |
+
weight: 1.5
|
| 36 |
+
dropout: 0.1
|
| 37 |
+
- name: proactiveness
|
| 38 |
+
num_labels: 3
|
| 39 |
+
weight: 1.0
|
| 40 |
+
dropout: 0.1
|
| 41 |
+
- name: resolution
|
| 42 |
+
num_labels: 5
|
| 43 |
+
weight: 2.0
|
| 44 |
+
dropout: 0.1
|
| 45 |
+
- name: hold
|
| 46 |
+
num_labels: 2
|
| 47 |
+
weight: 0.5
|
| 48 |
+
dropout: 0.1
|
| 49 |
+
- name: closing
|
| 50 |
+
num_labels: 1
|
| 51 |
+
weight: 1.0
|
| 52 |
+
dropout: 0.1
|
| 53 |
+
text_column: transcript
|
| 54 |
+
label_column: labels
|
| 55 |
+
label_maps:
|
| 56 |
+
opening:
|
| 57 |
+
0: greeting
|
| 58 |
+
listening:
|
| 59 |
+
0: acknowledgment
|
| 60 |
+
1: empathy
|
| 61 |
+
2: clarification
|
| 62 |
+
3: active_listening
|
| 63 |
+
4: patience
|
| 64 |
+
proactiveness:
|
| 65 |
+
0: initiative
|
| 66 |
+
1: follow_up
|
| 67 |
+
2: suggestions
|
| 68 |
+
resolution:
|
| 69 |
+
0: identified_issue
|
| 70 |
+
1: provided_solution
|
| 71 |
+
2: confirmed_resolution
|
| 72 |
+
3: set_expectations
|
| 73 |
+
4: offered_alternatives
|
| 74 |
+
hold:
|
| 75 |
+
0: asked_permission
|
| 76 |
+
1: explained_reason
|
| 77 |
+
closing:
|
| 78 |
+
0: proper_farewell
|
| 79 |
+
model:
|
| 80 |
+
base_model: roberta-base
|
| 81 |
+
hidden_size: 768
|
| 82 |
+
dropout: 0.1
|
| 83 |
+
fusion:
|
| 84 |
+
type: attention
|
| 85 |
+
dropout: 0.1
|
| 86 |
+
use_residual: true
|
| 87 |
+
num_attention_heads: 1
|
| 88 |
+
gate_init_value: 0.5
|
| 89 |
+
freeze_encoder_layers: 0
|
| 90 |
+
gradient_checkpointing: false
|
| 91 |
+
tokenizer:
|
| 92 |
+
max_length: 256
|
| 93 |
+
padding: max_length
|
| 94 |
+
truncation: true
|
| 95 |
+
training:
|
| 96 |
+
output_dir: ./results/roberta-base-3task_fusion
|
| 97 |
+
learning_rate: 2.0e-05
|
| 98 |
+
batch_size: 8
|
| 99 |
+
eval_batch_size: 8
|
| 100 |
+
num_epochs: 15
|
| 101 |
+
weight_decay: 0.01
|
| 102 |
+
warmup_steps: 20
|
| 103 |
+
max_grad_norm: 1.0
|
| 104 |
+
gradient_accumulation_steps: 1
|
| 105 |
+
use_amp: false
|
| 106 |
+
device: auto
|
| 107 |
+
task_sampling: proportional
|
| 108 |
+
temperature: 2.0
|
| 109 |
+
early_stopping_patience: 5
|
| 110 |
+
metric_for_best_model: eval_loss
|
| 111 |
+
greater_is_better: false
|
| 112 |
+
logging:
|
| 113 |
+
service: tensorboard
|
| 114 |
+
experiment_name: security_3task_fusion
|
| 115 |
+
tracking_uri: null
|
| 116 |
+
log_every_n_steps: 5
|
| 117 |
+
checkpoint:
|
| 118 |
+
save_every_n_epochs: 1
|
| 119 |
+
save_best: true
|
| 120 |
+
max_checkpoints: 2
|
| 121 |
+
data:
|
| 122 |
+
test_size: 0.2
|
| 123 |
+
seed: 42
|
| 124 |
+
num_workers: 0
|
| 125 |
+
pin_memory: true
|
| 126 |
+
pii_redaction:
|
| 127 |
+
enabled: false
|
| 128 |
+
strategy: mask
|
| 129 |
+
detect_types: null
|
| 130 |
+
hash_salt: jenga-ai-pii
|
| 131 |
+
log_detections: true
|
logs/security_3task_fusion/events.out.tfevents.1771237843.71520f106676.2965.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:767b24c6c27e9850bc86cf0eb369b6e7f678ab706012980d26fe9db54bcc9bc0
|
| 3 |
+
size 55381
|
metadata.json
ADDED
|
@@ -0,0 +1,16 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"base_model": "roberta-base",
|
| 3 |
+
"hidden_size": 768,
|
| 4 |
+
"num_tasks": 3,
|
| 5 |
+
"task_names": [
|
| 6 |
+
"fraud_detection",
|
| 7 |
+
"sentiment_analysis",
|
| 8 |
+
"call_quality"
|
| 9 |
+
],
|
| 10 |
+
"task_types": [
|
| 11 |
+
"single_label_classification",
|
| 12 |
+
"single_label_classification",
|
| 13 |
+
"question_answering"
|
| 14 |
+
],
|
| 15 |
+
"has_fusion": true
|
| 16 |
+
}
|
model.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9a65dcb9159567e935bd33163026780135a1e88e77f39e6141193cc50a2aacc3
|
| 3 |
+
size 503468023
|