Rogendo commited on
Commit
c838fec
·
verified ·
1 Parent(s): 2255a97

Upload JengaAI trained multi-task NLP model

Browse files
checkpoints/best/checkpoint_meta.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 2,
3
+ "metrics": {
4
+ "fraud_detection_fraud_accuracy": 0.5,
5
+ "fraud_detection_fraud_precision": 0.8333333333333334,
6
+ "fraud_detection_fraud_recall": 0.5,
7
+ "fraud_detection_fraud_f1": 0.5,
8
+ "sentiment_analysis_sentiment_accuracy": 0.6666666666666666,
9
+ "sentiment_analysis_sentiment_precision": 0.8333333333333334,
10
+ "sentiment_analysis_sentiment_recall": 0.6666666666666666,
11
+ "sentiment_analysis_sentiment_f1": 0.6666666666666666,
12
+ "call_quality_opening_precision": 0.966,
13
+ "call_quality_opening_recall": 0.966,
14
+ "call_quality_opening_f1": 0.966,
15
+ "call_quality_listening_precision": 0.8960259935016246,
16
+ "call_quality_listening_recall": 0.9421813403416557,
17
+ "call_quality_listening_f1": 0.9185242121445042,
18
+ "call_quality_proactiveness_precision": 0.7589473684210526,
19
+ "call_quality_proactiveness_recall": 0.828735632183908,
20
+ "call_quality_proactiveness_f1": 0.7923076923076923,
21
+ "call_quality_resolution_precision": 0.9283206553540082,
22
+ "call_quality_resolution_recall": 0.8880492583263364,
23
+ "call_quality_resolution_f1": 0.9077385209555142,
24
+ "call_quality_hold_precision": 0.7696629213483146,
25
+ "call_quality_hold_recall": 0.6650485436893204,
26
+ "call_quality_hold_f1": 0.7135416666666666,
27
+ "call_quality_closing_precision": 0.868,
28
+ "call_quality_closing_recall": 0.868,
29
+ "call_quality_closing_f1": 0.868,
30
+ "eval_loss": 1.9685399062990203,
31
+ "train_loss_avg": 1.9783029844836584
32
+ },
33
+ "is_best": true
34
+ }
checkpoints/best/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8b696a7fa15b75f2ac27d2c8f26442591698f223e5dce2e38de5190ba7db982
3
+ size 503468023
checkpoints/epoch_6/checkpoint_meta.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 6,
3
+ "metrics": {
4
+ "fraud_detection_fraud_accuracy": 1.0,
5
+ "fraud_detection_fraud_precision": 1.0,
6
+ "fraud_detection_fraud_recall": 1.0,
7
+ "fraud_detection_fraud_f1": 1.0,
8
+ "sentiment_analysis_sentiment_accuracy": 0.6666666666666666,
9
+ "sentiment_analysis_sentiment_precision": 0.8333333333333334,
10
+ "sentiment_analysis_sentiment_recall": 0.6666666666666666,
11
+ "sentiment_analysis_sentiment_f1": 0.6666666666666666,
12
+ "call_quality_opening_precision": 0.955,
13
+ "call_quality_opening_recall": 0.955,
14
+ "call_quality_opening_f1": 0.955,
15
+ "call_quality_listening_precision": 0.8877526174823472,
16
+ "call_quality_listening_recall": 0.9582128777923784,
17
+ "call_quality_listening_f1": 0.9216380182002022,
18
+ "call_quality_proactiveness_precision": 0.7404233870967742,
19
+ "call_quality_proactiveness_recall": 0.8442528735632184,
20
+ "call_quality_proactiveness_f1": 0.7889366272824919,
21
+ "call_quality_resolution_precision": 0.8931277056277056,
22
+ "call_quality_resolution_recall": 0.9238734956619088,
23
+ "call_quality_resolution_f1": 0.9082404732425368,
24
+ "call_quality_hold_precision": 0.7391304347826086,
25
+ "call_quality_hold_recall": 0.6601941747572816,
26
+ "call_quality_hold_f1": 0.6974358974358974,
27
+ "call_quality_closing_precision": 0.857,
28
+ "call_quality_closing_recall": 0.857,
29
+ "call_quality_closing_f1": 0.857,
30
+ "eval_loss": 2.2391653188511613,
31
+ "train_loss_avg": 1.4627334099144689
32
+ },
33
+ "is_best": false
34
+ }
checkpoints/epoch_6/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d69f2d1a5baa2f1f073d43ca46218f420a9f73473bd6375ecd45053b2001b834
3
+ size 503468023
checkpoints/epoch_7/checkpoint_meta.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 7,
3
+ "metrics": {
4
+ "fraud_detection_fraud_accuracy": 1.0,
5
+ "fraud_detection_fraud_precision": 1.0,
6
+ "fraud_detection_fraud_recall": 1.0,
7
+ "fraud_detection_fraud_f1": 1.0,
8
+ "sentiment_analysis_sentiment_accuracy": 0.3333333333333333,
9
+ "sentiment_analysis_sentiment_precision": 0.1111111111111111,
10
+ "sentiment_analysis_sentiment_recall": 0.3333333333333333,
11
+ "sentiment_analysis_sentiment_f1": 0.16666666666666666,
12
+ "call_quality_opening_precision": 0.955,
13
+ "call_quality_opening_recall": 0.955,
14
+ "call_quality_opening_f1": 0.955,
15
+ "call_quality_listening_precision": 0.8899975663178389,
16
+ "call_quality_listening_recall": 0.9611038107752956,
17
+ "call_quality_listening_f1": 0.9241849886277483,
18
+ "call_quality_proactiveness_precision": 0.7540134645261523,
19
+ "call_quality_proactiveness_recall": 0.8367816091954023,
20
+ "call_quality_proactiveness_f1": 0.7932443475892127,
21
+ "call_quality_resolution_precision": 0.9047486809219661,
22
+ "call_quality_resolution_recall": 0.9118387909319899,
23
+ "call_quality_resolution_f1": 0.9082798996375802,
24
+ "call_quality_hold_precision": 0.7282051282051282,
25
+ "call_quality_hold_recall": 0.6893203883495146,
26
+ "call_quality_hold_f1": 0.7082294264339152,
27
+ "call_quality_closing_precision": 0.853,
28
+ "call_quality_closing_recall": 0.853,
29
+ "call_quality_closing_f1": 0.853,
30
+ "eval_loss": 2.245173363469717,
31
+ "train_loss_avg": 1.345002403695299
32
+ },
33
+ "is_best": false
34
+ }
checkpoints/epoch_7/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9a65dcb9159567e935bd33163026780135a1e88e77f39e6141193cc50a2aacc3
3
+ size 503468023
encoder_config/config.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_cross_attention": false,
3
+ "architectures": [
4
+ "RobertaForMaskedLM"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "gelu",
11
+ "hidden_dropout_prob": 0.1,
12
+ "hidden_size": 768,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 3072,
15
+ "is_decoder": false,
16
+ "layer_norm_eps": 1e-05,
17
+ "max_position_embeddings": 514,
18
+ "model_type": "roberta",
19
+ "num_attention_heads": 12,
20
+ "num_hidden_layers": 12,
21
+ "pad_token_id": 1,
22
+ "tie_word_embeddings": true,
23
+ "transformers_version": "5.0.0",
24
+ "type_vocab_size": 1,
25
+ "use_cache": true,
26
+ "vocab_size": 50265
27
+ }
experiment_config.yaml ADDED
@@ -0,0 +1,131 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ project_name: security_nlp_3task
2
+ tasks:
3
+ - name: fraud_detection
4
+ type: single_label_classification
5
+ data_path: fixtures/sample_classification.jsonl
6
+ heads:
7
+ - name: fraud
8
+ num_labels: 2
9
+ weight: 1.0
10
+ dropout: 0.1
11
+ text_column: text
12
+ label_column: label
13
+ label_maps: null
14
+ - name: sentiment_analysis
15
+ type: single_label_classification
16
+ data_path: fixtures/sample_sentiment.jsonl
17
+ heads:
18
+ - name: sentiment
19
+ num_labels: 3
20
+ weight: 1.0
21
+ dropout: 0.1
22
+ text_column: text
23
+ label_column: label
24
+ label_maps: null
25
+ - name: call_quality
26
+ type: question_answering
27
+ data_path: /content/JengaAI/fixtures/synthetic_qa_metrics_data_v01x.json
28
+ heads:
29
+ - name: opening
30
+ num_labels: 1
31
+ weight: 1.0
32
+ dropout: 0.1
33
+ - name: listening
34
+ num_labels: 5
35
+ weight: 1.5
36
+ dropout: 0.1
37
+ - name: proactiveness
38
+ num_labels: 3
39
+ weight: 1.0
40
+ dropout: 0.1
41
+ - name: resolution
42
+ num_labels: 5
43
+ weight: 2.0
44
+ dropout: 0.1
45
+ - name: hold
46
+ num_labels: 2
47
+ weight: 0.5
48
+ dropout: 0.1
49
+ - name: closing
50
+ num_labels: 1
51
+ weight: 1.0
52
+ dropout: 0.1
53
+ text_column: transcript
54
+ label_column: labels
55
+ label_maps:
56
+ opening:
57
+ 0: greeting
58
+ listening:
59
+ 0: acknowledgment
60
+ 1: empathy
61
+ 2: clarification
62
+ 3: active_listening
63
+ 4: patience
64
+ proactiveness:
65
+ 0: initiative
66
+ 1: follow_up
67
+ 2: suggestions
68
+ resolution:
69
+ 0: identified_issue
70
+ 1: provided_solution
71
+ 2: confirmed_resolution
72
+ 3: set_expectations
73
+ 4: offered_alternatives
74
+ hold:
75
+ 0: asked_permission
76
+ 1: explained_reason
77
+ closing:
78
+ 0: proper_farewell
79
+ model:
80
+ base_model: roberta-base
81
+ hidden_size: 768
82
+ dropout: 0.1
83
+ fusion:
84
+ type: attention
85
+ dropout: 0.1
86
+ use_residual: true
87
+ num_attention_heads: 1
88
+ gate_init_value: 0.5
89
+ freeze_encoder_layers: 0
90
+ gradient_checkpointing: false
91
+ tokenizer:
92
+ max_length: 256
93
+ padding: max_length
94
+ truncation: true
95
+ training:
96
+ output_dir: ./results/roberta-base-3task_fusion
97
+ learning_rate: 2.0e-05
98
+ batch_size: 8
99
+ eval_batch_size: 8
100
+ num_epochs: 15
101
+ weight_decay: 0.01
102
+ warmup_steps: 20
103
+ max_grad_norm: 1.0
104
+ gradient_accumulation_steps: 1
105
+ use_amp: false
106
+ device: auto
107
+ task_sampling: proportional
108
+ temperature: 2.0
109
+ early_stopping_patience: 5
110
+ metric_for_best_model: eval_loss
111
+ greater_is_better: false
112
+ logging:
113
+ service: tensorboard
114
+ experiment_name: security_3task_fusion
115
+ tracking_uri: null
116
+ log_every_n_steps: 5
117
+ checkpoint:
118
+ save_every_n_epochs: 1
119
+ save_best: true
120
+ max_checkpoints: 2
121
+ data:
122
+ test_size: 0.2
123
+ seed: 42
124
+ num_workers: 0
125
+ pin_memory: true
126
+ pii_redaction:
127
+ enabled: false
128
+ strategy: mask
129
+ detect_types: null
130
+ hash_salt: jenga-ai-pii
131
+ log_detections: true
logs/security_3task_fusion/events.out.tfevents.1771237843.71520f106676.2965.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:767b24c6c27e9850bc86cf0eb369b6e7f678ab706012980d26fe9db54bcc9bc0
3
+ size 55381
metadata.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "base_model": "roberta-base",
3
+ "hidden_size": 768,
4
+ "num_tasks": 3,
5
+ "task_names": [
6
+ "fraud_detection",
7
+ "sentiment_analysis",
8
+ "call_quality"
9
+ ],
10
+ "task_types": [
11
+ "single_label_classification",
12
+ "single_label_classification",
13
+ "question_answering"
14
+ ],
15
+ "has_fusion": true
16
+ }
model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9a65dcb9159567e935bd33163026780135a1e88e77f39e6141193cc50a2aacc3
3
+ size 503468023