teru00801 commited on
Commit
dce6706
·
verified ·
1 Parent(s): 536bda9

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,6 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ 20260222_105221/analysis_bundle/cartography/task1_cartography.csv filter=lfs diff=lfs merge=lfs -text
37
+ 20260222_105221/analysis_bundle/cartography/task1_cartography.png filter=lfs diff=lfs merge=lfs -text
38
+ 20260222_105221/analysis_bundle/predictions/final_eval_predictions.csv filter=lfs diff=lfs merge=lfs -text
20260222_105221/analysis_bundle/cartography/summary.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "epochs": 6,
3
+ "task1_rows": 13552,
4
+ "task1_lowmean_threshold": 0.0,
5
+ "task1_suspects": 91,
6
+ "plot_saved": true
7
+ }
20260222_105221/analysis_bundle/cartography/task1_cartography.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1717e8994cb5fc001f3d2d58325d52baf250e97287ee20e3cd2b58a81f5a0f39
3
+ size 44563543
20260222_105221/analysis_bundle/cartography/task1_cartography.png ADDED

Git LFS Details

  • SHA256: be0a8f1f704f4f15729662dd6842c5a4b79f97bb72823d5c2b4acdadbffd05a5
  • Pointer size: 131 Bytes
  • Size of remote file: 435 kB
20260222_105221/analysis_bundle/metrics/metrics.jsonl ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {"epoch": 1, "train_loss": 0.36887128436509103, "acc": 0.8759964570416298, "f1": 0.826946847960445, "precision": 0.7559322033898305, "recall": 0.91268758526603, "score": 0.8759964570416298}
2
+ {"epoch": 2, "train_loss": 0.24266214143603582, "acc": 0.8742249778565102, "f1": 0.8246913580246914, "precision": 0.7531003382187148, "recall": 0.9113233287858117, "score": 0.8742249778565102}
3
+ {"epoch": 3, "train_loss": 0.17757338220792368, "acc": 0.8715677590788308, "f1": 0.8064085447263017, "precision": 0.7895424836601307, "recall": 0.8240109140518418, "score": 0.8715677590788308}
4
+ {"epoch": 4, "train_loss": 0.13203667834788174, "acc": 0.8782108060230293, "f1": 0.8175182481751825, "precision": 0.7958656330749354, "recall": 0.8403819918144612, "score": 0.8782108060230293}
5
+ {"epoch": 5, "train_loss": 0.1085118267381698, "acc": 0.87466784765279, "f1": 0.8279635258358663, "precision": 0.7467105263157895, "recall": 0.9290586630286494, "score": 0.87466784765279}
6
+ {"epoch": 6, "train_loss": 0.08938296143198661, "acc": 0.8782108060230293, "f1": 0.8206131767775603, "precision": 0.78625, "recall": 0.8581173260572987, "score": 0.8782108060230293}
20260222_105221/analysis_bundle/predictions/final_eval_predictions.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6e7f6499d2e0491c5974f1cf4287489efbcf8ce1f355d8a4869007f65f53567
3
+ size 11183672
20260222_105221/analysis_bundle/summaries/final_eval_summary.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ dataset,rows_evaluated,acc,f1,precision,recall,always_pick_acc,always_decline_acc
2
+ data_for_deberta/dedup_split/positive_test.csv,3389,0.8890528179403954,0.8362369337979094,0.803347280334728,0.8719346049046321,0.3248745942755975,0.6751254057244025
20260222_105221/analysis_bundle/summaries/summary.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_epoch": 4,
3
+ "best_score": 0.8782108060230293,
4
+ "train_rows": 13552,
5
+ "val_rows": 2258,
6
+ "final_eval_rows": 3389,
7
+ "val_metrics": {
8
+ "acc": 0.8782108060230293,
9
+ "f1": 0.8175182481751825,
10
+ "precision": 0.7958656330749354,
11
+ "recall": 0.8403819918144612
12
+ },
13
+ "final_eval_metrics": {
14
+ "acc": 0.8890528179403954,
15
+ "f1": 0.8362369337979094,
16
+ "precision": 0.803347280334728,
17
+ "recall": 0.8719346049046321
18
+ },
19
+ "final_eval_baseline": {
20
+ "always_pick_acc": 0.3248745942755975,
21
+ "always_decline_acc": 0.6751254057244025
22
+ }
23
+ }