bert-large-en / eval_results.txt
TransQuest's picture
First model version
f3cccb2
raw
history blame contribute delete
237 Bytes
accuracy = 0.9375307427447123
auprc = 0.9100596170919103
auroc = 0.9737253068309618
eval_loss = 0.1730787594569847
fn = 59
fp = 68
macro_f1 = 0.9021931544186028
mcc = 0.8044696056183317
tn = 1564
tp = 342
weighted_f1 = 0.937791003215199