roberta-large-mrpc / all_results.json
Ubuntu
initial upload
fc6dbea
raw
history blame contribute delete
485 Bytes
{
"epoch": 10.0,
"eval_accuracy": 0.8799019607843137,
"eval_combined_score": 0.8962789661319073,
"eval_f1": 0.9126559714795008,
"eval_loss": 0.42433881759643555,
"eval_runtime": 2.0487,
"eval_samples": 408,
"eval_samples_per_second": 199.151,
"eval_steps_per_second": 3.417,
"train_loss": 0.2565036392211914,
"train_runtime": 194.3012,
"train_samples": 3668,
"train_samples_per_second": 188.779,
"train_steps_per_second": 1.029
}