adapter-8classes-single_label / trainer_state.json
Faeze's picture
Upload folder using huggingface_hub
5f2f386 verified
raw
history blame
385 kB
{
"best_metric": 77.22228406219811,
"best_model_checkpoint": "/mounts/work/faeze/attempt/new_setting_outputs/source_adapter/crisis_8/100/checkpoint-1479",
"epoch": 200.0,
"global_step": 3400,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 1.0,
"learning_rate": 1.4999999999999999e-05,
"loss": 9.5935,
"step": 17
},
{
"epoch": 1.0,
"eval_accuracy": 0.0,
"eval_average_metrics": 0.0,
"eval_classification_report": "{\"\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"\\\"media\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"How Dar\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \")\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"-\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \".\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"0:\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \":\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": D\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": Lots\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": apparently\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": good\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": please\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": very\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"????\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Anders\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Emers\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Can use passport\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Congress needs to\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"False\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Hiding in\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"No Friday Im\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"No they have\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"PPC 2019\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"SCP-04\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Shawn\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"True\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a friend\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"apology\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 11.0}, \"canada is under\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 36.0}, \"classification:\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 39.0}, \"instagram\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"negative\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 13.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 16.0}, \"weather\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.0, \"macro avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 135.0}}",
"eval_f1_macro": 0.0,
"eval_f1_micro": 0.0,
"eval_f1_weighted": 0.0,
"eval_loss": 9.580382347106934,
"eval_runtime": 2.2821,
"eval_samples_per_second": 59.156,
"step": 17
},
{
"epoch": 2.0,
"learning_rate": 2.9999999999999997e-05,
"loss": 9.0928,
"step": 34
},
{
"epoch": 2.0,
"eval_accuracy": 0.0,
"eval_average_metrics": 0.0,
"eval_classification_report": "{\"\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"\\\"media\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"How Dar\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \")\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"-\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \".\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"99.5%\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": Lots\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": apparently\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": good\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": please\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": very\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Anders\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Emers\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Can use passport\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Climate change is\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Congress needs to\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Ddx\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"False\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Hiding in\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"No Friday Im\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"No they have\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"PPC 2019\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Shawn\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"True\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a friend\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"apology\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 11.0}, \"canada is under\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 39.0}, \"instagram\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"negative\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 13.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 16.0}, \"vac\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"weather\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.0, \"macro avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 135.0}}",
"eval_f1_macro": 0.0,
"eval_f1_micro": 0.0,
"eval_f1_weighted": 0.0,
"eval_loss": 9.190422058105469,
"eval_runtime": 2.2408,
"eval_samples_per_second": 60.245,
"step": 34
},
{
"epoch": 3.0,
"learning_rate": 4.4999999999999996e-05,
"loss": 8.1515,
"step": 51
},
{
"epoch": 3.0,
"eval_accuracy": 0.0,
"eval_average_metrics": 0.0,
"eval_classification_report": "{\"\\\"How Dar\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"-\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"0.0\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": Lots\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": good\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": please\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": very\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"?\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Anders\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Can use passport\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Congress needs to\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Ddx\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"False\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"No Friday Im\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"PPC 2019\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"SCP-04\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Shawn\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"True\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a friend\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a real\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"apology\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 11.0}, \"canada is under\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 36.0}, \"classification: climate\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"entail\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 39.0}, \"instagram\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"negative\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_e\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 13.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 16.0}, \"vac\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"weather\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.0, \"macro avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 135.0}}",
"eval_f1_macro": 0.0,
"eval_f1_micro": 0.0,
"eval_f1_weighted": 0.0,
"eval_loss": 8.223907470703125,
"eval_runtime": 2.3068,
"eval_samples_per_second": 58.523,
"step": 51
},
{
"epoch": 4.0,
"learning_rate": 5.9999999999999995e-05,
"loss": 6.6637,
"step": 68
},
{
"epoch": 4.0,
"eval_accuracy": 0.0,
"eval_average_metrics": 0.0,
"eval_classification_report": "{\"\\\"How Dar\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"Reality\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \".\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"0.0\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"?\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"????\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Andrew Neil\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Australian\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Bruh\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Congress needs to\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Ddx\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"False\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"SCP-04\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"True\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a friend\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a real\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"apology\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 11.0}, \"canada\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 36.0}, \"classification: good\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"eder\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"entail\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 39.0}, \"name\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"negative\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_e\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_en\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"panne\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"please show us\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 13.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 16.0}, \"vac\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"very goi\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"weather\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.0, \"macro avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 135.0}}",
"eval_f1_macro": 0.0,
"eval_f1_micro": 0.0,
"eval_f1_weighted": 0.0,
"eval_loss": 6.206799030303955,
"eval_runtime": 2.267,
"eval_samples_per_second": 59.549,
"step": 68
},
{
"epoch": 5.0,
"learning_rate": 7.5e-05,
"loss": 4.5728,
"step": 85
},
{
"epoch": 5.0,
"eval_accuracy": 0.0,
"eval_average_metrics": 0.0,
"eval_classification_report": "{\"\\\"Reality\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \".\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"0\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"1\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"33000f\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \":\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"?\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"????\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Andrew Neil\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Australian\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Bruh\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"CNN\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"COVID\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Congress needs to\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Ddx\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Florida\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"India\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"PPC\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"SCP-04\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"apology\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 11.0}, \"canada\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 36.0}, \"dal\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"doubtful\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"eder\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"entail\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"if de\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 39.0}, \"microwave\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"name\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"negative\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_d\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_e\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_en\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 13.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 16.0}, \"to\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"troll\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"trud\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"u\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"vac\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"weather\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.0, \"macro avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 135.0}}",
"eval_f1_macro": 0.0,
"eval_f1_micro": 0.0,
"eval_f1_weighted": 0.0,
"eval_loss": 3.01625657081604,
"eval_runtime": 2.3046,
"eval_samples_per_second": 58.579,
"step": 85
},
{
"epoch": 6.0,
"learning_rate": 8.999999999999999e-05,
"loss": 2.1563,
"step": 102
},
{
"epoch": 6.0,
"eval_accuracy": 25.925925925925924,
"eval_average_metrics": 18.478859129440288,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"apology\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 11.0}, \"challenge\": {\"precision\": 0.27049180327868855, \"recall\": 0.9166666666666666, \"f1-score\": 0.4177215189873418, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 39.0}, \"not_e\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 13.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.125, \"f1-score\": 0.2222222222222222, \"support\": 16.0}, \"statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.25925925925925924, \"macro avg\": {\"precision\": 0.11549925484351714, \"recall\": 0.09469696969696968, \"f1-score\": 0.058176703746324, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.19064966605950215, \"recall\": 0.25925925925925924, \"f1-score\": 0.13772985362296192, \"support\": 135.0}}",
"eval_f1_macro": 7.99929676511955,
"eval_f1_micro": 26.217228464419474,
"eval_f1_weighted": 13.772985362296192,
"eval_loss": 1.1285314559936523,
"eval_runtime": 2.054,
"eval_samples_per_second": 65.725,
"step": 102
},
{
"epoch": 7.0,
"learning_rate": 0.00010499999999999999,
"loss": 1.0166,
"step": 119
},
{
"epoch": 7.0,
"eval_accuracy": 40.0,
"eval_average_metrics": 33.589515948427554,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"apology\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.2727272727272727, \"f1-score\": 0.39999999999999997, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.16666666666666666, \"f1-score\": 0.26666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.3392857142857143, \"recall\": 0.9743589743589743, \"f1-score\": 0.5033112582781457, \"support\": 39.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 13.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.4375, \"f1-score\": 0.6086956521739131, \"support\": 16.0}, \"accuracy\": 0.4, \"macro avg\": {\"precision\": 0.3444940476190476, \"recall\": 0.23140661421911424, \"f1-score\": 0.22233419713984065, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.45542328042328045, \"recall\": 0.4, \"f1-score\": 0.3212464407972614, \"support\": 135.0}}",
"eval_f1_macro": 22.233419713984066,
"eval_f1_micro": 40.00000000000001,
"eval_f1_weighted": 32.12464407972614,
"eval_loss": 0.6725317239761353,
"eval_runtime": 2.2366,
"eval_samples_per_second": 60.359,
"step": 119
},
{
"epoch": 8.0,
"learning_rate": 0.00011999999999999999,
"loss": 0.7319,
"step": 136
},
{
"epoch": 8.0,
"eval_accuracy": 48.888888888888886,
"eval_average_metrics": 46.684741490249316,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.4, \"recall\": 0.2222222222222222, \"f1-score\": 0.2857142857142857, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.45454545454545453, \"f1-score\": 0.5555555555555556, \"support\": 11.0}, \"challenge\": {\"precision\": 0.4426229508196721, \"recall\": 0.75, \"f1-score\": 0.556701030927835, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.47619047619047616, \"recall\": 0.5128205128205128, \"f1-score\": 0.49382716049382713, \"support\": 39.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"rejection\": {\"precision\": 0.25, \"recall\": 0.15384615384615385, \"f1-score\": 0.1904761904761905, \"support\": 13.0}, \"request\": {\"precision\": 0.8181818181818182, \"recall\": 0.5625, \"f1-score\": 0.6666666666666666, \"support\": 16.0}, \"accuracy\": 0.4888888888888889, \"macro avg\": {\"precision\": 0.5126601199347101, \"recall\": 0.39449179292929293, \"f1-score\": 0.4269509445626284, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.47632523517769415, \"recall\": 0.4888888888888889, \"f1-score\": 0.46266093726956653, \"support\": 135.0}}",
"eval_f1_macro": 42.69509445626284,
"eval_f1_micro": 48.888888888888886,
"eval_f1_weighted": 46.26609372695665,
"eval_loss": 0.5566144585609436,
"eval_runtime": 2.1877,
"eval_samples_per_second": 61.709,
"step": 136
},
{
"epoch": 9.0,
"learning_rate": 0.000135,
"loss": 0.6661,
"step": 153
},
{
"epoch": 9.0,
"eval_accuracy": 54.074074074074076,
"eval_average_metrics": 54.03755296584174,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.3333333333333333, \"f1-score\": 0.42857142857142855, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7777777777777778, \"recall\": 0.6363636363636364, \"f1-score\": 0.7000000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.5581395348837209, \"recall\": 0.6666666666666666, \"f1-score\": 0.6075949367088608, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7272727272727273, \"recall\": 0.41025641025641024, \"f1-score\": 0.5245901639344261, \"support\": 39.0}, \"question\": {\"precision\": 0.75, \"recall\": 0.3333333333333333, \"f1-score\": 0.46153846153846156, \"support\": 9.0}, \"rejection\": {\"precision\": 0.2727272727272727, \"recall\": 0.46153846153846156, \"f1-score\": 0.3428571428571428, \"support\": 13.0}, \"request\": {\"precision\": 0.4482758620689655, \"recall\": 0.8125, \"f1-score\": 0.5777777777777777, \"support\": 16.0}, \"accuracy\": 0.5407407407407407, \"macro avg\": {\"precision\": 0.6417741468413081, \"recall\": 0.5192489801864801, \"f1-score\": 0.5386995722568455, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.60651913713751, \"recall\": 0.5407407407407407, \"f1-score\": 0.5413210648953426, \"support\": 135.0}}",
"eval_f1_macro": 53.86995722568455,
"eval_f1_micro": 54.074074074074076,
"eval_f1_weighted": 54.13210648953427,
"eval_loss": 0.511962890625,
"eval_runtime": 2.1819,
"eval_samples_per_second": 61.874,
"step": 153
},
{
"epoch": 10.0,
"learning_rate": 0.00015,
"loss": 0.6131,
"step": 170
},
{
"epoch": 10.0,
"eval_accuracy": 57.03703703703704,
"eval_average_metrics": 57.547429298101335,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.7142857142857143, \"recall\": 0.5555555555555556, \"f1-score\": 0.6250000000000001, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7272727272727273, \"recall\": 0.7272727272727273, \"f1-score\": 0.7272727272727273, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6111111111111112, \"recall\": 0.6111111111111112, \"f1-score\": 0.6111111111111112, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7391304347826086, \"recall\": 0.4358974358974359, \"f1-score\": 0.5483870967741935, \"support\": 39.0}, \"question\": {\"precision\": 0.8, \"recall\": 0.4444444444444444, \"f1-score\": 0.5714285714285714, \"support\": 9.0}, \"rejection\": {\"precision\": 0.3157894736842105, \"recall\": 0.46153846153846156, \"f1-score\": 0.37499999999999994, \"support\": 13.0}, \"request\": {\"precision\": 0.42424242424242425, \"recall\": 0.875, \"f1-score\": 0.5714285714285714, \"support\": 16.0}, \"accuracy\": 0.5703703703703704, \"macro avg\": {\"precision\": 0.6664789856723495, \"recall\": 0.576352466977467, \"f1-score\": 0.5870368430852302, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.6322059283767903, \"recall\": 0.5703703703703704, \"f1-score\": 0.5741195880980826, \"support\": 135.0}}",
"eval_f1_macro": 58.703684308523016,
"eval_f1_micro": 57.03703703703704,
"eval_f1_weighted": 57.41195880980826,
"eval_loss": 0.4787319600582123,
"eval_runtime": 2.2206,
"eval_samples_per_second": 60.794,
"step": 170
},
{
"epoch": 11.0,
"learning_rate": 0.000165,
"loss": 0.5787,
"step": 187
},
{
"epoch": 11.0,
"eval_accuracy": 64.44444444444444,
"eval_average_metrics": 65.62118906429522,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6428571428571429, \"recall\": 0.8181818181818182, \"f1-score\": 0.7200000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6333333333333333, \"recall\": 0.5277777777777778, \"f1-score\": 0.5757575757575758, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.5833333333333334, \"recall\": 0.717948717948718, \"f1-score\": 0.6436781609195402, \"support\": 39.0}, \"question\": {\"precision\": 0.625, \"recall\": 0.5555555555555556, \"f1-score\": 0.5882352941176471, \"support\": 9.0}, \"rejection\": {\"precision\": 0.7, \"recall\": 0.5384615384615384, \"f1-score\": 0.608695652173913, \"support\": 13.0}, \"request\": {\"precision\": 0.7333333333333333, \"recall\": 0.6875, \"f1-score\": 0.7096774193548386, \"support\": 16.0}, \"accuracy\": 0.6444444444444445, \"macro avg\": {\"precision\": 0.7084821428571428, \"recall\": 0.6890115093240092, \"f1-score\": 0.6939908069080865, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.6505908289241622, \"recall\": 0.6444444444444445, \"f1-score\": 0.6419678667748333, \"support\": 135.0}}",
"eval_f1_macro": 69.39908069080865,
"eval_f1_micro": 64.44444444444444,
"eval_f1_weighted": 64.19678667748333,
"eval_loss": 0.4315463900566101,
"eval_runtime": 2.336,
"eval_samples_per_second": 57.791,
"step": 187
},
{
"epoch": 12.0,
"learning_rate": 0.00017999999999999998,
"loss": 0.5394,
"step": 204
},
{
"epoch": 12.0,
"eval_accuracy": 68.14814814814815,
"eval_average_metrics": 69.08949458669183,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6, \"recall\": 0.8181818181818182, \"f1-score\": 0.6923076923076923, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6444444444444445, \"recall\": 0.8055555555555556, \"f1-score\": 0.7160493827160493, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6896551724137931, \"recall\": 0.5128205128205128, \"f1-score\": 0.5882352941176471, \"support\": 39.0}, \"question\": {\"precision\": 0.8333333333333334, \"recall\": 0.5555555555555556, \"f1-score\": 0.6666666666666667, \"support\": 9.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.6153846153846154, \"f1-score\": 0.6956521739130435, \"support\": 13.0}, \"request\": {\"precision\": 0.7333333333333333, \"recall\": 0.6875, \"f1-score\": 0.7096774193548386, \"support\": 16.0}, \"accuracy\": 0.6814814814814815, \"macro avg\": {\"precision\": 0.7395188623636899, \"recall\": 0.7354858682983683, \"f1-score\": 0.7244826695435831, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.6953210858957987, \"recall\": 0.6814814814814815, \"f1-score\": 0.6761341509611271, \"support\": 135.0}}",
"eval_f1_macro": 72.4482669543583,
"eval_f1_micro": 68.14814814814815,
"eval_f1_weighted": 67.6134150961127,
"eval_loss": 0.3929964601993561,
"eval_runtime": 2.2653,
"eval_samples_per_second": 59.595,
"step": 204
},
{
"epoch": 13.0,
"learning_rate": 0.000195,
"loss": 0.4445,
"step": 221
},
{
"epoch": 13.0,
"eval_accuracy": 68.14814814814815,
"eval_average_metrics": 69.54531108529271,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6136363636363636, \"recall\": 0.75, \"f1-score\": 0.6749999999999999, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.631578947368421, \"recall\": 0.6153846153846154, \"f1-score\": 0.6233766233766234, \"support\": 39.0}, \"question\": {\"precision\": 0.8571428571428571, \"recall\": 0.6666666666666666, \"f1-score\": 0.75, \"support\": 9.0}, \"rejection\": {\"precision\": 0.8571428571428571, \"recall\": 0.46153846153846156, \"f1-score\": 0.6, \"support\": 13.0}, \"request\": {\"precision\": 0.6875, \"recall\": 0.6875, \"f1-score\": 0.6875, \"support\": 16.0}, \"accuracy\": 0.6814814814814815, \"macro avg\": {\"precision\": 0.7767885896997739, \"recall\": 0.7221311674436675, \"f1-score\": 0.7386757543926662, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.6968149297096665, \"recall\": 0.6814814814814815, \"f1-score\": 0.6801737260560791, \"support\": 135.0}}",
"eval_f1_macro": 73.86757543926662,
"eval_f1_micro": 68.14814814814815,
"eval_f1_weighted": 68.0173726056079,
"eval_loss": 0.3637169301509857,
"eval_runtime": 2.1262,
"eval_samples_per_second": 63.494,
"step": 221
},
{
"epoch": 14.0,
"learning_rate": 0.00020999999999999998,
"loss": 0.4284,
"step": 238
},
{
"epoch": 14.0,
"eval_accuracy": 68.88888888888889,
"eval_average_metrics": 70.04225470037278,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7, \"recall\": 0.6363636363636364, \"f1-score\": 0.6666666666666666, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6511627906976745, \"recall\": 0.7777777777777778, \"f1-score\": 0.7088607594936709, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7241379310344828, \"recall\": 0.5384615384615384, \"f1-score\": 0.6176470588235294, \"support\": 39.0}, \"question\": {\"precision\": 0.7142857142857143, \"recall\": 0.5555555555555556, \"f1-score\": 0.6250000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 0.6923076923076923, \"f1-score\": 0.7826086956521738, \"support\": 13.0}, \"request\": {\"precision\": 0.5384615384615384, \"recall\": 0.875, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.6888888888888889, \"macro avg\": {\"precision\": 0.7628809968099262, \"recall\": 0.7316554972804973, \"f1-score\": 0.7363724073834267, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.711127376440127, \"recall\": 0.6888888888888889, \"f1-score\": 0.6875400028537069, \"support\": 135.0}}",
"eval_f1_macro": 73.63724073834267,
"eval_f1_micro": 68.88888888888889,
"eval_f1_weighted": 68.7540002853707,
"eval_loss": 0.35276439785957336,
"eval_runtime": 2.1486,
"eval_samples_per_second": 62.832,
"step": 238
},
{
"epoch": 15.0,
"learning_rate": 0.000225,
"loss": 0.4125,
"step": 255
},
{
"epoch": 15.0,
"eval_accuracy": 66.66666666666666,
"eval_average_metrics": 67.82947247449044,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.5918367346938775, \"recall\": 0.8055555555555556, \"f1-score\": 0.6823529411764706, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8571428571428571, \"recall\": 0.3076923076923077, \"f1-score\": 0.45283018867924535, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.6153846153846154, \"f1-score\": 0.761904761904762, \"support\": 13.0}, \"request\": {\"precision\": 0.5384615384615384, \"recall\": 0.875, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.6666666666666666, \"macro avg\": {\"precision\": 0.761308929166072, \"recall\": 0.7724237567987569, \"f1-score\": 0.7343863238519761, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7304495152114199, \"recall\": 0.6666666666666666, \"f1-score\": 0.6454592417943081, \"support\": 135.0}}",
"eval_f1_macro": 73.43863238519761,
"eval_f1_micro": 66.66666666666666,
"eval_f1_weighted": 64.54592417943081,
"eval_loss": 0.36039280891418457,
"eval_runtime": 2.0664,
"eval_samples_per_second": 65.332,
"step": 255
},
{
"epoch": 16.0,
"learning_rate": 0.00023999999999999998,
"loss": 0.3537,
"step": 272
},
{
"epoch": 16.0,
"eval_accuracy": 70.37037037037037,
"eval_average_metrics": 71.72736523652416,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.625, \"recall\": 0.9090909090909091, \"f1-score\": 0.7407407407407406, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6521739130434783, \"recall\": 0.8333333333333334, \"f1-score\": 0.7317073170731708, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8095238095238095, \"recall\": 0.4358974358974359, \"f1-score\": 0.5666666666666667, \"support\": 39.0}, \"question\": {\"precision\": 0.8333333333333334, \"recall\": 0.5555555555555556, \"f1-score\": 0.6666666666666667, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.6923076923076923, \"f1-score\": 0.8181818181818181, \"support\": 13.0}, \"request\": {\"precision\": 0.5185185185185185, \"recall\": 0.875, \"f1-score\": 0.6511627906976744, \"support\": 16.0}, \"accuracy\": 0.7037037037037037, \"macro avg\": {\"precision\": 0.8048186968023925, \"recall\": 0.7737592268842269, \"f1-score\": 0.7645378088268715, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7534887832391858, \"recall\": 0.7037037037037037, \"f1-score\": 0.6971493932266877, \"support\": 135.0}}",
"eval_f1_macro": 76.45378088268716,
"eval_f1_micro": 70.37037037037037,
"eval_f1_weighted": 69.71493932266877,
"eval_loss": 0.3415175974369049,
"eval_runtime": 2.2011,
"eval_samples_per_second": 61.334,
"step": 272
},
{
"epoch": 17.0,
"learning_rate": 0.00025499999999999996,
"loss": 0.3307,
"step": 289
},
{
"epoch": 17.0,
"eval_accuracy": 72.5925925925926,
"eval_average_metrics": 73.80907924352218,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6585365853658537, \"recall\": 0.75, \"f1-score\": 0.7012987012987012, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7741935483870968, \"recall\": 0.6153846153846154, \"f1-score\": 0.6857142857142857, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.6875, \"recall\": 0.6875, \"f1-score\": 0.6875, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7720217737121258, \"recall\": 0.7883704836829837, \"f1-score\": 0.7742597755505364, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7391776099879956, \"recall\": 0.725925925925926, \"f1-score\": 0.7262515423384989, \"support\": 135.0}}",
"eval_f1_macro": 77.42597755505363,
"eval_f1_micro": 72.59259259259261,
"eval_f1_weighted": 72.62515423384988,
"eval_loss": 0.320279061794281,
"eval_runtime": 2.1547,
"eval_samples_per_second": 62.654,
"step": 289
},
{
"epoch": 18.0,
"learning_rate": 0.00027,
"loss": 0.2829,
"step": 306
},
{
"epoch": 18.0,
"eval_accuracy": 75.55555555555556,
"eval_average_metrics": 76.66401588989308,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6829268292682927, \"recall\": 0.7777777777777778, \"f1-score\": 0.7272727272727273, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7878787878787878, \"recall\": 0.6666666666666666, \"f1-score\": 0.7222222222222221, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.7857142857142857, \"recall\": 0.6875, \"f1-score\": 0.7333333333333334, \"support\": 16.0}, \"accuracy\": 0.7555555555555555, \"macro avg\": {\"precision\": 0.7981489039415868, \"recall\": 0.807868346930847, \"f1-score\": 0.7989500265816055, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7661238851482753, \"recall\": 0.7555555555555555, \"f1-score\": 0.7564994979030066, \"support\": 135.0}}",
"eval_f1_macro": 79.89500265816055,
"eval_f1_micro": 75.55555555555556,
"eval_f1_weighted": 75.64994979030067,
"eval_loss": 0.3107147812843323,
"eval_runtime": 2.1712,
"eval_samples_per_second": 62.178,
"step": 306
},
{
"epoch": 19.0,
"learning_rate": 0.000285,
"loss": 0.27,
"step": 323
},
{
"epoch": 19.0,
"eval_accuracy": 69.62962962962963,
"eval_average_metrics": 70.65875801590278,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6078431372549019, \"recall\": 0.8611111111111112, \"f1-score\": 0.7126436781609194, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8823529411764706, \"recall\": 0.38461538461538464, \"f1-score\": 0.5357142857142858, \"support\": 39.0}, \"question\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"rejection\": {\"precision\": 0.8461538461538461, \"recall\": 0.8461538461538461, \"f1-score\": 0.8461538461538461, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.7684180606974724, \"recall\": 0.7708758255633257, \"f1-score\": 0.7515478083829743, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7415855386443622, \"recall\": 0.6962962962962963, \"f1-score\": 0.6822099196605441, \"support\": 135.0}}",
"eval_f1_macro": 75.15478083829744,
"eval_f1_micro": 69.62962962962963,
"eval_f1_weighted": 68.22099196605443,
"eval_loss": 0.3932558298110962,
"eval_runtime": 2.1067,
"eval_samples_per_second": 64.082,
"step": 323
},
{
"epoch": 20.0,
"learning_rate": 0.0003,
"loss": 0.2814,
"step": 340
},
{
"epoch": 20.0,
"eval_accuracy": 73.33333333333333,
"eval_average_metrics": 73.63929543961528,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6904761904761905, \"recall\": 0.7435897435897436, \"f1-score\": 0.7160493827160495, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.6875, \"recall\": 0.6875, \"f1-score\": 0.6875, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7865632631257631, \"recall\": 0.731479458041958, \"f1-score\": 0.7439023969088441, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7466218966218966, \"recall\": 0.7333333333333333, \"f1-score\": 0.7350027540091005, \"support\": 135.0}}",
"eval_f1_macro": 74.39023969088441,
"eval_f1_micro": 73.33333333333333,
"eval_f1_weighted": 73.50027540091006,
"eval_loss": 0.35566049814224243,
"eval_runtime": 2.1665,
"eval_samples_per_second": 62.312,
"step": 340
},
{
"epoch": 21.0,
"learning_rate": 0.00029833333333333334,
"loss": 0.242,
"step": 357
},
{
"epoch": 21.0,
"eval_accuracy": 73.33333333333333,
"eval_average_metrics": 74.95524042950514,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6410256410256411, \"f1-score\": 0.6944444444444444, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.631578947368421, \"recall\": 0.75, \"f1-score\": 0.6857142857142857, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7941800524037366, \"recall\": 0.8099504662004662, \"f1-score\": 0.7964169000933707, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7463545226703122, \"recall\": 0.7333333333333333, \"f1-score\": 0.7351260504201681, \"support\": 135.0}}",
"eval_f1_macro": 79.64169000933707,
"eval_f1_micro": 73.33333333333333,
"eval_f1_weighted": 73.51260504201682,
"eval_loss": 0.33881324529647827,
"eval_runtime": 2.177,
"eval_samples_per_second": 62.012,
"step": 357
},
{
"epoch": 22.0,
"learning_rate": 0.00029666666666666665,
"loss": 0.1889,
"step": 374
},
{
"epoch": 22.0,
"eval_accuracy": 73.33333333333333,
"eval_average_metrics": 74.62607689125785,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6511627906976745, \"recall\": 0.7777777777777778, \"f1-score\": 0.7088607594936709, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7857142857142857, \"recall\": 0.5641025641025641, \"f1-score\": 0.6567164179104477, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.75, \"f1-score\": 0.7058823529411765, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7855620155038759, \"recall\": 0.8003350815850816, \"f1-score\": 0.7871122658545654, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7467298306057996, \"recall\": 0.7333333333333333, \"f1-score\": 0.7312641431290826, \"support\": 135.0}}",
"eval_f1_macro": 78.71122658545654,
"eval_f1_micro": 73.33333333333333,
"eval_f1_weighted": 73.12641431290825,
"eval_loss": 0.38223153352737427,
"eval_runtime": 2.1885,
"eval_samples_per_second": 61.685,
"step": 374
},
{
"epoch": 23.0,
"learning_rate": 0.00029499999999999996,
"loss": 0.1787,
"step": 391
},
{
"epoch": 23.0,
"eval_accuracy": 71.85185185185186,
"eval_average_metrics": 73.05134386837818,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6428571428571429, \"recall\": 0.8181818181818182, \"f1-score\": 0.7200000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7105263157894737, \"recall\": 0.75, \"f1-score\": 0.7297297297297298, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7241379310344828, \"recall\": 0.5384615384615384, \"f1-score\": 0.6176470588235294, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6190476190476191, \"recall\": 0.8125, \"f1-score\": 0.7027027027027026, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7613135503335141, \"recall\": 0.7901065947940948, \"f1-score\": 0.7691766030736619, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7284590496326734, \"recall\": 0.7185185185185186, \"f1-score\": 0.7158401146244283, \"support\": 135.0}}",
"eval_f1_macro": 76.91766030736619,
"eval_f1_micro": 71.85185185185186,
"eval_f1_weighted": 71.58401146244283,
"eval_loss": 0.40378537774086,
"eval_runtime": 2.2888,
"eval_samples_per_second": 58.982,
"step": 391
},
{
"epoch": 24.0,
"learning_rate": 0.00029333333333333327,
"loss": 0.1458,
"step": 408
},
{
"epoch": 24.0,
"eval_accuracy": 70.37037037037037,
"eval_average_metrics": 71.88129852922421,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.6111111111111112, \"f1-score\": 0.6470588235294118, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8, \"recall\": 0.5128205128205128, \"f1-score\": 0.625, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 0.9230769230769231, \"f1-score\": 0.8275862068965517, \"support\": 13.0}, \"request\": {\"precision\": 0.5384615384615384, \"recall\": 0.875, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7037037037037037, \"macro avg\": {\"precision\": 0.7539396367521367, \"recall\": 0.8122207653457654, \"f1-score\": 0.7701591825594307, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7255460588793923, \"recall\": 0.7037037037037037, \"f1-score\": 0.6976853512021299, \"support\": 135.0}}",
"eval_f1_macro": 77.01591825594308,
"eval_f1_micro": 70.37037037037037,
"eval_f1_weighted": 69.768535120213,
"eval_loss": 0.4484809637069702,
"eval_runtime": 2.2039,
"eval_samples_per_second": 61.256,
"step": 408
},
{
"epoch": 25.0,
"learning_rate": 0.00029166666666666664,
"loss": 0.1755,
"step": 425
},
{
"epoch": 25.0,
"eval_accuracy": 73.33333333333333,
"eval_average_metrics": 73.30296325475025,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.7272727272727273, \"f1-score\": 0.6956521739130435, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6595744680851063, \"recall\": 0.8611111111111112, \"f1-score\": 0.7469879518072289, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8461538461538461, \"recall\": 0.5641025641025641, \"f1-score\": 0.676923076923077, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.75, \"f1-score\": 0.7058823529411765, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7791251301889599, \"recall\": 0.7255244755244756, \"f1-score\": 0.7344306944480657, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7577047066408767, \"recall\": 0.7333333333333333, \"f1-score\": 0.7310211690752781, \"support\": 135.0}}",
"eval_f1_macro": 73.44306944480657,
"eval_f1_micro": 73.33333333333333,
"eval_f1_weighted": 73.1021169075278,
"eval_loss": 0.47192898392677307,
"eval_runtime": 2.1604,
"eval_samples_per_second": 62.489,
"step": 425
},
{
"epoch": 26.0,
"learning_rate": 0.00029,
"loss": 0.1193,
"step": 442
},
{
"epoch": 26.0,
"eval_accuracy": 73.33333333333333,
"eval_average_metrics": 74.94665416513004,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6578947368421053, \"recall\": 0.6944444444444444, \"f1-score\": 0.6756756756756757, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6410256410256411, \"f1-score\": 0.6944444444444444, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.8037195260879472, \"recall\": 0.7995337995337994, \"f1-score\": 0.7956619448395763, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7480504291030607, \"recall\": 0.7333333333333333, \"f1-score\": 0.7355375550989586, \"support\": 135.0}}",
"eval_f1_macro": 79.56619448395764,
"eval_f1_micro": 73.33333333333333,
"eval_f1_weighted": 73.55375550989585,
"eval_loss": 0.45392414927482605,
"eval_runtime": 2.2435,
"eval_samples_per_second": 60.174,
"step": 442
},
{
"epoch": 27.0,
"learning_rate": 0.0002883333333333333,
"loss": 0.1263,
"step": 459
},
{
"epoch": 27.0,
"eval_accuracy": 72.5925925925926,
"eval_average_metrics": 74.0950271183071,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.65, \"recall\": 0.7222222222222222, \"f1-score\": 0.6842105263157895, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8214285714285714, \"recall\": 0.5897435897435898, \"f1-score\": 0.6865671641791046, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5454545454545454, \"recall\": 0.75, \"f1-score\": 0.631578947368421, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7858405483405484, \"recall\": 0.7965957653457654, \"f1-score\": 0.7843223575106921, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7482972582972582, \"recall\": 0.725925925925926, \"f1-score\": 0.7276268753697399, \"support\": 135.0}}",
"eval_f1_macro": 78.43223575106921,
"eval_f1_micro": 72.59259259259261,
"eval_f1_weighted": 72.762687536974,
"eval_loss": 0.46841171383857727,
"eval_runtime": 2.1169,
"eval_samples_per_second": 63.773,
"step": 459
},
{
"epoch": 28.0,
"learning_rate": 0.0002866666666666667,
"loss": 0.1043,
"step": 476
},
{
"epoch": 28.0,
"eval_accuracy": 68.88888888888889,
"eval_average_metrics": 69.42866207163016,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.625, \"recall\": 0.9090909090909091, \"f1-score\": 0.7407407407407406, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6571428571428571, \"recall\": 0.6388888888888888, \"f1-score\": 0.647887323943662, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8333333333333334, \"recall\": 0.5128205128205128, \"f1-score\": 0.6349206349206349, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.52, \"recall\": 0.8125, \"f1-score\": 0.6341463414634146, \"support\": 16.0}, \"accuracy\": 0.6888888888888889, \"macro avg\": {\"precision\": 0.7394724025974025, \"recall\": 0.7357651029526029, \"f1-score\": 0.7130452134668899, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7225299021965689, \"recall\": 0.6888888888888889, \"f1-score\": 0.6863234916205385, \"support\": 135.0}}",
"eval_f1_macro": 71.30452134668899,
"eval_f1_micro": 68.88888888888889,
"eval_f1_weighted": 68.63234916205386,
"eval_loss": 0.49356982111930847,
"eval_runtime": 2.0925,
"eval_samples_per_second": 64.518,
"step": 476
},
{
"epoch": 29.0,
"learning_rate": 0.000285,
"loss": 0.1174,
"step": 493
},
{
"epoch": 29.0,
"eval_accuracy": 68.14814814814815,
"eval_average_metrics": 69.89790682417316,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6206896551724138, \"recall\": 0.5, \"f1-score\": 0.5538461538461539, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7741935483870968, \"recall\": 0.6153846153846154, \"f1-score\": 0.6857142857142857, \"support\": 39.0}, \"question\": {\"precision\": 0.5454545454545454, \"recall\": 0.6666666666666666, \"f1-score\": 0.6, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.4827586206896552, \"recall\": 0.875, \"f1-score\": 0.6222222222222222, \"support\": 16.0}, \"accuracy\": 0.6814814814814815, \"macro avg\": {\"precision\": 0.7515477604986782, \"recall\": 0.7737592268842268, \"f1-score\": 0.7502473425267543, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7103981308931253, \"recall\": 0.6814814814814815, \"f1-score\": 0.6827059674772094, \"support\": 135.0}}",
"eval_f1_macro": 75.02473425267543,
"eval_f1_micro": 68.14814814814815,
"eval_f1_weighted": 68.27059674772093,
"eval_loss": 0.5587795972824097,
"eval_runtime": 2.2187,
"eval_samples_per_second": 60.847,
"step": 493
},
{
"epoch": 30.0,
"learning_rate": 0.0002833333333333333,
"loss": 0.0689,
"step": 510
},
{
"epoch": 30.0,
"eval_accuracy": 72.5925925925926,
"eval_average_metrics": 73.94861619800345,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6585365853658537, \"recall\": 0.75, \"f1-score\": 0.7012987012987012, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7931034482758621, \"recall\": 0.5897435897435898, \"f1-score\": 0.676470588235294, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.75, \"f1-score\": 0.6486486486486486, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7848776232528335, \"recall\": 0.7861790986790986, \"f1-score\": 0.7791600854100853, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7445433448587359, \"recall\": 0.725925925925926, \"f1-score\": 0.7269327106582008, \"support\": 135.0}}",
"eval_f1_macro": 77.91600854100852,
"eval_f1_micro": 72.59259259259261,
"eval_f1_weighted": 72.69327106582007,
"eval_loss": 0.539673388004303,
"eval_runtime": 2.1316,
"eval_samples_per_second": 63.333,
"step": 510
},
{
"epoch": 31.0,
"learning_rate": 0.00028166666666666666,
"loss": 0.0539,
"step": 527
},
{
"epoch": 31.0,
"eval_accuracy": 69.62962962962963,
"eval_average_metrics": 70.33781120741372,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 1.0, \"f1-score\": 0.8571428571428571, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.6111111111111112, \"f1-score\": 0.6376811594202899, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7096774193548387, \"recall\": 0.5641025641025641, \"f1-score\": 0.6285714285714286, \"support\": 39.0}, \"question\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.7561203917050692, \"recall\": 0.7387031371406372, \"f1-score\": 0.7282430580928023, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7082807077430734, \"recall\": 0.6962962962962963, \"f1-score\": 0.6926767976111539, \"support\": 135.0}}",
"eval_f1_macro": 72.82430580928023,
"eval_f1_micro": 69.62962962962963,
"eval_f1_weighted": 69.26767976111539,
"eval_loss": 0.5683187246322632,
"eval_runtime": 2.1767,
"eval_samples_per_second": 62.021,
"step": 527
},
{
"epoch": 32.0,
"learning_rate": 0.00028,
"loss": 0.0591,
"step": 544
},
{
"epoch": 32.0,
"eval_accuracy": 70.37037037037037,
"eval_average_metrics": 71.58032751159402,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7027027027027027, \"recall\": 0.7222222222222222, \"f1-score\": 0.7123287671232876, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7916666666666666, \"recall\": 0.48717948717948717, \"f1-score\": 0.6031746031746031, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 0.8461538461538461, \"recall\": 0.8461538461538461, \"f1-score\": 0.8461538461538461, \"support\": 13.0}, \"request\": {\"precision\": 0.5454545454545454, \"recall\": 0.75, \"f1-score\": 0.631578947368421, \"support\": 16.0}, \"accuracy\": 0.7037037037037037, \"macro avg\": {\"precision\": 0.738858691983692, \"recall\": 0.7976641414141414, \"f1-score\": 0.757487853810853, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7226086476086476, \"recall\": 0.7037037037037037, \"f1-score\": 0.6983178392455008, \"support\": 135.0}}",
"eval_f1_macro": 75.7487853810853,
"eval_f1_micro": 70.37037037037037,
"eval_f1_weighted": 69.83178392455008,
"eval_loss": 0.6107626557350159,
"eval_runtime": 2.1901,
"eval_samples_per_second": 61.641,
"step": 544
},
{
"epoch": 33.0,
"learning_rate": 0.00027833333333333334,
"loss": 0.0539,
"step": 561
},
{
"epoch": 33.0,
"eval_accuracy": 73.33333333333333,
"eval_average_metrics": 73.67097529451648,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7272727272727273, \"recall\": 0.6666666666666666, \"f1-score\": 0.6956521739130435, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7297297297297297, \"recall\": 0.6923076923076923, \"f1-score\": 0.7105263157894737, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7804387991887991, \"recall\": 0.7399718337218337, \"f1-score\": 0.7468834223239258, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7408524975191642, \"recall\": 0.7333333333333333, \"f1-score\": 0.7332889227900669, \"support\": 135.0}}",
"eval_f1_macro": 74.68834223239257,
"eval_f1_micro": 73.33333333333333,
"eval_f1_weighted": 73.32889227900668,
"eval_loss": 0.5886625051498413,
"eval_runtime": 2.2726,
"eval_samples_per_second": 59.404,
"step": 561
},
{
"epoch": 34.0,
"learning_rate": 0.00027666666666666665,
"loss": 0.0734,
"step": 578
},
{
"epoch": 34.0,
"eval_accuracy": 74.81481481481481,
"eval_average_metrics": 75.0892478093565,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7575757575757576, \"recall\": 0.6944444444444444, \"f1-score\": 0.7246376811594203, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7368421052631579, \"recall\": 0.717948717948718, \"f1-score\": 0.7272727272727273, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.631578947368421, \"recall\": 0.75, \"f1-score\": 0.6857142857142857, \"support\": 16.0}, \"accuracy\": 0.7481481481481481, \"macro avg\": {\"precision\": 0.7929446284709443, \"recall\": 0.7491744366744366, \"f1-score\": 0.757480864546082, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7589443694706852, \"recall\": 0.7481481481481481, \"f1-score\": 0.749792751531882, \"support\": 135.0}}",
"eval_f1_macro": 75.7480864546082,
"eval_f1_micro": 74.81481481481481,
"eval_f1_weighted": 74.9792751531882,
"eval_loss": 0.5679999589920044,
"eval_runtime": 2.2089,
"eval_samples_per_second": 61.116,
"step": 578
},
{
"epoch": 35.0,
"learning_rate": 0.00027499999999999996,
"loss": 0.0343,
"step": 595
},
{
"epoch": 35.0,
"eval_accuracy": 71.85185185185186,
"eval_average_metrics": 72.3823414203499,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6585365853658537, \"recall\": 0.75, \"f1-score\": 0.7012987012987012, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7857142857142857, \"recall\": 0.5641025641025641, \"f1-score\": 0.6567164179104477, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 0.8461538461538461, \"recall\": 0.8461538461538461, \"f1-score\": 0.8461538461538461, \"support\": 13.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.75, \"f1-score\": 0.6486486486486486, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7803482087018672, \"recall\": 0.7343628593628594, \"f1-score\": 0.7409159272416515, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7359272164150212, \"recall\": 0.7185185185185186, \"f1-score\": 0.7173406925353076, \"support\": 135.0}}",
"eval_f1_macro": 74.09159272416514,
"eval_f1_micro": 71.85185185185186,
"eval_f1_weighted": 71.73406925353078,
"eval_loss": 0.64279705286026,
"eval_runtime": 2.2258,
"eval_samples_per_second": 60.654,
"step": 595
},
{
"epoch": 36.0,
"learning_rate": 0.00027333333333333333,
"loss": 0.0256,
"step": 612
},
{
"epoch": 36.0,
"eval_accuracy": 70.37037037037037,
"eval_average_metrics": 71.27192027743499,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6470588235294118, \"recall\": 0.6111111111111112, \"f1-score\": 0.6285714285714287, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7741935483870968, \"recall\": 0.6153846153846154, \"f1-score\": 0.6857142857142857, \"support\": 39.0}, \"question\": {\"precision\": 0.5384615384615384, \"recall\": 0.7777777777777778, \"f1-score\": 0.6363636363636364, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5217391304347826, \"recall\": 0.75, \"f1-score\": 0.6153846153846153, \"support\": 16.0}, \"accuracy\": 0.7037037037037037, \"macro avg\": {\"precision\": 0.774467344387318, \"recall\": 0.7373008935508936, \"f1-score\": 0.7363179712444419, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7299169542548802, \"recall\": 0.7037037037037037, \"f1-score\": 0.7071514324455502, \"support\": 135.0}}",
"eval_f1_macro": 73.63179712444419,
"eval_f1_micro": 70.37037037037037,
"eval_f1_weighted": 70.71514324455502,
"eval_loss": 0.6609192490577698,
"eval_runtime": 2.2472,
"eval_samples_per_second": 60.074,
"step": 612
},
{
"epoch": 37.0,
"learning_rate": 0.00027166666666666664,
"loss": 0.0369,
"step": 629
},
{
"epoch": 37.0,
"eval_accuracy": 73.33333333333333,
"eval_average_metrics": 74.77168024471725,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7297297297297297, \"recall\": 0.6923076923076923, \"f1-score\": 0.7105263157894737, \"support\": 39.0}, \"question\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7861613248710023, \"recall\": 0.8058651418026419, \"f1-score\": 0.7918380042511899, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7395884688357807, \"recall\": 0.7333333333333333, \"f1-score\": 0.7323625388708334, \"support\": 135.0}}",
"eval_f1_macro": 79.18380042511899,
"eval_f1_micro": 73.33333333333333,
"eval_f1_weighted": 73.23625388708334,
"eval_loss": 0.6587929725646973,
"eval_runtime": 2.1685,
"eval_samples_per_second": 62.254,
"step": 629
},
{
"epoch": 38.0,
"learning_rate": 0.00027,
"loss": 0.0411,
"step": 646
},
{
"epoch": 38.0,
"eval_accuracy": 71.85185185185186,
"eval_average_metrics": 72.4984910288592,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.71875, \"recall\": 0.6388888888888888, \"f1-score\": 0.676470588235294, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7916610299422799, \"recall\": 0.7314915986790986, \"f1-score\": 0.741379475703325, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7429092512425847, \"recall\": 0.7185185185185186, \"f1-score\": 0.7215231284140066, \"support\": 135.0}}",
"eval_f1_macro": 74.13794757033249,
"eval_f1_micro": 71.85185185185186,
"eval_f1_weighted": 72.15231284140066,
"eval_loss": 0.717117190361023,
"eval_runtime": 2.2579,
"eval_samples_per_second": 59.789,
"step": 646
},
{
"epoch": 39.0,
"learning_rate": 0.0002683333333333333,
"loss": 0.0188,
"step": 663
},
{
"epoch": 39.0,
"eval_accuracy": 74.81481481481481,
"eval_average_metrics": 75.21675953993977,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7222222222222222, \"recall\": 0.7222222222222222, \"f1-score\": 0.7222222222222222, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.6666666666666666, \"f1-score\": 0.7123287671232877, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.631578947368421, \"recall\": 0.75, \"f1-score\": 0.6857142857142857, \"support\": 16.0}, \"accuracy\": 0.7481481481481481, \"macro avg\": {\"precision\": 0.7982278836574502, \"recall\": 0.7576000388500388, \"f1-score\": 0.7635135515393372, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7587398134869756, \"recall\": 0.7481481481481481, \"f1-score\": 0.7488605337619573, \"support\": 135.0}}",
"eval_f1_macro": 76.35135515393372,
"eval_f1_micro": 74.81481481481481,
"eval_f1_weighted": 74.88605337619573,
"eval_loss": 0.6625694036483765,
"eval_runtime": 2.1219,
"eval_samples_per_second": 63.622,
"step": 663
},
{
"epoch": 40.0,
"learning_rate": 0.0002666666666666666,
"loss": 0.0232,
"step": 680
},
{
"epoch": 40.0,
"eval_accuracy": 74.07407407407408,
"eval_average_metrics": 75.41849073937567,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7352941176470589, \"recall\": 0.6944444444444444, \"f1-score\": 0.7142857142857144, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.631578947368421, \"recall\": 0.75, \"f1-score\": 0.6857142857142857, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.7912581597759617, \"recall\": 0.8052641802641802, \"f1-score\": 0.7945949432404541, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7460020576119648, \"recall\": 0.7407407407407407, \"f1-score\": 0.7406632048530915, \"support\": 135.0}}",
"eval_f1_macro": 79.45949432404542,
"eval_f1_micro": 74.07407407407408,
"eval_f1_weighted": 74.06632048530913,
"eval_loss": 0.7407746315002441,
"eval_runtime": 2.1843,
"eval_samples_per_second": 61.805,
"step": 680
},
{
"epoch": 41.0,
"learning_rate": 0.000265,
"loss": 0.034,
"step": 697
},
{
"epoch": 41.0,
"eval_accuracy": 70.37037037037037,
"eval_average_metrics": 72.10801080339358,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6285714285714286, \"recall\": 0.6111111111111112, \"f1-score\": 0.619718309859155, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8214285714285714, \"recall\": 0.5897435897435898, \"f1-score\": 0.6865671641791046, \"support\": 39.0}, \"question\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.8125, \"f1-score\": 0.6190476190476191, \"support\": 16.0}, \"accuracy\": 0.7037037037037037, \"macro avg\": {\"precision\": 0.7691468253968254, \"recall\": 0.7905193764568765, \"f1-score\": 0.7703084495012319, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7327513227513227, \"recall\": 0.7037037037037037, \"f1-score\": 0.7066045752271038, \"support\": 135.0}}",
"eval_f1_macro": 77.0308449501232,
"eval_f1_micro": 70.37037037037037,
"eval_f1_weighted": 70.66045752271039,
"eval_loss": 0.8014240264892578,
"eval_runtime": 2.1576,
"eval_samples_per_second": 62.569,
"step": 697
},
{
"epoch": 42.0,
"learning_rate": 0.0002633333333333333,
"loss": 0.0393,
"step": 714
},
{
"epoch": 42.0,
"eval_accuracy": 72.5925925925926,
"eval_average_metrics": 72.91137239666652,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.8181818181818182, \"recall\": 0.8181818181818182, \"f1-score\": 0.8181818181818182, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6428571428571429, \"recall\": 0.75, \"f1-score\": 0.6923076923076924, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7878787878787878, \"recall\": 0.6666666666666666, \"f1-score\": 0.7222222222222221, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9090909090909091, \"recall\": 0.7692307692307693, \"f1-score\": 0.8333333333333333, \"support\": 13.0}, \"request\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7874594155844157, \"recall\": 0.7123154623154623, \"f1-score\": 0.7361968097262215, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.741950456950457, \"recall\": 0.725925925925926, \"f1-score\": 0.7284062342885872, \"support\": 135.0}}",
"eval_f1_macro": 73.61968097262215,
"eval_f1_micro": 72.59259259259261,
"eval_f1_weighted": 72.84062342885873,
"eval_loss": 0.7516341805458069,
"eval_runtime": 2.2389,
"eval_samples_per_second": 60.297,
"step": 714
},
{
"epoch": 43.0,
"learning_rate": 0.00026166666666666667,
"loss": 0.0187,
"step": 731
},
{
"epoch": 43.0,
"eval_accuracy": 69.62962962962963,
"eval_average_metrics": 71.6853552122398,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6363636363636364, \"recall\": 0.5833333333333334, \"f1-score\": 0.6086956521739131, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7419354838709677, \"recall\": 0.5897435897435898, \"f1-score\": 0.6571428571428573, \"support\": 39.0}, \"question\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.48148148148148145, \"recall\": 0.8125, \"f1-score\": 0.6046511627906976, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.784555908547844, \"recall\": 0.7870471542346542, \"f1-score\": 0.7754307414417895, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7231971245710792, \"recall\": 0.6962962962962963, \"f1-score\": 0.6993908744552095, \"support\": 135.0}}",
"eval_f1_macro": 77.54307414417895,
"eval_f1_micro": 69.62962962962963,
"eval_f1_weighted": 69.93908744552097,
"eval_loss": 0.8274635672569275,
"eval_runtime": 2.2726,
"eval_samples_per_second": 59.402,
"step": 731
},
{
"epoch": 44.0,
"learning_rate": 0.00026,
"loss": 0.0326,
"step": 748
},
{
"epoch": 44.0,
"eval_accuracy": 71.85185185185186,
"eval_average_metrics": 73.29350182491167,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.71875, \"recall\": 0.6388888888888888, \"f1-score\": 0.676470588235294, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6923076923076923, \"recall\": 0.6923076923076923, \"f1-score\": 0.6923076923076923, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.75, \"f1-score\": 0.6486486486486486, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.778793688949939, \"recall\": 0.7780205905205906, \"f1-score\": 0.7740684627673118, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7306166056166056, \"recall\": 0.7185185185185186, \"f1-score\": 0.7206345731921179, \"support\": 135.0}}",
"eval_f1_macro": 77.40684627673117,
"eval_f1_micro": 71.85185185185186,
"eval_f1_weighted": 72.06345731921179,
"eval_loss": 0.7537038326263428,
"eval_runtime": 2.251,
"eval_samples_per_second": 59.972,
"step": 748
},
{
"epoch": 45.0,
"learning_rate": 0.00025833333333333334,
"loss": 0.0446,
"step": 765
},
{
"epoch": 45.0,
"eval_accuracy": 73.33333333333333,
"eval_average_metrics": 74.55936282460405,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6428571428571429, \"recall\": 0.8181818181818182, \"f1-score\": 0.7200000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7352941176470589, \"recall\": 0.6944444444444444, \"f1-score\": 0.7142857142857144, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.6666666666666666, \"f1-score\": 0.7027027027027027, \"support\": 39.0}, \"question\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7762371615312792, \"recall\": 0.7913752913752914, \"f1-score\": 0.7800986983223825, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7445440398381575, \"recall\": 0.7333333333333333, \"f1-score\": 0.735609147995113, \"support\": 135.0}}",
"eval_f1_macro": 78.00986983223825,
"eval_f1_micro": 73.33333333333333,
"eval_f1_weighted": 73.5609147995113,
"eval_loss": 0.7628147006034851,
"eval_runtime": 2.1831,
"eval_samples_per_second": 61.838,
"step": 765
},
{
"epoch": 46.0,
"learning_rate": 0.00025666666666666665,
"loss": 0.0222,
"step": 782
},
{
"epoch": 46.0,
"eval_accuracy": 71.85185185185186,
"eval_average_metrics": 73.61743270582245,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.5833333333333334, \"f1-score\": 0.6461538461538462, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7352941176470589, \"recall\": 0.6410256410256411, \"f1-score\": 0.6849315068493151, \"support\": 39.0}, \"question\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.8125, \"f1-score\": 0.6190476190476191, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7865956727518593, \"recall\": 0.8073462995337995, \"f1-score\": 0.7881557617778407, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7388550822627902, \"recall\": 0.7185185185185186, \"f1-score\": 0.71950450941802, \"support\": 135.0}}",
"eval_f1_macro": 78.81557617778408,
"eval_f1_micro": 71.85185185185186,
"eval_f1_weighted": 71.950450941802,
"eval_loss": 0.8409448266029358,
"eval_runtime": 2.2893,
"eval_samples_per_second": 58.969,
"step": 782
},
{
"epoch": 47.0,
"learning_rate": 0.00025499999999999996,
"loss": 0.0137,
"step": 799
},
{
"epoch": 47.0,
"eval_accuracy": 71.11111111111111,
"eval_average_metrics": 72.17480498243586,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6756756756756757, \"recall\": 0.6944444444444444, \"f1-score\": 0.684931506849315, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8064516129032258, \"recall\": 0.6410256410256411, \"f1-score\": 0.7142857142857142, \"support\": 39.0}, \"question\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"rejection\": {\"precision\": 0.8461538461538461, \"recall\": 0.8461538461538461, \"f1-score\": 0.8461538461538461, \"support\": 13.0}, \"request\": {\"precision\": 0.55, \"recall\": 0.6875, \"f1-score\": 0.6111111111111112, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7435458256022771, \"recall\": 0.7664687742812744, \"f1-score\": 0.7519798629432732, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7228986803180352, \"recall\": 0.7111111111111111, \"f1-score\": 0.712790114131939, \"support\": 135.0}}",
"eval_f1_macro": 75.19798629432732,
"eval_f1_micro": 71.11111111111111,
"eval_f1_weighted": 71.27901141319391,
"eval_loss": 0.8047698140144348,
"eval_runtime": 2.1643,
"eval_samples_per_second": 62.375,
"step": 799
},
{
"epoch": 48.0,
"learning_rate": 0.00025333333333333333,
"loss": 0.0209,
"step": 816
},
{
"epoch": 48.0,
"eval_accuracy": 71.85185185185186,
"eval_average_metrics": 73.5841574785611,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.6111111111111112, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7352941176470589, \"recall\": 0.6410256410256411, \"f1-score\": 0.6849315068493151, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.48, \"recall\": 0.75, \"f1-score\": 0.5853658536585366, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.78078081232493, \"recall\": 0.8030060217560218, \"f1-score\": 0.7854538367301481, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7372613341632949, \"recall\": 0.7185185185185186, \"f1-score\": 0.7208754253752583, \"support\": 135.0}}",
"eval_f1_macro": 78.54538367301481,
"eval_f1_micro": 71.85185185185186,
"eval_f1_weighted": 72.08754253752583,
"eval_loss": 0.8688657283782959,
"eval_runtime": 2.1847,
"eval_samples_per_second": 61.794,
"step": 816
},
{
"epoch": 49.0,
"learning_rate": 0.00025166666666666664,
"loss": 0.0091,
"step": 833
},
{
"epoch": 49.0,
"eval_accuracy": 73.33333333333333,
"eval_average_metrics": 74.75127888430008,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8125, \"recall\": 0.6666666666666666, \"f1-score\": 0.7323943661971831, \"support\": 39.0}, \"question\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.75, \"f1-score\": 0.6486486486486486, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7851934523809524, \"recall\": 0.7992667055167055, \"f1-score\": 0.7875581374336444, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.748179012345679, \"recall\": 0.7333333333333333, \"f1-score\": 0.735826351271692, \"support\": 135.0}}",
"eval_f1_macro": 78.75581374336443,
"eval_f1_micro": 73.33333333333333,
"eval_f1_weighted": 73.58263512716921,
"eval_loss": 0.8018043637275696,
"eval_runtime": 2.2936,
"eval_samples_per_second": 58.859,
"step": 833
},
{
"epoch": 50.0,
"learning_rate": 0.00025,
"loss": 0.0097,
"step": 850
},
{
"epoch": 50.0,
"eval_accuracy": 74.07407407407408,
"eval_average_metrics": 75.57898288245715,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.8181818181818182, \"f1-score\": 0.7826086956521738, \"support\": 11.0}, \"challenge\": {\"precision\": 0.65, \"recall\": 0.7222222222222222, \"f1-score\": 0.6842105263157895, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8387096774193549, \"recall\": 0.6666666666666666, \"f1-score\": 0.7428571428571428, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5454545454545454, \"recall\": 0.75, \"f1-score\": 0.631578947368421, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.8063538611925709, \"recall\": 0.7948475135975136, \"f1-score\": 0.795720639514387, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7636070381231672, \"recall\": 0.7407407407407407, \"f1-score\": 0.7459571943024179, \"support\": 135.0}}",
"eval_f1_macro": 79.5720639514387,
"eval_f1_micro": 74.07407407407408,
"eval_f1_weighted": 74.59571943024179,
"eval_loss": 0.8517288565635681,
"eval_runtime": 2.1756,
"eval_samples_per_second": 62.051,
"step": 850
},
{
"epoch": 51.0,
"learning_rate": 0.0002483333333333333,
"loss": 0.0193,
"step": 867
},
{
"epoch": 51.0,
"eval_accuracy": 71.11111111111111,
"eval_average_metrics": 72.8732174399158,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.625, \"recall\": 0.9090909090909091, \"f1-score\": 0.7407407407407406, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.5833333333333334, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7058823529411765, \"recall\": 0.6153846153846154, \"f1-score\": 0.6575342465753424, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.75, \"f1-score\": 0.6, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7702113042186571, \"recall\": 0.8102175602175602, \"f1-score\": 0.7806177067478437, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7336914900640391, \"recall\": 0.7111111111111111, \"f1-score\": 0.7120887686265667, \"support\": 135.0}}",
"eval_f1_macro": 78.06177067478437,
"eval_f1_micro": 71.11111111111111,
"eval_f1_weighted": 71.20887686265667,
"eval_loss": 0.9536928534507751,
"eval_runtime": 2.1272,
"eval_samples_per_second": 63.463,
"step": 867
},
{
"epoch": 52.0,
"learning_rate": 0.0002466666666666666,
"loss": 0.0046,
"step": 884
},
{
"epoch": 52.0,
"eval_accuracy": 71.11111111111111,
"eval_average_metrics": 71.64170538878341,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.8181818181818182, \"f1-score\": 0.7826086956521738, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6829268292682927, \"recall\": 0.717948717948718, \"f1-score\": 0.7000000000000001, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.55, \"recall\": 0.6875, \"f1-score\": 0.6111111111111112, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7706033088556692, \"recall\": 0.7170563811188811, \"f1-score\": 0.7315823058619944, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7196237044807725, \"recall\": 0.7111111111111111, \"f1-score\": 0.7118636874671196, \"support\": 135.0}}",
"eval_f1_macro": 73.15823058619944,
"eval_f1_micro": 71.11111111111111,
"eval_f1_weighted": 71.18636874671196,
"eval_loss": 0.8713467121124268,
"eval_runtime": 2.2641,
"eval_samples_per_second": 59.627,
"step": 884
},
{
"epoch": 53.0,
"learning_rate": 0.000245,
"loss": 0.0085,
"step": 901
},
{
"epoch": 53.0,
"eval_accuracy": 68.88888888888889,
"eval_average_metrics": 69.60698215446021,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7037037037037037, \"recall\": 0.5277777777777778, \"f1-score\": 0.6031746031746033, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.75, \"f1-score\": 0.6, \"support\": 16.0}, \"accuracy\": 0.6888888888888889, \"macro avg\": {\"precision\": 0.7484503734503735, \"recall\": 0.721930846930847, \"f1-score\": 0.7174509189640768, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7075614075614075, \"recall\": 0.6888888888888889, \"f1-score\": 0.6890505894365544, \"support\": 135.0}}",
"eval_f1_macro": 71.74509189640769,
"eval_f1_micro": 68.88888888888889,
"eval_f1_weighted": 68.90505894365543,
"eval_loss": 0.9699578285217285,
"eval_runtime": 2.2097,
"eval_samples_per_second": 61.095,
"step": 901
},
{
"epoch": 54.0,
"learning_rate": 0.0002433333333333333,
"loss": 0.0181,
"step": 918
},
{
"epoch": 54.0,
"eval_accuracy": 71.11111111111111,
"eval_average_metrics": 71.51024557847745,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7, \"recall\": 0.6363636363636364, \"f1-score\": 0.6666666666666666, \"support\": 11.0}, \"challenge\": {\"precision\": 0.696969696969697, \"recall\": 0.6388888888888888, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7, \"recall\": 0.717948717948718, \"f1-score\": 0.708860759493671, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.75, \"f1-score\": 0.6486486486486486, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7724702380952382, \"recall\": 0.7056138306138307, \"f1-score\": 0.7247131858217115, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7247113997113998, \"recall\": 0.7111111111111111, \"f1-score\": 0.7134744150951646, \"support\": 135.0}}",
"eval_f1_macro": 72.47131858217115,
"eval_f1_micro": 71.11111111111111,
"eval_f1_weighted": 71.34744150951646,
"eval_loss": 0.9296635389328003,
"eval_runtime": 2.1343,
"eval_samples_per_second": 63.252,
"step": 918
},
{
"epoch": 55.0,
"learning_rate": 0.00024166666666666664,
"loss": 0.0055,
"step": 935
},
{
"epoch": 55.0,
"eval_accuracy": 68.14814814814815,
"eval_average_metrics": 69.88210719785789,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6451612903225806, \"recall\": 0.5555555555555556, \"f1-score\": 0.5970149253731343, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7857142857142857, \"recall\": 0.5641025641025641, \"f1-score\": 0.6567164179104477, \"support\": 39.0}, \"question\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"rejection\": {\"precision\": 0.7857142857142857, \"recall\": 0.8461538461538461, \"f1-score\": 0.8148148148148148, \"support\": 13.0}, \"request\": {\"precision\": 0.4642857142857143, \"recall\": 0.8125, \"f1-score\": 0.5909090909090908, \"support\": 16.0}, \"accuracy\": 0.6814814814814815, \"macro avg\": {\"precision\": 0.7426706358157971, \"recall\": 0.7828950563325563, \"f1-score\": 0.751917200243583, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7054854941951716, \"recall\": 0.6814814814814815, \"f1-score\": 0.6804041247077698, \"support\": 135.0}}",
"eval_f1_macro": 75.1917200243583,
"eval_f1_micro": 68.14814814814815,
"eval_f1_weighted": 68.04041247077699,
"eval_loss": 0.9763553142547607,
"eval_runtime": 2.1762,
"eval_samples_per_second": 62.035,
"step": 935
},
{
"epoch": 56.0,
"learning_rate": 0.00023999999999999998,
"loss": 0.0051,
"step": 952
},
{
"epoch": 56.0,
"eval_accuracy": 71.85185185185186,
"eval_average_metrics": 73.26569664902999,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.5555555555555556, \"f1-score\": 0.634920634920635, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6923076923076923, \"f1-score\": 0.7199999999999999, \"support\": 39.0}, \"question\": {\"precision\": 0.5, \"recall\": 0.7777777777777778, \"f1-score\": 0.6086956521739131, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7671443391552086, \"recall\": 0.7989206973581974, \"f1-score\": 0.7732298136645963, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7413002188364506, \"recall\": 0.7185185185185186, \"f1-score\": 0.720361015259566, \"support\": 135.0}}",
"eval_f1_macro": 77.32298136645963,
"eval_f1_micro": 71.85185185185186,
"eval_f1_weighted": 72.0361015259566,
"eval_loss": 0.9423373341560364,
"eval_runtime": 2.238,
"eval_samples_per_second": 60.322,
"step": 952
},
{
"epoch": 57.0,
"learning_rate": 0.0002383333333333333,
"loss": 0.0114,
"step": 969
},
{
"epoch": 57.0,
"eval_accuracy": 71.11111111111111,
"eval_average_metrics": 72.53354760332923,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 1.0, \"f1-score\": 0.8571428571428571, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.6363636363636364, \"f1-score\": 0.6363636363636364, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6764705882352942, \"recall\": 0.6388888888888888, \"f1-score\": 0.6571428571428571, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6153846153846154, \"f1-score\": 0.676056338028169, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7597564519879099, \"recall\": 0.7908835955710956, \"f1-score\": 0.7683601409092514, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7236771810173345, \"recall\": 0.7111111111111111, \"f1-score\": 0.7107595410016957, \"support\": 135.0}}",
"eval_f1_macro": 76.83601409092513,
"eval_f1_micro": 71.11111111111111,
"eval_f1_weighted": 71.07595410016957,
"eval_loss": 0.9203009605407715,
"eval_runtime": 2.1386,
"eval_samples_per_second": 63.126,
"step": 969
},
{
"epoch": 58.0,
"learning_rate": 0.00023666666666666663,
"loss": 0.0072,
"step": 986
},
{
"epoch": 58.0,
"eval_accuracy": 71.85185185185186,
"eval_average_metrics": 73.6642490769342,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"cancellation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.6388888888888888, \"f1-score\": 0.6865671641791045, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7071727257211129, \"recall\": 0.7029213070879737, \"f1-score\": 0.6958756590822236, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7471616316777607, \"recall\": 0.7185185185185186, \"f1-score\": 0.7240017370132812, \"support\": 135.0}}",
"eval_f1_macro": 78.28601164675014,
"eval_f1_micro": 72.1189591078067,
"eval_f1_weighted": 72.40017370132811,
"eval_loss": 0.9572186470031738,
"eval_runtime": 2.1451,
"eval_samples_per_second": 62.935,
"step": 986
},
{
"epoch": 59.0,
"learning_rate": 0.00023499999999999997,
"loss": 0.0042,
"step": 1003
},
{
"epoch": 59.0,
"eval_accuracy": 71.11111111111111,
"eval_average_metrics": 71.7649842113928,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.6388888888888888, \"f1-score\": 0.6865671641791045, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6857142857142857, \"recall\": 0.6153846153846154, \"f1-score\": 0.6486486486486486, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7765705163233214, \"recall\": 0.7346967268842268, \"f1-score\": 0.736597489423982, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7294317448174391, \"recall\": 0.7111111111111111, \"f1-score\": 0.7117796568095075, \"support\": 135.0}}",
"eval_f1_macro": 73.6597489423982,
"eval_f1_micro": 71.11111111111111,
"eval_f1_weighted": 71.17796568095075,
"eval_loss": 0.8926369547843933,
"eval_runtime": 2.1674,
"eval_samples_per_second": 62.286,
"step": 1003
},
{
"epoch": 60.0,
"learning_rate": 0.0002333333333333333,
"loss": 0.0108,
"step": 1020
},
{
"epoch": 60.0,
"eval_accuracy": 67.4074074074074,
"eval_average_metrics": 67.96942582571397,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.7272727272727273, \"f1-score\": 0.6956521739130435, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6896551724137931, \"recall\": 0.5555555555555556, \"f1-score\": 0.6153846153846154, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6756756756756757, \"recall\": 0.6410256410256411, \"f1-score\": 0.6578947368421052, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.6923076923076923, \"f1-score\": 0.8181818181818181, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.674074074074074, \"macro avg\": {\"precision\": 0.7355337802536078, \"recall\": 0.6994160353535354, \"f1-score\": 0.6963058347068645, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.6961066081755738, \"recall\": 0.674074074074074, \"f1-score\": 0.674323050173546, \"support\": 135.0}}",
"eval_f1_macro": 69.63058347068645,
"eval_f1_micro": 67.4074074074074,
"eval_f1_weighted": 67.4323050173546,
"eval_loss": 0.9399789571762085,
"eval_runtime": 2.1159,
"eval_samples_per_second": 63.802,
"step": 1020
},
{
"epoch": 61.0,
"learning_rate": 0.00023166666666666667,
"loss": 0.0056,
"step": 1037
},
{
"epoch": 61.0,
"eval_accuracy": 69.62962962962963,
"eval_average_metrics": 71.17152096147001,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 1.0, \"f1-score\": 0.8571428571428571, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7037037037037037, \"recall\": 0.5277777777777778, \"f1-score\": 0.6031746031746033, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6153846153846154, \"f1-score\": 0.676056338028169, \"support\": 39.0}, \"question\": {\"precision\": 0.5384615384615384, \"recall\": 0.7777777777777778, \"f1-score\": 0.6363636363636364, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.52, \"recall\": 0.8125, \"f1-score\": 0.6341463414634146, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.7443091168091167, \"recall\": 0.7997219794094794, \"f1-score\": 0.7591938053549183, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.719369420702754, \"recall\": 0.6962962962962963, \"f1-score\": 0.6950744405112893, \"support\": 135.0}}",
"eval_f1_macro": 75.91938053549183,
"eval_f1_micro": 69.62962962962963,
"eval_f1_weighted": 69.50744405112893,
"eval_loss": 0.9959737062454224,
"eval_runtime": 2.1462,
"eval_samples_per_second": 62.902,
"step": 1037
},
{
"epoch": 62.0,
"learning_rate": 0.00023,
"loss": 0.0091,
"step": 1054
},
{
"epoch": 62.0,
"eval_accuracy": 74.07407407407408,
"eval_average_metrics": 74.43021846765396,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6923076923076923, \"recall\": 0.75, \"f1-score\": 0.7199999999999999, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8275862068965517, \"recall\": 0.6153846153846154, \"f1-score\": 0.7058823529411765, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.7951003737641669, \"recall\": 0.7485856157731158, \"f1-score\": 0.754229181551163, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7622480329376882, \"recall\": 0.7407407407407407, \"f1-score\": 0.7414980756735141, \"support\": 135.0}}",
"eval_f1_macro": 75.42291815511629,
"eval_f1_micro": 74.07407407407408,
"eval_f1_weighted": 74.14980756735142,
"eval_loss": 0.9423182010650635,
"eval_runtime": 2.0725,
"eval_samples_per_second": 65.137,
"step": 1054
},
{
"epoch": 63.0,
"learning_rate": 0.0002283333333333333,
"loss": 0.006,
"step": 1071
},
{
"epoch": 63.0,
"eval_accuracy": 70.37037037037037,
"eval_average_metrics": 71.13630195174072,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6774193548387096, \"recall\": 0.5833333333333334, \"f1-score\": 0.626865671641791, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7037037037037037, \"macro avg\": {\"precision\": 0.7705405145929339, \"recall\": 0.7309574106449107, \"f1-score\": 0.7339954289925987, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7191460431245379, \"recall\": 0.7037037037037037, \"f1-score\": 0.7040492416696227, \"support\": 135.0}}",
"eval_f1_macro": 73.39954289925987,
"eval_f1_micro": 70.37037037037037,
"eval_f1_weighted": 70.40492416696227,
"eval_loss": 0.9775393009185791,
"eval_runtime": 2.1426,
"eval_samples_per_second": 63.009,
"step": 1071
},
{
"epoch": 64.0,
"learning_rate": 0.00022666666666666663,
"loss": 0.0071,
"step": 1088
},
{
"epoch": 64.0,
"eval_accuracy": 71.85185185185186,
"eval_average_metrics": 72.50660628560995,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6571428571428571, \"recall\": 0.6388888888888888, \"f1-score\": 0.647887323943662, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8333333333333334, \"recall\": 0.6410256410256411, \"f1-score\": 0.7246376811594204, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5185185185185185, \"recall\": 0.875, \"f1-score\": 0.6511627906976744, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7855794668294669, \"recall\": 0.7343507187257188, \"f1-score\": 0.7410246999652906, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7480406698925217, \"recall\": 0.7185185185185186, \"f1-score\": 0.7222025144220701, \"support\": 135.0}}",
"eval_f1_macro": 74.10246999652907,
"eval_f1_micro": 71.85185185185186,
"eval_f1_weighted": 72.22025144220703,
"eval_loss": 1.0484141111373901,
"eval_runtime": 2.2145,
"eval_samples_per_second": 60.962,
"step": 1088
},
{
"epoch": 65.0,
"learning_rate": 0.000225,
"loss": 0.0034,
"step": 1105
},
{
"epoch": 65.0,
"eval_accuracy": 73.33333333333333,
"eval_average_metrics": 73.9244200133918,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7741935483870968, \"recall\": 0.6666666666666666, \"f1-score\": 0.7164179104477612, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.675, \"recall\": 0.6923076923076923, \"f1-score\": 0.6835443037974683, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5789473684210527, \"recall\": 0.6875, \"f1-score\": 0.6285714285714286, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7900560761394801, \"recall\": 0.7485734751359752, \"f1-score\": 0.7557596878082226, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7442556314967181, \"recall\": 0.7333333333333333, \"f1-score\": 0.7345504460607825, \"support\": 135.0}}",
"eval_f1_macro": 75.57596878082227,
"eval_f1_micro": 73.33333333333333,
"eval_f1_weighted": 73.45504460607825,
"eval_loss": 0.943881094455719,
"eval_runtime": 2.2664,
"eval_samples_per_second": 59.565,
"step": 1105
},
{
"epoch": 66.0,
"learning_rate": 0.00022333333333333333,
"loss": 0.007,
"step": 1122
},
{
"epoch": 66.0,
"eval_accuracy": 71.11111111111111,
"eval_average_metrics": 71.91255932928726,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7692307692307693, \"recall\": 0.9090909090909091, \"f1-score\": 0.8333333333333333, \"support\": 11.0}, \"challenge\": {\"precision\": 0.65625, \"recall\": 0.5833333333333334, \"f1-score\": 0.6176470588235293, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6923076923076923, \"f1-score\": 0.7199999999999999, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.8125, \"f1-score\": 0.6190476190476191, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7927684294871795, \"recall\": 0.7234787781662781, \"f1-score\": 0.7393785014005602, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7358262108262108, \"recall\": 0.7111111111111111, \"f1-score\": 0.7149016495487084, \"support\": 135.0}}",
"eval_f1_macro": 73.93785014005603,
"eval_f1_micro": 71.11111111111111,
"eval_f1_weighted": 71.49016495487082,
"eval_loss": 1.0599663257598877,
"eval_runtime": 2.1786,
"eval_samples_per_second": 61.965,
"step": 1122
},
{
"epoch": 67.0,
"learning_rate": 0.00022166666666666667,
"loss": 0.0042,
"step": 1139
},
{
"epoch": 67.0,
"eval_accuracy": 73.33333333333333,
"eval_average_metrics": 75.12632131391055,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.6111111111111112, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8064516129032258, \"recall\": 0.6410256410256411, \"f1-score\": 0.7142857142857142, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.48148148148148145, \"recall\": 0.8125, \"f1-score\": 0.6046511627906976, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7984916367980883, \"recall\": 0.8247074106449106, \"f1-score\": 0.801042885612959, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7643603699278729, \"recall\": 0.7333333333333333, \"f1-score\": 0.7373433002767967, \"support\": 135.0}}",
"eval_f1_macro": 80.1042885612959,
"eval_f1_micro": 73.33333333333333,
"eval_f1_weighted": 73.73433002767968,
"eval_loss": 1.0870040655136108,
"eval_runtime": 2.253,
"eval_samples_per_second": 59.919,
"step": 1139
},
{
"epoch": 68.0,
"learning_rate": 0.00021999999999999995,
"loss": 0.0048,
"step": 1156
},
{
"epoch": 68.0,
"eval_accuracy": 71.85185185185186,
"eval_average_metrics": 73.55212011915016,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.5833333333333334, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.48148148148148145, \"recall\": 0.8125, \"f1-score\": 0.6046511627906976, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.791516354016354, \"recall\": 0.7966625388500388, \"f1-score\": 0.7820397286821705, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7515169329984145, \"recall\": 0.7185185185185186, \"f1-score\": 0.7230080390467987, \"support\": 135.0}}",
"eval_f1_macro": 78.20397286821705,
"eval_f1_micro": 71.85185185185186,
"eval_f1_weighted": 72.30080390467987,
"eval_loss": 1.0407289266586304,
"eval_runtime": 2.2127,
"eval_samples_per_second": 61.011,
"step": 1156
},
{
"epoch": 69.0,
"learning_rate": 0.0002183333333333333,
"loss": 0.0026,
"step": 1173
},
{
"epoch": 69.0,
"eval_accuracy": 68.88888888888889,
"eval_average_metrics": 70.6530531645798,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.72, \"recall\": 0.5, \"f1-score\": 0.5901639344262295, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6410256410256411, \"f1-score\": 0.6944444444444444, \"support\": 39.0}, \"question\": {\"precision\": 0.5, \"recall\": 0.7777777777777778, \"f1-score\": 0.6086956521739131, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.4642857142857143, \"recall\": 0.8125, \"f1-score\": 0.5909090909090908, \"support\": 16.0}, \"accuracy\": 0.6888888888888889, \"macro avg\": {\"precision\": 0.7635660173160174, \"recall\": 0.783040743978244, \"f1-score\": 0.7556388197313891, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.731313772647106, \"recall\": 0.6888888888888889, \"f1-score\": 0.6927055290740252, \"support\": 135.0}}",
"eval_f1_macro": 75.56388197313892,
"eval_f1_micro": 68.88888888888889,
"eval_f1_weighted": 69.27055290740252,
"eval_loss": 1.0904052257537842,
"eval_runtime": 2.1381,
"eval_samples_per_second": 63.141,
"step": 1173
},
{
"epoch": 70.0,
"learning_rate": 0.00021666666666666666,
"loss": 0.0014,
"step": 1190
},
{
"epoch": 70.0,
"eval_accuracy": 73.33333333333333,
"eval_average_metrics": 74.87504553531681,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.6388888888888888, \"f1-score\": 0.6865671641791045, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6410256410256411, \"f1-score\": 0.7042253521126761, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7893565188172043, \"recall\": 0.814290743978244, \"f1-score\": 0.7918166361805148, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7587290919952211, \"recall\": 0.7333333333333333, \"f1-score\": 0.7365185185654917, \"support\": 135.0}}",
"eval_f1_macro": 79.18166361805147,
"eval_f1_micro": 73.33333333333333,
"eval_f1_weighted": 73.65185185654917,
"eval_loss": 1.0420516729354858,
"eval_runtime": 2.151,
"eval_samples_per_second": 62.763,
"step": 1190
},
{
"epoch": 71.0,
"learning_rate": 0.000215,
"loss": 0.0047,
"step": 1207
},
{
"epoch": 71.0,
"eval_accuracy": 73.33333333333333,
"eval_average_metrics": 73.96275924849695,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7058823529411765, \"recall\": 0.6666666666666666, \"f1-score\": 0.6857142857142857, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6410256410256411, \"f1-score\": 0.7042253521126761, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7944225059094784, \"recall\": 0.7552629662004662, \"f1-score\": 0.7562933597057457, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7554414742767257, \"recall\": 0.7333333333333333, \"f1-score\": 0.735550343567466, \"support\": 135.0}}",
"eval_f1_macro": 75.62933597057457,
"eval_f1_micro": 73.33333333333333,
"eval_f1_weighted": 73.5550343567466,
"eval_loss": 1.0715558528900146,
"eval_runtime": 2.2105,
"eval_samples_per_second": 61.073,
"step": 1207
},
{
"epoch": 72.0,
"learning_rate": 0.00021333333333333333,
"loss": 0.0019,
"step": 1224
},
{
"epoch": 72.0,
"eval_accuracy": 71.85185185185186,
"eval_average_metrics": 72.36672605045894,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6428571428571429, \"recall\": 0.8181818181818182, \"f1-score\": 0.7200000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7666666666666667, \"recall\": 0.6388888888888888, \"f1-score\": 0.696969696969697, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7352941176470589, \"recall\": 0.6410256410256411, \"f1-score\": 0.6849315068493151, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7698383520074696, \"recall\": 0.7404271076146076, \"f1-score\": 0.7363487615884876, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7427004876024484, \"recall\": 0.7185185185185186, \"f1-score\": 0.7212832433928325, \"support\": 135.0}}",
"eval_f1_macro": 73.63487615884875,
"eval_f1_micro": 71.85185185185186,
"eval_f1_weighted": 72.12832433928324,
"eval_loss": 1.1170759201049805,
"eval_runtime": 2.1545,
"eval_samples_per_second": 62.659,
"step": 1224
},
{
"epoch": 73.0,
"learning_rate": 0.00021166666666666667,
"loss": 0.0009,
"step": 1241
},
{
"epoch": 73.0,
"eval_accuracy": 73.33333333333333,
"eval_average_metrics": 74.45112685839807,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6923076923076923, \"recall\": 0.75, \"f1-score\": 0.7199999999999999, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8214285714285714, \"recall\": 0.5897435897435898, \"f1-score\": 0.6865671641791046, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7725357975357976, \"recall\": 0.7965168512043512, \"f1-score\": 0.7786937025399319, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7492250958917624, \"recall\": 0.7333333333333333, \"f1-score\": 0.7326847051293242, \"support\": 135.0}}",
"eval_f1_macro": 77.8693702539932,
"eval_f1_micro": 73.33333333333333,
"eval_f1_weighted": 73.26847051293241,
"eval_loss": 1.0454473495483398,
"eval_runtime": 2.0945,
"eval_samples_per_second": 64.453,
"step": 1241
},
{
"epoch": 74.0,
"learning_rate": 0.00020999999999999998,
"loss": 0.0137,
"step": 1258
},
{
"epoch": 74.0,
"eval_accuracy": 71.11111111111111,
"eval_average_metrics": 71.70350070614597,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.6363636363636364, \"f1-score\": 0.6363636363636364, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6097560975609756, \"recall\": 0.6944444444444444, \"f1-score\": 0.6493506493506495, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8275862068965517, \"recall\": 0.6153846153846154, \"f1-score\": 0.7058823529411765, \"support\": 39.0}, \"question\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.56, \"recall\": 0.875, \"f1-score\": 0.6829268292682927, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7840743537137566, \"recall\": 0.7153627622377623, \"f1-score\": 0.7315785053858942, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7402746783789677, \"recall\": 0.7111111111111111, \"f1-score\": 0.7143393006377224, \"support\": 135.0}}",
"eval_f1_macro": 73.15785053858941,
"eval_f1_micro": 71.11111111111111,
"eval_f1_weighted": 71.43393006377225,
"eval_loss": 1.0791348218917847,
"eval_runtime": 2.1939,
"eval_samples_per_second": 61.535,
"step": 1258
},
{
"epoch": 75.0,
"learning_rate": 0.00020833333333333332,
"loss": 0.0254,
"step": 1275
},
{
"epoch": 75.0,
"eval_accuracy": 74.07407407407408,
"eval_average_metrics": 74.53066318197898,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7352941176470589, \"recall\": 0.6944444444444444, \"f1-score\": 0.7142857142857144, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.6666666666666666, \"f1-score\": 0.7027027027027027, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.7937821165762342, \"recall\": 0.7505766802641802, \"f1-score\": 0.7575329088486984, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7541625694566871, \"recall\": 0.7407407407407407, \"f1-score\": 0.7422121369489791, \"support\": 135.0}}",
"eval_f1_macro": 75.75329088486984,
"eval_f1_micro": 74.07407407407408,
"eval_f1_weighted": 74.22121369489791,
"eval_loss": 0.9704490303993225,
"eval_runtime": 2.225,
"eval_samples_per_second": 60.674,
"step": 1275
},
{
"epoch": 76.0,
"learning_rate": 0.00020666666666666666,
"loss": 0.0039,
"step": 1292
},
{
"epoch": 76.0,
"eval_accuracy": 71.11111111111111,
"eval_average_metrics": 71.88592559677284,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.5833333333333334, \"f1-score\": 0.6461538461538462, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.52, \"recall\": 0.8125, \"f1-score\": 0.6341463414634146, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7785046518667209, \"recall\": 0.7366877913752914, \"f1-score\": 0.7404468801773039, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7311833743557882, \"recall\": 0.7111111111111111, \"f1-score\": 0.7127679214713875, \"support\": 135.0}}",
"eval_f1_macro": 74.04468801773038,
"eval_f1_micro": 71.11111111111111,
"eval_f1_weighted": 71.27679214713875,
"eval_loss": 1.000522255897522,
"eval_runtime": 2.2253,
"eval_samples_per_second": 60.667,
"step": 1292
},
{
"epoch": 77.0,
"learning_rate": 0.000205,
"loss": 0.0095,
"step": 1309
},
{
"epoch": 77.0,
"eval_accuracy": 71.85185185185186,
"eval_average_metrics": 72.50938740714625,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6428571428571429, \"recall\": 0.8181818181818182, \"f1-score\": 0.7200000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.696969696969697, \"recall\": 0.6388888888888888, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6410256410256411, \"f1-score\": 0.7042253521126761, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.52, \"recall\": 0.8125, \"f1-score\": 0.6341463414634146, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7787457160894661, \"recall\": 0.7404271076146076, \"f1-score\": 0.7417628359660171, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7426006493506493, \"recall\": 0.7185185185185186, \"f1-score\": 0.7215756232827957, \"support\": 135.0}}",
"eval_f1_macro": 74.17628359660172,
"eval_f1_micro": 71.85185185185186,
"eval_f1_weighted": 72.15756232827957,
"eval_loss": 0.9987503886222839,
"eval_runtime": 2.0847,
"eval_samples_per_second": 64.758,
"step": 1309
},
{
"epoch": 78.0,
"learning_rate": 0.00020333333333333333,
"loss": 0.0016,
"step": 1326
},
{
"epoch": 78.0,
"eval_accuracy": 71.85185185185186,
"eval_average_metrics": 72.50535308482088,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6428571428571429, \"recall\": 0.8181818181818182, \"f1-score\": 0.7200000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.6388888888888888, \"f1-score\": 0.6865671641791045, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7787042370711725, \"recall\": 0.7404271076146076, \"f1-score\": 0.74266339591752, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7378141889969847, \"recall\": 0.7185185185185186, \"f1-score\": 0.7205136904382781, \"support\": 135.0}}",
"eval_f1_macro": 74.266339591752,
"eval_f1_micro": 71.85185185185186,
"eval_f1_weighted": 72.0513690438278,
"eval_loss": 1.0020543336868286,
"eval_runtime": 2.1954,
"eval_samples_per_second": 61.492,
"step": 1326
},
{
"epoch": 79.0,
"learning_rate": 0.00020166666666666667,
"loss": 0.0011,
"step": 1343
},
{
"epoch": 79.0,
"eval_accuracy": 72.5925925925926,
"eval_average_metrics": 73.48732962316897,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6764705882352942, \"recall\": 0.6388888888888888, \"f1-score\": 0.6571428571428571, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6410256410256411, \"f1-score\": 0.7042253521126761, \"support\": 39.0}, \"question\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.52, \"recall\": 0.8125, \"f1-score\": 0.6341463414634146, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.8027706290849673, \"recall\": 0.751790743978244, \"f1-score\": 0.7583791127061329, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7516668482207698, \"recall\": 0.725925925925926, \"f1-score\": 0.7292622203687736, \"support\": 135.0}}",
"eval_f1_macro": 75.8379112706133,
"eval_f1_micro": 72.59259259259261,
"eval_f1_weighted": 72.92622203687736,
"eval_loss": 1.041650652885437,
"eval_runtime": 2.0975,
"eval_samples_per_second": 64.363,
"step": 1343
},
{
"epoch": 80.0,
"learning_rate": 0.00019999999999999998,
"loss": 0.0022,
"step": 1360
},
{
"epoch": 80.0,
"eval_accuracy": 72.5925925925926,
"eval_average_metrics": 73.38039917163168,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.6388888888888888, \"f1-score\": 0.6865671641791045, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7955693164362518, \"recall\": 0.751790743978244, \"f1-score\": 0.7553531897837845, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7471616316777607, \"recall\": 0.725925925925926, \"f1-score\": 0.7280109252296305, \"support\": 135.0}}",
"eval_f1_macro": 75.53531897837846,
"eval_f1_micro": 72.59259259259261,
"eval_f1_weighted": 72.80109252296305,
"eval_loss": 1.0494800806045532,
"eval_runtime": 2.2088,
"eval_samples_per_second": 61.119,
"step": 1360
},
{
"epoch": 81.0,
"learning_rate": 0.00019833333333333332,
"loss": 0.0126,
"step": 1377
},
{
"epoch": 81.0,
"eval_accuracy": 68.88888888888889,
"eval_average_metrics": 69.48276901714948,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.72, \"recall\": 0.5, \"f1-score\": 0.5901639344262295, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.5, \"recall\": 0.7777777777777778, \"f1-score\": 0.6086956521739131, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.52, \"recall\": 0.8125, \"f1-score\": 0.6341463414634146, \"support\": 16.0}, \"accuracy\": 0.6888888888888889, \"macro avg\": {\"precision\": 0.7443162393162392, \"recall\": 0.7262711247086246, \"f1-score\": 0.7127222322360148, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.716459639126306, \"recall\": 0.6888888888888889, \"f1-score\": 0.6888107506721864, \"support\": 135.0}}",
"eval_f1_macro": 71.27222322360149,
"eval_f1_micro": 68.88888888888889,
"eval_f1_weighted": 68.88107506721865,
"eval_loss": 1.0959280729293823,
"eval_runtime": 2.2009,
"eval_samples_per_second": 61.34,
"step": 1377
},
{
"epoch": 82.0,
"learning_rate": 0.00019666666666666666,
"loss": 0.0033,
"step": 1394
},
{
"epoch": 82.0,
"eval_accuracy": 73.33333333333333,
"eval_average_metrics": 74.0222652316685,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.8, \"recall\": 0.7272727272727273, \"f1-score\": 0.761904761904762, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6571428571428571, \"recall\": 0.6388888888888888, \"f1-score\": 0.647887323943662, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8, \"recall\": 0.717948717948718, \"f1-score\": 0.7567567567567569, \"support\": 39.0}, \"question\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.52, \"recall\": 0.8125, \"f1-score\": 0.6341463414634146, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.8054761904761905, \"recall\": 0.738678855866356, \"f1-score\": 0.7563368980085744, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7553862433862435, \"recall\": 0.7333333333333333, \"f1-score\": 0.737887044591499, \"support\": 135.0}}",
"eval_f1_macro": 75.63368980085744,
"eval_f1_micro": 73.33333333333333,
"eval_f1_weighted": 73.78870445914991,
"eval_loss": 1.07533860206604,
"eval_runtime": 2.33,
"eval_samples_per_second": 57.939,
"step": 1394
},
{
"epoch": 83.0,
"learning_rate": 0.000195,
"loss": 0.0029,
"step": 1411
},
{
"epoch": 83.0,
"eval_accuracy": 74.07407407407408,
"eval_average_metrics": 74.76246669403774,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6842105263157895, \"recall\": 0.7222222222222222, \"f1-score\": 0.7027027027027027, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6410256410256411, \"f1-score\": 0.7042253521126761, \"support\": 39.0}, \"question\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.75, \"f1-score\": 0.6486486486486486, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.8101666927736006, \"recall\": 0.7543949106449106, \"f1-score\": 0.7658868817992679, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7598260698041399, \"recall\": 0.7407407407407407, \"f1-score\": 0.7431303044807601, \"support\": 135.0}}",
"eval_f1_macro": 76.58868817992679,
"eval_f1_micro": 74.07407407407408,
"eval_f1_weighted": 74.313030448076,
"eval_loss": 1.025146722793579,
"eval_runtime": 2.189,
"eval_samples_per_second": 61.672,
"step": 1411
},
{
"epoch": 84.0,
"learning_rate": 0.00019333333333333333,
"loss": 0.0008,
"step": 1428
},
{
"epoch": 84.0,
"eval_accuracy": 71.11111111111111,
"eval_average_metrics": 71.65410439323483,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7586206896551724, \"recall\": 0.6111111111111112, \"f1-score\": 0.676923076923077, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.5384615384615384, \"recall\": 0.7777777777777778, \"f1-score\": 0.6363636363636364, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.766517611157791, \"recall\": 0.7387031371406372, \"f1-score\": 0.7312475747258356, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7362255787043392, \"recall\": 0.7111111111111111, \"f1-score\": 0.7126943787813353, \"support\": 135.0}}",
"eval_f1_macro": 73.12475747258355,
"eval_f1_micro": 71.11111111111111,
"eval_f1_weighted": 71.26943787813353,
"eval_loss": 1.0640238523483276,
"eval_runtime": 2.1727,
"eval_samples_per_second": 62.136,
"step": 1428
},
{
"epoch": 85.0,
"learning_rate": 0.00019166666666666665,
"loss": 0.0007,
"step": 1445
},
{
"epoch": 85.0,
"eval_accuracy": 71.11111111111111,
"eval_average_metrics": 71.63498328794381,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7586206896551724, \"recall\": 0.6111111111111112, \"f1-score\": 0.676923076923077, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7612767620310724, \"recall\": 0.7369548853923854, \"f1-score\": 0.730588001969581, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7325891775317063, \"recall\": 0.7111111111111111, \"f1-score\": 0.7125891073259495, \"support\": 135.0}}",
"eval_f1_macro": 73.05880019695809,
"eval_f1_micro": 71.11111111111111,
"eval_f1_weighted": 71.25891073259494,
"eval_loss": 1.0694376230239868,
"eval_runtime": 2.1511,
"eval_samples_per_second": 62.759,
"step": 1445
},
{
"epoch": 86.0,
"learning_rate": 0.00018999999999999998,
"loss": 0.0013,
"step": 1462
},
{
"epoch": 86.0,
"eval_accuracy": 73.33333333333333,
"eval_average_metrics": 73.88851983050755,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6944444444444444, \"recall\": 0.6944444444444444, \"f1-score\": 0.6944444444444444, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6410256410256411, \"f1-score\": 0.7042253521126761, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7967806461352657, \"recall\": 0.7448462995337995, \"f1-score\": 0.7532179628803489, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.754411567364466, \"recall\": 0.7333333333333333, \"f1-score\": 0.7356561636732861, \"support\": 135.0}}",
"eval_f1_macro": 75.3217962880349,
"eval_f1_micro": 73.33333333333333,
"eval_f1_weighted": 73.56561636732862,
"eval_loss": 1.076446771621704,
"eval_runtime": 2.2436,
"eval_samples_per_second": 60.171,
"step": 1462
},
{
"epoch": 87.0,
"learning_rate": 0.00018833333333333332,
"loss": 0.0052,
"step": 1479
},
{
"epoch": 87.0,
"eval_accuracy": 77.03703703703704,
"eval_average_metrics": 77.53590008051296,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7692307692307693, \"recall\": 0.9090909090909091, \"f1-score\": 0.8333333333333333, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7105263157894737, \"recall\": 0.75, \"f1-score\": 0.7297297297297298, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7941176470588235, \"recall\": 0.6923076923076923, \"f1-score\": 0.7397260273972601, \"support\": 39.0}, \"question\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6190476190476191, \"recall\": 0.8125, \"f1-score\": 0.7027027027027026, \"support\": 16.0}, \"accuracy\": 0.7703703703703704, \"macro avg\": {\"precision\": 0.833837516113058, \"recall\": 0.7720898892773893, \"f1-score\": 0.7884724218577965, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.784561747926041, \"recall\": 0.7703703703703704, \"f1-score\": 0.7722228406219811, \"support\": 135.0}}",
"eval_f1_macro": 78.84724218577965,
"eval_f1_micro": 77.03703703703704,
"eval_f1_weighted": 77.22228406219811,
"eval_loss": 1.0223504304885864,
"eval_runtime": 2.1607,
"eval_samples_per_second": 62.479,
"step": 1479
},
{
"epoch": 88.0,
"learning_rate": 0.00018666666666666666,
"loss": 0.0004,
"step": 1496
},
{
"epoch": 88.0,
"eval_accuracy": 75.55555555555556,
"eval_average_metrics": 76.05592523834689,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7692307692307693, \"recall\": 0.9090909090909091, \"f1-score\": 0.8333333333333333, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7575757575757576, \"recall\": 0.6944444444444444, \"f1-score\": 0.7246376811594203, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6923076923076923, \"f1-score\": 0.7199999999999999, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7555555555555555, \"macro avg\": {\"precision\": 0.8130099067599068, \"recall\": 0.7651454448329449, \"f1-score\": 0.773336418091264, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7716006216006216, \"recall\": 0.7555555555555555, \"f1-score\": 0.7577894803315004, \"support\": 135.0}}",
"eval_f1_macro": 77.33364180912639,
"eval_f1_micro": 75.55555555555556,
"eval_f1_weighted": 75.77894803315003,
"eval_loss": 1.0519832372665405,
"eval_runtime": 2.2144,
"eval_samples_per_second": 60.966,
"step": 1496
},
{
"epoch": 89.0,
"learning_rate": 0.000185,
"loss": 0.0004,
"step": 1513
},
{
"epoch": 89.0,
"eval_accuracy": 75.55555555555556,
"eval_average_metrics": 76.11648469654651,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7692307692307693, \"recall\": 0.9090909090909091, \"f1-score\": 0.8333333333333333, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6923076923076923, \"recall\": 0.75, \"f1-score\": 0.7199999999999999, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7714285714285715, \"recall\": 0.6923076923076923, \"f1-score\": 0.7297297297297296, \"support\": 39.0}, \"question\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 16.0}, \"accuracy\": 0.7555555555555555, \"macro avg\": {\"precision\": 0.8263431013431014, \"recall\": 0.7546620046620047, \"f1-score\": 0.7756144827692142, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7708913308913309, \"recall\": 0.7555555555555555, \"f1-score\": 0.7579337939815347, \"support\": 135.0}}",
"eval_f1_macro": 77.56144827692142,
"eval_f1_micro": 75.55555555555556,
"eval_f1_weighted": 75.79337939815348,
"eval_loss": 1.0824300050735474,
"eval_runtime": 2.2212,
"eval_samples_per_second": 60.777,
"step": 1513
},
{
"epoch": 90.0,
"learning_rate": 0.00018333333333333334,
"loss": 0.0023,
"step": 1530
},
{
"epoch": 90.0,
"eval_accuracy": 71.11111111111111,
"eval_average_metrics": 71.89906381584439,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.5833333333333334, \"f1-score\": 0.6562499999999999, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.8125, \"f1-score\": 0.6190476190476191, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7837301587301587, \"recall\": 0.7384360431235432, \"f1-score\": 0.7390882467117281, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7427689594356262, \"recall\": 0.7111111111111111, \"f1-score\": 0.7146520836998245, \"support\": 135.0}}",
"eval_f1_macro": 73.90882467117281,
"eval_f1_micro": 71.11111111111111,
"eval_f1_weighted": 71.46520836998246,
"eval_loss": 1.1680539846420288,
"eval_runtime": 2.2118,
"eval_samples_per_second": 61.037,
"step": 1530
},
{
"epoch": 91.0,
"learning_rate": 0.00018166666666666665,
"loss": 0.0017,
"step": 1547
},
{
"epoch": 91.0,
"eval_accuracy": 71.85185185185186,
"eval_average_metrics": 73.54780705462385,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.5833333333333334, \"f1-score\": 0.6562499999999999, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6410256410256411, \"f1-score\": 0.6944444444444444, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.8125, \"f1-score\": 0.6190476190476191, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7821969696969697, \"recall\": 0.8073462995337995, \"f1-score\": 0.7829353295805501, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7491021324354658, \"recall\": 0.7185185185185186, \"f1-score\": 0.7219399155673665, \"support\": 135.0}}",
"eval_f1_macro": 78.29353295805501,
"eval_f1_micro": 71.85185185185186,
"eval_f1_weighted": 72.19399155673665,
"eval_loss": 1.1076750755310059,
"eval_runtime": 2.1138,
"eval_samples_per_second": 63.867,
"step": 1547
},
{
"epoch": 92.0,
"learning_rate": 0.00017999999999999998,
"loss": 0.0072,
"step": 1564
},
{
"epoch": 92.0,
"eval_accuracy": 72.5925925925926,
"eval_average_metrics": 73.88573213660933,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.8, \"recall\": 0.7272727272727273, \"f1-score\": 0.761904761904762, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.5833333333333334, \"f1-score\": 0.6461538461538462, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7567567567567568, \"recall\": 0.717948717948718, \"f1-score\": 0.736842105263158, \"support\": 39.0}, \"question\": {\"precision\": 0.5, \"recall\": 0.7777777777777778, \"f1-score\": 0.6086956521739131, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7721699167819858, \"recall\": 0.7942344114219114, \"f1-score\": 0.7758369725875447, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.742619934459015, \"recall\": 0.725925925925926, \"f1-score\": 0.7277404610249767, \"support\": 135.0}}",
"eval_f1_macro": 77.58369725875447,
"eval_f1_micro": 72.59259259259261,
"eval_f1_weighted": 72.77404610249766,
"eval_loss": 1.0486587285995483,
"eval_runtime": 2.2394,
"eval_samples_per_second": 60.284,
"step": 1564
},
{
"epoch": 93.0,
"learning_rate": 0.00017833333333333332,
"loss": 0.0057,
"step": 1581
},
{
"epoch": 93.0,
"eval_accuracy": 68.88888888888889,
"eval_average_metrics": 69.75310065197965,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.8, \"recall\": 0.7272727272727273, \"f1-score\": 0.761904761904762, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6923076923076923, \"recall\": 0.5, \"f1-score\": 0.5806451612903226, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7368421052631579, \"recall\": 0.717948717948718, \"f1-score\": 0.7272727272727273, \"support\": 39.0}, \"question\": {\"precision\": 0.5, \"recall\": 0.7777777777777778, \"f1-score\": 0.6086956521739131, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.4642857142857143, \"recall\": 0.8125, \"f1-score\": 0.5909090909090908, \"support\": 16.0}, \"accuracy\": 0.6888888888888889, \"macro avg\": {\"precision\": 0.7741794389820706, \"recall\": 0.7117023601398601, \"f1-score\": 0.7183544685246277, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7288036330141594, \"recall\": 0.6888888888888889, \"f1-score\": 0.693991779776781, \"support\": 135.0}}",
"eval_f1_macro": 71.83544685246277,
"eval_f1_micro": 68.88888888888889,
"eval_f1_weighted": 69.3991779776781,
"eval_loss": 1.1878422498703003,
"eval_runtime": 2.1986,
"eval_samples_per_second": 61.403,
"step": 1581
},
{
"epoch": 94.0,
"learning_rate": 0.00017666666666666666,
"loss": 0.0195,
"step": 1598
},
{
"epoch": 94.0,
"eval_accuracy": 74.07407407407408,
"eval_average_metrics": 74.61304930742271,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7575757575757576, \"recall\": 0.6944444444444444, \"f1-score\": 0.7246376811594203, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6410256410256411, \"f1-score\": 0.6944444444444444, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.7979249011857708, \"recall\": 0.7587351884226884, \"f1-score\": 0.7599361706778587, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7623871565900551, \"recall\": 0.7407407407407407, \"f1-score\": 0.7431043201375682, \"support\": 135.0}}",
"eval_f1_macro": 75.99361706778586,
"eval_f1_micro": 74.07407407407408,
"eval_f1_weighted": 74.31043201375682,
"eval_loss": 1.0754647254943848,
"eval_runtime": 2.2267,
"eval_samples_per_second": 60.629,
"step": 1598
},
{
"epoch": 95.0,
"learning_rate": 0.000175,
"loss": 0.0009,
"step": 1615
},
{
"epoch": 95.0,
"eval_accuracy": 74.07407407407408,
"eval_average_metrics": 74.42176515283398,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.7272727272727273, \"f1-score\": 0.6956521739130435, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7105263157894737, \"recall\": 0.75, \"f1-score\": 0.7297297297297298, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6410256410256411, \"f1-score\": 0.6944444444444444, \"support\": 39.0}, \"question\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.6190476190476191, \"recall\": 0.8125, \"f1-score\": 0.7027027027027026, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.7921437115516063, \"recall\": 0.7429523601398601, \"f1-score\": 0.7544827980154067, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7502160282862037, \"recall\": 0.7407407407407407, \"f1-score\": 0.7409063266164715, \"support\": 135.0}}",
"eval_f1_macro": 75.44827980154068,
"eval_f1_micro": 74.07407407407408,
"eval_f1_weighted": 74.09063266164715,
"eval_loss": 1.0278376340866089,
"eval_runtime": 2.155,
"eval_samples_per_second": 62.646,
"step": 1615
},
{
"epoch": 96.0,
"learning_rate": 0.0001733333333333333,
"loss": 0.0004,
"step": 1632
},
{
"epoch": 96.0,
"eval_accuracy": 74.81481481481481,
"eval_average_metrics": 75.17056142900891,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7297297297297297, \"recall\": 0.75, \"f1-score\": 0.7397260273972601, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6410256410256411, \"f1-score\": 0.6944444444444444, \"support\": 39.0}, \"question\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.6190476190476191, \"recall\": 0.8125, \"f1-score\": 0.7027027027027026, \"support\": 16.0}, \"accuracy\": 0.7481481481481481, \"macro avg\": {\"precision\": 0.7977492664992665, \"recall\": 0.7543159965034965, \"f1-score\": 0.7625258134847175, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7574262074262075, \"recall\": 0.7481481481481481, \"f1-score\": 0.7480003473793427, \"support\": 135.0}}",
"eval_f1_macro": 76.25258134847175,
"eval_f1_micro": 74.81481481481481,
"eval_f1_weighted": 74.80003473793427,
"eval_loss": 1.0315518379211426,
"eval_runtime": 2.137,
"eval_samples_per_second": 63.172,
"step": 1632
},
{
"epoch": 97.0,
"learning_rate": 0.00017166666666666665,
"loss": 0.0004,
"step": 1649
},
{
"epoch": 97.0,
"eval_accuracy": 74.81481481481481,
"eval_average_metrics": 75.17542277410698,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7647058823529411, \"recall\": 0.7222222222222222, \"f1-score\": 0.7428571428571428, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7105263157894737, \"recall\": 0.6923076923076923, \"f1-score\": 0.7012987012987013, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.631578947368421, \"recall\": 0.75, \"f1-score\": 0.6857142857142857, \"support\": 16.0}, \"accuracy\": 0.7481481481481481, \"macro avg\": {\"precision\": 0.7985009658384271, \"recall\": 0.7494415306915307, \"f1-score\": 0.7611168071694387, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7574858211390718, \"recall\": 0.7481481481481481, \"f1-score\": 0.7496038074985443, \"support\": 135.0}}",
"eval_f1_macro": 76.11168071694388,
"eval_f1_micro": 74.81481481481481,
"eval_f1_weighted": 74.96038074985442,
"eval_loss": 1.0185447931289673,
"eval_runtime": 2.2095,
"eval_samples_per_second": 61.099,
"step": 1649
},
{
"epoch": 98.0,
"learning_rate": 0.00016999999999999999,
"loss": 0.0081,
"step": 1666
},
{
"epoch": 98.0,
"eval_accuracy": 73.33333333333333,
"eval_average_metrics": 74.67129308878316,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6428571428571429, \"recall\": 0.8181818181818182, \"f1-score\": 0.7200000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7741935483870968, \"recall\": 0.6666666666666666, \"f1-score\": 0.7164179104477612, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7352941176470589, \"recall\": 0.6410256410256411, \"f1-score\": 0.6849315068493151, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7769345152528264, \"recall\": 0.8063993298368298, \"f1-score\": 0.784722770729386, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7505437952307024, \"recall\": 0.7333333333333333, \"f1-score\": 0.735462286155274, \"support\": 135.0}}",
"eval_f1_macro": 78.4722770729386,
"eval_f1_micro": 73.33333333333333,
"eval_f1_weighted": 73.5462286155274,
"eval_loss": 1.0540401935577393,
"eval_runtime": 2.1385,
"eval_samples_per_second": 63.127,
"step": 1666
},
{
"epoch": 99.0,
"learning_rate": 0.00016833333333333332,
"loss": 0.0003,
"step": 1683
},
{
"epoch": 99.0,
"eval_accuracy": 75.55555555555556,
"eval_average_metrics": 76.05696734786595,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7428571428571429, \"recall\": 0.7222222222222222, \"f1-score\": 0.732394366197183, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.6666666666666666, \"f1-score\": 0.7123287671232877, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7555555555555555, \"macro avg\": {\"precision\": 0.8140947288006112, \"recall\": 0.7654125388500388, \"f1-score\": 0.7737856961026234, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7716894434541492, \"recall\": 0.7555555555555555, \"f1-score\": 0.7573818867009036, \"support\": 135.0}}",
"eval_f1_macro": 77.37856961026233,
"eval_f1_micro": 75.55555555555556,
"eval_f1_weighted": 75.73818867009035,
"eval_loss": 1.054269790649414,
"eval_runtime": 2.1572,
"eval_samples_per_second": 62.58,
"step": 1683
},
{
"epoch": 100.0,
"learning_rate": 0.00016666666666666666,
"loss": 0.0016,
"step": 1700
},
{
"epoch": 100.0,
"eval_accuracy": 74.81481481481481,
"eval_average_metrics": 75.19184257143311,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.8181818181818182, \"recall\": 0.8181818181818182, \"f1-score\": 0.8181818181818182, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7741935483870968, \"recall\": 0.6666666666666666, \"f1-score\": 0.7164179104477612, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7368421052631579, \"recall\": 0.717948717948718, \"f1-score\": 0.7272727272727273, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7481481481481481, \"macro avg\": {\"precision\": 0.7990435981204232, \"recall\": 0.7535147144522145, \"f1-score\": 0.7606214838883731, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7652767059388452, \"recall\": 0.7481481481481481, \"f1-score\": 0.7507559226726549, \"support\": 135.0}}",
"eval_f1_macro": 76.06214838883731,
"eval_f1_micro": 74.81481481481481,
"eval_f1_weighted": 75.07559226726548,
"eval_loss": 1.0762172937393188,
"eval_runtime": 2.2054,
"eval_samples_per_second": 61.213,
"step": 1700
},
{
"epoch": 101.0,
"learning_rate": 0.000165,
"loss": 0.0008,
"step": 1717
},
{
"epoch": 101.0,
"eval_accuracy": 76.29629629629629,
"eval_average_metrics": 76.48385951719736,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.8, \"recall\": 0.7272727272727273, \"f1-score\": 0.761904761904762, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8235294117647058, \"recall\": 0.717948717948718, \"f1-score\": 0.767123287671233, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 16.0}, \"accuracy\": 0.762962962962963, \"macro avg\": {\"precision\": 0.8140522875816993, \"recall\": 0.748227466977467, \"f1-score\": 0.7677001436238999, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7779084967320261, \"recall\": 0.762962962962963, \"f1-score\": 0.7657283111380684, \"support\": 135.0}}",
"eval_f1_macro": 76.77001436238999,
"eval_f1_micro": 76.29629629629629,
"eval_f1_weighted": 76.57283111380684,
"eval_loss": 1.0158424377441406,
"eval_runtime": 2.1602,
"eval_samples_per_second": 62.494,
"step": 1717
},
{
"epoch": 102.0,
"learning_rate": 0.0001633333333333333,
"loss": 0.0005,
"step": 1734
},
{
"epoch": 102.0,
"eval_accuracy": 72.5925925925926,
"eval_average_metrics": 72.98289629063926,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.8, \"recall\": 0.7272727272727273, \"f1-score\": 0.761904761904762, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.6388888888888888, \"f1-score\": 0.6865671641791045, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7567567567567568, \"recall\": 0.717948717948718, \"f1-score\": 0.736842105263158, \"support\": 39.0}, \"question\": {\"precision\": 0.5384615384615384, \"recall\": 0.7777777777777778, \"f1-score\": 0.6363636363636364, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7834636668306023, \"recall\": 0.7290634712509713, \"f1-score\": 0.7370998050821901, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.752118603301399, \"recall\": 0.725925925925926, \"f1-score\": 0.7303641946915284, \"support\": 135.0}}",
"eval_f1_macro": 73.709980508219,
"eval_f1_micro": 72.59259259259261,
"eval_f1_weighted": 73.03641946915283,
"eval_loss": 1.0503590106964111,
"eval_runtime": 2.2165,
"eval_samples_per_second": 60.907,
"step": 1734
},
{
"epoch": 103.0,
"learning_rate": 0.00016166666666666665,
"loss": 0.0013,
"step": 1751
},
{
"epoch": 103.0,
"eval_accuracy": 74.07407407407408,
"eval_average_metrics": 74.67005589980079,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7352941176470589, \"recall\": 0.6944444444444444, \"f1-score\": 0.7142857142857144, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.6666666666666666, \"f1-score\": 0.7123287671232877, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.8069940476190476, \"recall\": 0.752324932012432, \"f1-score\": 0.7613581176647959, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7638364975619877, \"recall\": 0.7407407407407407, \"f1-score\": 0.7439626368457539, \"support\": 135.0}}",
"eval_f1_macro": 76.13581176647959,
"eval_f1_micro": 74.07407407407408,
"eval_f1_weighted": 74.39626368457539,
"eval_loss": 1.0846141576766968,
"eval_runtime": 2.2097,
"eval_samples_per_second": 61.093,
"step": 1751
},
{
"epoch": 104.0,
"learning_rate": 0.00015999999999999999,
"loss": 0.0017,
"step": 1768
},
{
"epoch": 104.0,
"eval_accuracy": 67.4074074074074,
"eval_average_metrics": 67.85678034161822,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6428571428571429, \"recall\": 0.8181818181818182, \"f1-score\": 0.7200000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.5833333333333334, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.65, \"recall\": 0.6666666666666666, \"f1-score\": 0.6582278481012659, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.6153846153846154, \"f1-score\": 0.761904761904762, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.6875, \"f1-score\": 0.5789473684210527, \"support\": 16.0}, \"accuracy\": 0.674074074074074, \"macro avg\": {\"precision\": 0.7351551226551227, \"recall\": 0.6922166375291375, \"f1-score\": 0.689884997303385, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7053102453102452, \"recall\": 0.674074074074074, \"f1-score\": 0.6762380682131959, \"support\": 135.0}}",
"eval_f1_macro": 68.9884997303385,
"eval_f1_micro": 67.4074074074074,
"eval_f1_weighted": 67.62380682131959,
"eval_loss": 1.1709874868392944,
"eval_runtime": 2.145,
"eval_samples_per_second": 62.936,
"step": 1768
},
{
"epoch": 105.0,
"learning_rate": 0.00015833333333333332,
"loss": 0.0039,
"step": 1785
},
{
"epoch": 105.0,
"eval_accuracy": 69.62962962962963,
"eval_average_metrics": 70.57626910983969,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7586206896551724, \"recall\": 0.6111111111111112, \"f1-score\": 0.676923076923077, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6410256410256411, \"recall\": 0.6410256410256411, \"f1-score\": 0.6410256410256411, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.6875, \"f1-score\": 0.5789473684210527, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.7753345792138895, \"recall\": 0.7230781371406372, \"f1-score\": 0.7304419012808433, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7212663028754984, \"recall\": 0.6962962962962963, \"f1-score\": 0.7000162705201517, \"support\": 135.0}}",
"eval_f1_macro": 73.04419012808432,
"eval_f1_micro": 69.62962962962963,
"eval_f1_weighted": 70.00162705201517,
"eval_loss": 1.1890654563903809,
"eval_runtime": 2.2179,
"eval_samples_per_second": 60.869,
"step": 1785
},
{
"epoch": 106.0,
"learning_rate": 0.00015666666666666666,
"loss": 0.0005,
"step": 1802
},
{
"epoch": 106.0,
"eval_accuracy": 69.62962962962963,
"eval_average_metrics": 70.60340419650365,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.5555555555555556, \"f1-score\": 0.6451612903225806, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6410256410256411, \"f1-score\": 0.6666666666666666, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.4642857142857143, \"recall\": 0.8125, \"f1-score\": 0.5909090909090908, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.7788739038739039, \"recall\": 0.7317586926961928, \"f1-score\": 0.7311720214719142, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7352949519616188, \"recall\": 0.6962962962962963, \"f1-score\": 0.7003715537956388, \"support\": 135.0}}",
"eval_f1_macro": 73.11720214719142,
"eval_f1_micro": 69.62962962962963,
"eval_f1_weighted": 70.03715537956387,
"eval_loss": 1.2529244422912598,
"eval_runtime": 2.2315,
"eval_samples_per_second": 60.498,
"step": 1802
},
{
"epoch": 107.0,
"learning_rate": 0.000155,
"loss": 0.001,
"step": 1819
},
{
"epoch": 107.0,
"eval_accuracy": 69.62962962962963,
"eval_average_metrics": 70.58287100496284,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7142857142857143, \"recall\": 0.5555555555555556, \"f1-score\": 0.6250000000000001, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.48148148148148145, \"recall\": 0.8125, \"f1-score\": 0.6046511627906976, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.7766354016354016, \"recall\": 0.7317586926961928, \"f1-score\": 0.7314957452929186, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7284128765610247, \"recall\": 0.6962962962962963, \"f1-score\": 0.6992265023130029, \"support\": 135.0}}",
"eval_f1_macro": 73.14957452929185,
"eval_f1_micro": 69.62962962962963,
"eval_f1_weighted": 69.92265023130028,
"eval_loss": 1.2114461660385132,
"eval_runtime": 2.1003,
"eval_samples_per_second": 64.278,
"step": 1819
},
{
"epoch": 108.0,
"learning_rate": 0.0001533333333333333,
"loss": 0.0002,
"step": 1836
},
{
"epoch": 108.0,
"eval_accuracy": 69.62962962962963,
"eval_average_metrics": 70.58367232228676,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.5555555555555556, \"f1-score\": 0.634920634920635, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6756756756756757, \"recall\": 0.6410256410256411, \"f1-score\": 0.6578947368421052, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.8125, \"f1-score\": 0.6190476190476191, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.77743083993084, \"recall\": 0.7317586926961928, \"f1-score\": 0.7323127643359169, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7265083265083264, \"recall\": 0.6962962962962963, \"f1-score\": 0.698441535962961, \"support\": 135.0}}",
"eval_f1_macro": 73.23127643359169,
"eval_f1_micro": 69.62962962962963,
"eval_f1_weighted": 69.8441535962961,
"eval_loss": 1.21000075340271,
"eval_runtime": 2.1341,
"eval_samples_per_second": 63.257,
"step": 1836
},
{
"epoch": 109.0,
"learning_rate": 0.00015166666666666665,
"loss": 0.0041,
"step": 1853
},
{
"epoch": 109.0,
"eval_accuracy": 73.33333333333333,
"eval_average_metrics": 73.92307044896559,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7222222222222222, \"recall\": 0.7222222222222222, \"f1-score\": 0.7222222222222222, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7983946608946609, \"recall\": 0.744779526029526, \"f1-score\": 0.7552466149013464, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7484559884559885, \"recall\": 0.7333333333333333, \"f1-score\": 0.7350095363906106, \"support\": 135.0}}",
"eval_f1_macro": 75.52466149013463,
"eval_f1_micro": 73.33333333333333,
"eval_f1_weighted": 73.50095363906107,
"eval_loss": 1.1299362182617188,
"eval_runtime": 2.2527,
"eval_samples_per_second": 59.928,
"step": 1853
},
{
"epoch": 110.0,
"learning_rate": 0.00015,
"loss": 0.0012,
"step": 1870
},
{
"epoch": 110.0,
"eval_accuracy": 73.33333333333333,
"eval_average_metrics": 73.93367080496068,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7352941176470589, \"recall\": 0.6944444444444444, \"f1-score\": 0.7142857142857144, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.798580771374889, \"recall\": 0.744512432012432, \"f1-score\": 0.7553214077392971, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7485956238897415, \"recall\": 0.7333333333333333, \"f1-score\": 0.7353587577924635, \"support\": 135.0}}",
"eval_f1_macro": 75.53214077392971,
"eval_f1_micro": 73.33333333333333,
"eval_f1_weighted": 73.53587577924635,
"eval_loss": 1.1259392499923706,
"eval_runtime": 2.1641,
"eval_samples_per_second": 62.382,
"step": 1870
},
{
"epoch": 111.0,
"learning_rate": 0.00014833333333333332,
"loss": 0.0016,
"step": 1887
},
{
"epoch": 111.0,
"eval_accuracy": 71.85185185185186,
"eval_average_metrics": 72.5803126086232,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7272727272727273, \"recall\": 0.6666666666666666, \"f1-score\": 0.6956521739130435, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6842105263157895, \"recall\": 0.6666666666666666, \"f1-score\": 0.6753246753246753, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.55, \"recall\": 0.6875, \"f1-score\": 0.6111111111111112, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7890165755297334, \"recall\": 0.7332277097902098, \"f1-score\": 0.7449370393743795, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7351884762411078, \"recall\": 0.7185185185185186, \"f1-score\": 0.7212384279335117, \"support\": 135.0}}",
"eval_f1_macro": 74.49370393743794,
"eval_f1_micro": 71.85185185185186,
"eval_f1_weighted": 72.12384279335117,
"eval_loss": 1.1060606241226196,
"eval_runtime": 2.1888,
"eval_samples_per_second": 61.678,
"step": 1887
},
{
"epoch": 112.0,
"learning_rate": 0.00014666666666666664,
"loss": 0.0008,
"step": 1904
},
{
"epoch": 112.0,
"eval_accuracy": 69.62962962962963,
"eval_average_metrics": 70.58815358410081,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.5833333333333334, \"f1-score\": 0.6562499999999999, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6578947368421053, \"recall\": 0.6410256410256411, \"f1-score\": 0.6493506493506495, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.75, \"f1-score\": 0.6, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.776365629984051, \"recall\": 0.727418414918415, \"f1-score\": 0.7315299716534531, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7238407466477642, \"recall\": 0.6962962962962963, \"f1-score\": 0.6994035791179866, \"support\": 135.0}}",
"eval_f1_macro": 73.15299716534531,
"eval_f1_micro": 69.62962962962963,
"eval_f1_weighted": 69.94035791179867,
"eval_loss": 1.204077959060669,
"eval_runtime": 2.1515,
"eval_samples_per_second": 62.746,
"step": 1904
},
{
"epoch": 113.0,
"learning_rate": 0.000145,
"loss": 0.0009,
"step": 1921
},
{
"epoch": 113.0,
"eval_accuracy": 71.85185185185186,
"eval_average_metrics": 72.56981914255991,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.71875, \"recall\": 0.6388888888888888, \"f1-score\": 0.676470588235294, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7352941176470589, \"recall\": 0.6410256410256411, \"f1-score\": 0.6849315068493151, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7873426359180036, \"recall\": 0.7421753593628594, \"f1-score\": 0.7447551523701981, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7428055060408002, \"recall\": 0.7185185185185186, \"f1-score\": 0.721000576295161, \"support\": 135.0}}",
"eval_f1_macro": 74.4755152370198,
"eval_f1_micro": 71.85185185185186,
"eval_f1_weighted": 72.1000576295161,
"eval_loss": 1.2206599712371826,
"eval_runtime": 2.1932,
"eval_samples_per_second": 61.553,
"step": 1921
},
{
"epoch": 114.0,
"learning_rate": 0.00014333333333333334,
"loss": 0.0008,
"step": 1938
},
{
"epoch": 114.0,
"eval_accuracy": 71.85185185185186,
"eval_average_metrics": 72.50895922645417,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.6388888888888888, \"f1-score\": 0.6865671641791045, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7352941176470589, \"recall\": 0.6410256410256411, \"f1-score\": 0.6849315068493151, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7836120335230867, \"recall\": 0.7421753593628594, \"f1-score\": 0.7418505576965078, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7454529482043714, \"recall\": 0.7185185185185186, \"f1-score\": 0.7214707743246217, \"support\": 135.0}}",
"eval_f1_macro": 74.18505576965077,
"eval_f1_micro": 71.85185185185186,
"eval_f1_weighted": 72.14707743246217,
"eval_loss": 1.201111078262329,
"eval_runtime": 2.1314,
"eval_samples_per_second": 63.338,
"step": 1938
},
{
"epoch": 115.0,
"learning_rate": 0.00014166666666666665,
"loss": 0.0003,
"step": 1955
},
{
"epoch": 115.0,
"eval_accuracy": 71.11111111111111,
"eval_average_metrics": 71.89831172580645,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.71875, \"recall\": 0.6388888888888888, \"f1-score\": 0.676470588235294, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6410256410256411, \"f1-score\": 0.6666666666666666, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5454545454545454, \"recall\": 0.75, \"f1-score\": 0.631578947368421, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7827099116161615, \"recall\": 0.7343628593628594, \"f1-score\": 0.7401694157684197, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7314534231200898, \"recall\": 0.7111111111111111, \"f1-score\": 0.7135408310416161, \"support\": 135.0}}",
"eval_f1_macro": 74.01694157684197,
"eval_f1_micro": 71.11111111111111,
"eval_f1_weighted": 71.35408310416163,
"eval_loss": 1.181020736694336,
"eval_runtime": 2.237,
"eval_samples_per_second": 60.349,
"step": 1955
},
{
"epoch": 116.0,
"learning_rate": 0.00014,
"loss": 0.0002,
"step": 1972
},
{
"epoch": 116.0,
"eval_accuracy": 70.37037037037037,
"eval_average_metrics": 71.25460880133727,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7586206896551724, \"recall\": 0.6111111111111112, \"f1-score\": 0.676923076923077, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6578947368421053, \"recall\": 0.6410256410256411, \"f1-score\": 0.6493506493506495, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5217391304347826, \"recall\": 0.75, \"f1-score\": 0.6153846153846153, \"support\": 16.0}, \"accuracy\": 0.7037037037037037, \"macro avg\": {\"precision\": 0.7801606074952954, \"recall\": 0.7308906371406372, \"f1-score\": 0.7360371831919146, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7287160867554882, \"recall\": 0.7037037037037037, \"f1-score\": 0.7067397614541691, \"support\": 135.0}}",
"eval_f1_macro": 73.60371831919146,
"eval_f1_micro": 70.37037037037037,
"eval_f1_weighted": 70.6739761454169,
"eval_loss": 1.216983675956726,
"eval_runtime": 2.1268,
"eval_samples_per_second": 63.474,
"step": 1972
},
{
"epoch": 117.0,
"learning_rate": 0.00013833333333333333,
"loss": 0.0015,
"step": 1989
},
{
"epoch": 117.0,
"eval_accuracy": 72.5925925925926,
"eval_average_metrics": 73.36956241306592,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.71875, \"recall\": 0.6388888888888888, \"f1-score\": 0.676470588235294, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.8001194760365956, \"recall\": 0.7453804875679876, \"f1-score\": 0.7551997801408894, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7443039001734654, \"recall\": 0.725925925925926, \"f1-score\": 0.7277308645298954, \"support\": 135.0}}",
"eval_f1_macro": 75.51997801408893,
"eval_f1_micro": 72.59259259259261,
"eval_f1_weighted": 72.77308645298953,
"eval_loss": 1.1754764318466187,
"eval_runtime": 2.3156,
"eval_samples_per_second": 58.299,
"step": 1989
},
{
"epoch": 118.0,
"learning_rate": 0.00013666666666666666,
"loss": 0.0011,
"step": 2006
},
{
"epoch": 118.0,
"eval_accuracy": 70.37037037037037,
"eval_average_metrics": 71.17485160829882,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.5833333333333334, \"f1-score\": 0.6562499999999999, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6842105263157895, \"recall\": 0.6666666666666666, \"f1-score\": 0.6753246753246753, \"support\": 39.0}, \"question\": {\"precision\": 0.5384615384615384, \"recall\": 0.7777777777777778, \"f1-score\": 0.6363636363636364, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5217391304347826, \"recall\": 0.75, \"f1-score\": 0.6153846153846153, \"support\": 16.0}, \"accuracy\": 0.7037037037037037, \"macro avg\": {\"precision\": 0.7760871136872282, \"recall\": 0.7306235431235432, \"f1-score\": 0.7325914102148916, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7313728393819928, \"recall\": 0.7037037037037037, \"f1-score\": 0.7069952467096542, \"support\": 135.0}}",
"eval_f1_macro": 73.25914102148916,
"eval_f1_micro": 70.37037037037037,
"eval_f1_weighted": 70.69952467096542,
"eval_loss": 1.1750222444534302,
"eval_runtime": 2.2858,
"eval_samples_per_second": 59.061,
"step": 2006
},
{
"epoch": 119.0,
"learning_rate": 0.000135,
"loss": 0.0001,
"step": 2023
},
{
"epoch": 119.0,
"eval_accuracy": 71.11111111111111,
"eval_average_metrics": 71.73002733490209,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.71875, \"recall\": 0.6388888888888888, \"f1-score\": 0.676470588235294, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5454545454545454, \"recall\": 0.75, \"f1-score\": 0.631578947368421, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7730584332146831, \"recall\": 0.7262043512043512, \"f1-score\": 0.7334226043582956, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7285210035210035, \"recall\": 0.7111111111111111, \"f1-score\": 0.7135562668155659, \"support\": 135.0}}",
"eval_f1_macro": 73.34226043582956,
"eval_f1_micro": 71.11111111111111,
"eval_f1_weighted": 71.35562668155659,
"eval_loss": 1.1494803428649902,
"eval_runtime": 2.2653,
"eval_samples_per_second": 59.593,
"step": 2023
},
{
"epoch": 120.0,
"learning_rate": 0.0001333333333333333,
"loss": 0.0004,
"step": 2040
},
{
"epoch": 120.0,
"eval_accuracy": 74.07407407407408,
"eval_average_metrics": 74.61782293955142,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7352941176470589, \"recall\": 0.6944444444444444, \"f1-score\": 0.7142857142857144, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.6190476190476191, \"recall\": 0.8125, \"f1-score\": 0.7027027027027026, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.8034016636957814, \"recall\": 0.752324932012432, \"f1-score\": 0.7609662631209947, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7564920917862095, \"recall\": 0.7407407407407407, \"f1-score\": 0.7422651729795805, \"support\": 135.0}}",
"eval_f1_macro": 76.09662631209946,
"eval_f1_micro": 74.07407407407408,
"eval_f1_weighted": 74.22651729795805,
"eval_loss": 1.1421889066696167,
"eval_runtime": 2.1875,
"eval_samples_per_second": 61.713,
"step": 2040
},
{
"epoch": 121.0,
"learning_rate": 0.00013166666666666665,
"loss": 0.0023,
"step": 2057
},
{
"epoch": 121.0,
"eval_accuracy": 74.07407407407408,
"eval_average_metrics": 75.5217489806628,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7105263157894737, \"recall\": 0.75, \"f1-score\": 0.7297297297297298, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7666666666666667, \"recall\": 0.5897435897435898, \"f1-score\": 0.6666666666666667, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9090909090909091, \"recall\": 0.7692307692307693, \"f1-score\": 0.8333333333333333, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.7989348370927318, \"recall\": 0.8121539918414918, \"f1-score\": 0.7989948539871141, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7548801296169717, \"recall\": 0.7407407407407407, \"f1-score\": 0.7403936237579168, \"support\": 135.0}}",
"eval_f1_macro": 79.8994853987114,
"eval_f1_micro": 74.07407407407408,
"eval_f1_weighted": 74.03936237579168,
"eval_loss": 1.1364786624908447,
"eval_runtime": 2.1228,
"eval_samples_per_second": 63.596,
"step": 2057
},
{
"epoch": 122.0,
"learning_rate": 0.00013,
"loss": 0.0011,
"step": 2074
},
{
"epoch": 122.0,
"eval_accuracy": 72.5925925925926,
"eval_average_metrics": 74.26085371819231,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.6111111111111112, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6410256410256411, \"f1-score\": 0.7042253521126761, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9090909090909091, \"recall\": 0.7692307692307693, \"f1-score\": 0.8333333333333333, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.875, \"f1-score\": 0.6363636363636364, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7922449945887446, \"recall\": 0.8090156371406372, \"f1-score\": 0.7898259455409633, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7544005531505531, \"recall\": 0.725925925925926, \"f1-score\": 0.728756351334877, \"support\": 135.0}}",
"eval_f1_macro": 78.98259455409632,
"eval_f1_micro": 72.59259259259261,
"eval_f1_weighted": 72.87563513348772,
"eval_loss": 1.221017599105835,
"eval_runtime": 2.2505,
"eval_samples_per_second": 59.985,
"step": 2074
},
{
"epoch": 123.0,
"learning_rate": 0.00012833333333333333,
"loss": 0.0008,
"step": 2091
},
{
"epoch": 123.0,
"eval_accuracy": 71.85185185185186,
"eval_average_metrics": 72.57727733650582,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7586206896551724, \"recall\": 0.6111111111111112, \"f1-score\": 0.676923076923077, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.875, \"f1-score\": 0.6363636363636364, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7952244116037219, \"recall\": 0.7358318764568764, \"f1-score\": 0.7438368834916149, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7506233655658944, \"recall\": 0.7185185185185186, \"f1-score\": 0.7222171729315804, \"support\": 135.0}}",
"eval_f1_macro": 74.38368834916149,
"eval_f1_micro": 71.85185185185186,
"eval_f1_weighted": 72.22171729315805,
"eval_loss": 1.2404636144638062,
"eval_runtime": 2.218,
"eval_samples_per_second": 60.867,
"step": 2091
},
{
"epoch": 124.0,
"learning_rate": 0.00012666666666666666,
"loss": 0.0048,
"step": 2108
},
{
"epoch": 124.0,
"eval_accuracy": 73.33333333333333,
"eval_average_metrics": 74.91986381095504,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.71875, \"recall\": 0.6388888888888888, \"f1-score\": 0.676470588235294, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.6666666666666666, \"f1-score\": 0.7027027027027027, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7971842356013552, \"recall\": 0.8078804875679876, \"f1-score\": 0.7945727056980254, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7516616475312128, \"recall\": 0.7333333333333333, \"f1-score\": 0.7355551800735092, \"support\": 135.0}}",
"eval_f1_macro": 79.45727056980255,
"eval_f1_micro": 73.33333333333333,
"eval_f1_weighted": 73.55551800735093,
"eval_loss": 1.1756157875061035,
"eval_runtime": 2.1659,
"eval_samples_per_second": 62.329,
"step": 2108
},
{
"epoch": 125.0,
"learning_rate": 0.000125,
"loss": 0.0007,
"step": 2125
},
{
"epoch": 125.0,
"eval_accuracy": 71.85185185185186,
"eval_average_metrics": 73.36438346817907,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.6388888888888888, \"f1-score\": 0.6865671641791045, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5454545454545454, \"recall\": 0.75, \"f1-score\": 0.631578947368421, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7759566186985541, \"recall\": 0.7887043512043512, \"f1-score\": 0.7763513430179385, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7347037992199282, \"recall\": 0.7185185185185186, \"f1-score\": 0.721186958672187, \"support\": 135.0}}",
"eval_f1_macro": 77.63513430179385,
"eval_f1_micro": 71.85185185185186,
"eval_f1_weighted": 72.11869586721869,
"eval_loss": 1.212998628616333,
"eval_runtime": 2.2937,
"eval_samples_per_second": 58.857,
"step": 2125
},
{
"epoch": 126.0,
"learning_rate": 0.0001233333333333333,
"loss": 0.0008,
"step": 2142
},
{
"epoch": 126.0,
"eval_accuracy": 69.62962962962963,
"eval_average_metrics": 71.27406229672498,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6896551724137931, \"recall\": 0.5555555555555556, \"f1-score\": 0.6153846153846154, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6410256410256411, \"f1-score\": 0.6944444444444444, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.48148148148148145, \"recall\": 0.8125, \"f1-score\": 0.6046511627906976, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.7571729675177952, \"recall\": 0.7828950563325563, \"f1-score\": 0.7595188378961195, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7231066799904603, \"recall\": 0.6962962962962963, \"f1-score\": 0.6988510613802864, \"support\": 135.0}}",
"eval_f1_macro": 75.95188378961197,
"eval_f1_micro": 69.62962962962963,
"eval_f1_weighted": 69.88510613802865,
"eval_loss": 1.2854887247085571,
"eval_runtime": 2.074,
"eval_samples_per_second": 65.093,
"step": 2142
},
{
"epoch": 127.0,
"learning_rate": 0.00012166666666666665,
"loss": 0.0004,
"step": 2159
},
{
"epoch": 127.0,
"eval_accuracy": 69.62962962962963,
"eval_average_metrics": 70.4292424329562,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.5555555555555556, \"f1-score\": 0.634920634920635, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.4642857142857143, \"recall\": 0.8125, \"f1-score\": 0.5909090909090908, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.7681011118511119, \"recall\": 0.7236001845376845, \"f1-score\": 0.72428547901374, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7304041637374971, \"recall\": 0.6962962962962963, \"f1-score\": 0.7002916257119156, \"support\": 135.0}}",
"eval_f1_macro": 72.42854790137399,
"eval_f1_micro": 69.62962962962963,
"eval_f1_weighted": 70.02916257119156,
"eval_loss": 1.351884365081787,
"eval_runtime": 2.1241,
"eval_samples_per_second": 63.557,
"step": 2159
},
{
"epoch": 128.0,
"learning_rate": 0.00011999999999999999,
"loss": 0.0002,
"step": 2176
},
{
"epoch": 128.0,
"eval_accuracy": 69.62962962962963,
"eval_average_metrics": 70.41850794610482,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.5555555555555556, \"f1-score\": 0.634920634920635, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.48148148148148145, \"recall\": 0.8125, \"f1-score\": 0.6046511627906976, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.7678106428106428, \"recall\": 0.7236001845376845, \"f1-score\": 0.7248628871217477, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7268032082846898, \"recall\": 0.6962962962962963, \"f1-score\": 0.6992848381298528, \"support\": 135.0}}",
"eval_f1_macro": 72.48628871217477,
"eval_f1_micro": 69.62962962962963,
"eval_f1_weighted": 69.92848381298528,
"eval_loss": 1.3572306632995605,
"eval_runtime": 2.1876,
"eval_samples_per_second": 61.71,
"step": 2176
},
{
"epoch": 129.0,
"learning_rate": 0.00011833333333333331,
"loss": 0.001,
"step": 2193
},
{
"epoch": 129.0,
"eval_accuracy": 74.07407407407408,
"eval_average_metrics": 74.61782293955142,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7352941176470589, \"recall\": 0.6944444444444444, \"f1-score\": 0.7142857142857144, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.6190476190476191, \"recall\": 0.8125, \"f1-score\": 0.7027027027027026, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.8034016636957814, \"recall\": 0.752324932012432, \"f1-score\": 0.7609662631209947, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7564920917862095, \"recall\": 0.7407407407407407, \"f1-score\": 0.7422651729795805, \"support\": 135.0}}",
"eval_f1_macro": 76.09662631209946,
"eval_f1_micro": 74.07407407407408,
"eval_f1_weighted": 74.22651729795805,
"eval_loss": 1.2295937538146973,
"eval_runtime": 2.2337,
"eval_samples_per_second": 60.438,
"step": 2193
},
{
"epoch": 130.0,
"learning_rate": 0.00011666666666666665,
"loss": 0.0028,
"step": 2210
},
{
"epoch": 130.0,
"eval_accuracy": 74.07407407407408,
"eval_average_metrics": 75.55886340237367,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.625, \"recall\": 0.9090909090909091, \"f1-score\": 0.7407407407407406, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7352941176470589, \"recall\": 0.6944444444444444, \"f1-score\": 0.7142857142857144, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6153846153846154, \"f1-score\": 0.676056338028169, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6190476190476191, \"recall\": 0.8125, \"f1-score\": 0.7027027027027026, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.7957131716322893, \"recall\": 0.8180300602175602, \"f1-score\": 0.7989535791265285, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7572416508691019, \"recall\": 0.7407407407407407, \"f1-score\": 0.7419194754869369, \"support\": 135.0}}",
"eval_f1_macro": 79.89535791265286,
"eval_f1_micro": 74.07407407407408,
"eval_f1_weighted": 74.1919475486937,
"eval_loss": 1.259170651435852,
"eval_runtime": 2.1066,
"eval_samples_per_second": 64.083,
"step": 2210
},
{
"epoch": 131.0,
"learning_rate": 0.000115,
"loss": 0.0033,
"step": 2227
},
{
"epoch": 131.0,
"eval_accuracy": 70.37037037037037,
"eval_average_metrics": 72.25202775378126,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.625, \"recall\": 0.9090909090909091, \"f1-score\": 0.7407407407407406, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.5555555555555556, \"f1-score\": 0.6451612903225806, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6410256410256411, \"f1-score\": 0.6666666666666666, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.48148148148148145, \"recall\": 0.8125, \"f1-score\": 0.6046511627906976, \"support\": 16.0}, \"accuracy\": 0.7037037037037037, \"macro avg\": {\"precision\": 0.7837695868945869, \"recall\": 0.7942586926961928, \"f1-score\": 0.7756004567204229, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7381803313284795, \"recall\": 0.7037037037037037, \"f1-score\": 0.7070732460234199, \"support\": 135.0}}",
"eval_f1_macro": 77.56004567204229,
"eval_f1_micro": 70.37037037037037,
"eval_f1_weighted": 70.707324602342,
"eval_loss": 1.3266923427581787,
"eval_runtime": 2.1701,
"eval_samples_per_second": 62.208,
"step": 2227
},
{
"epoch": 132.0,
"learning_rate": 0.00011333333333333331,
"loss": 0.0004,
"step": 2244
},
{
"epoch": 132.0,
"eval_accuracy": 69.62962962962963,
"eval_average_metrics": 70.6257688194083,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.625, \"recall\": 0.9090909090909091, \"f1-score\": 0.7407407407407406, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.5555555555555556, \"f1-score\": 0.6451612903225806, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6756756756756757, \"recall\": 0.6410256410256411, \"f1-score\": 0.6578947368421052, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.48148148148148145, \"recall\": 0.8125, \"f1-score\": 0.6046511627906976, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.7814234907984907, \"recall\": 0.7317586926961928, \"f1-score\": 0.7328372988256859, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7327582425730573, \"recall\": 0.6962962962962963, \"f1-score\": 0.699600861358053, \"support\": 135.0}}",
"eval_f1_macro": 73.28372988256861,
"eval_f1_micro": 69.62962962962963,
"eval_f1_weighted": 69.9600861358053,
"eval_loss": 1.33100426197052,
"eval_runtime": 2.1896,
"eval_samples_per_second": 61.654,
"step": 2244
},
{
"epoch": 133.0,
"learning_rate": 0.00011166666666666667,
"loss": 0.0006,
"step": 2261
},
{
"epoch": 133.0,
"eval_accuracy": 71.85185185185186,
"eval_average_metrics": 72.70429974221537,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7960415628762403, \"recall\": 0.7419082653457654, \"f1-score\": 0.7506471755169501, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7390933483406602, \"recall\": 0.7185185185185186, \"f1-score\": 0.7204877771346277, \"support\": 135.0}}",
"eval_f1_macro": 75.06471755169501,
"eval_f1_micro": 71.85185185185186,
"eval_f1_weighted": 72.04877771346277,
"eval_loss": 1.2336255311965942,
"eval_runtime": 2.2249,
"eval_samples_per_second": 60.678,
"step": 2261
},
{
"epoch": 134.0,
"learning_rate": 0.00010999999999999998,
"loss": 0.0001,
"step": 2278
},
{
"epoch": 134.0,
"eval_accuracy": 71.11111111111111,
"eval_average_metrics": 72.02848654875856,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.5555555555555556, \"f1-score\": 0.6451612903225806, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6923076923076923, \"recall\": 0.6923076923076923, \"f1-score\": 0.6923076923076923, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.48148148148148145, \"recall\": 0.8125, \"f1-score\": 0.6046511627906976, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7946632071632072, \"recall\": 0.7381689491064491, \"f1-score\": 0.7445463256662919, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7448381796529944, \"recall\": 0.7111111111111111, \"f1-score\": 0.7143709140618288, \"support\": 135.0}}",
"eval_f1_macro": 74.45463256662919,
"eval_f1_micro": 71.11111111111111,
"eval_f1_weighted": 71.43709140618289,
"eval_loss": 1.3270906209945679,
"eval_runtime": 2.1891,
"eval_samples_per_second": 61.669,
"step": 2278
},
{
"epoch": 135.0,
"learning_rate": 0.00010833333333333333,
"loss": 0.0014,
"step": 2295
},
{
"epoch": 135.0,
"eval_accuracy": 74.07407407407408,
"eval_average_metrics": 74.72547585668991,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7142857142857143, \"recall\": 0.6944444444444444, \"f1-score\": 0.7042253521126761, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6410256410256411, \"f1-score\": 0.7042253521126761, \"support\": 39.0}, \"question\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.56, \"recall\": 0.875, \"f1-score\": 0.6829268292682927, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.8124975198412698, \"recall\": 0.7569323038073038, \"f1-score\": 0.7644743043935498, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7664916225749558, \"recall\": 0.7407407407407407, \"f1-score\": 0.7430632483925649, \"support\": 135.0}}",
"eval_f1_macro": 76.44743043935497,
"eval_f1_micro": 74.07407407407408,
"eval_f1_weighted": 74.30632483925649,
"eval_loss": 1.251107931137085,
"eval_runtime": 2.1766,
"eval_samples_per_second": 62.023,
"step": 2295
},
{
"epoch": 136.0,
"learning_rate": 0.00010666666666666667,
"loss": 0.0024,
"step": 2312
},
{
"epoch": 136.0,
"eval_accuracy": 74.81481481481481,
"eval_average_metrics": 76.36714659524762,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6842105263157895, \"recall\": 0.7222222222222222, \"f1-score\": 0.7027027027027027, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8275862068965517, \"recall\": 0.6153846153846154, \"f1-score\": 0.7058823529411765, \"support\": 39.0}, \"question\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5833333333333334, \"recall\": 0.875, \"f1-score\": 0.7000000000000001, \"support\": 16.0}, \"accuracy\": 0.7481481481481481, \"macro avg\": {\"precision\": 0.8174468138737649, \"recall\": 0.8196993978243978, \"f1-score\": 0.8082919113289957, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7746230198740786, \"recall\": 0.7481481481481481, \"f1-score\": 0.7500976561846127, \"support\": 135.0}}",
"eval_f1_macro": 80.82919113289957,
"eval_f1_micro": 74.81481481481481,
"eval_f1_weighted": 75.00976561846126,
"eval_loss": 1.2201787233352661,
"eval_runtime": 2.1085,
"eval_samples_per_second": 64.026,
"step": 2312
},
{
"epoch": 137.0,
"learning_rate": 0.00010499999999999999,
"loss": 0.0006,
"step": 2329
},
{
"epoch": 137.0,
"eval_accuracy": 71.85185185185186,
"eval_average_metrics": 73.34865168139474,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.76, \"recall\": 0.5277777777777778, \"f1-score\": 0.6229508196721312, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.6666666666666666, \"f1-score\": 0.7123287671232877, \"support\": 39.0}, \"question\": {\"precision\": 0.5384615384615384, \"recall\": 0.7777777777777778, \"f1-score\": 0.6363636363636364, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5384615384615384, \"recall\": 0.875, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7731202865761689, \"recall\": 0.8114194832944832, \"f1-score\": 0.7785896412055908, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7473708731355789, \"recall\": 0.7185185185185186, \"f1-score\": 0.7183193890131623, \"support\": 135.0}}",
"eval_f1_macro": 77.85896412055907,
"eval_f1_micro": 71.85185185185186,
"eval_f1_weighted": 71.83193890131622,
"eval_loss": 1.2545726299285889,
"eval_runtime": 2.0734,
"eval_samples_per_second": 65.11,
"step": 2329
},
{
"epoch": 138.0,
"learning_rate": 0.00010333333333333333,
"loss": 0.0018,
"step": 2346
},
{
"epoch": 138.0,
"eval_accuracy": 74.07407407407408,
"eval_average_metrics": 75.43196457326891,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7272727272727273, \"recall\": 0.6666666666666666, \"f1-score\": 0.6956521739130435, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7878787878787878, \"recall\": 0.6666666666666666, \"f1-score\": 0.7222222222222221, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.7872411405020101, \"recall\": 0.809604458041958, \"f1-score\": 0.792512077294686, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7577424186119838, \"recall\": 0.7407407407407407, \"f1-score\": 0.7432850241545893, \"support\": 135.0}}",
"eval_f1_macro": 79.2512077294686,
"eval_f1_micro": 74.07407407407408,
"eval_f1_weighted": 74.32850241545893,
"eval_loss": 1.1618794202804565,
"eval_runtime": 2.1519,
"eval_samples_per_second": 62.735,
"step": 2346
},
{
"epoch": 139.0,
"learning_rate": 0.00010166666666666667,
"loss": 0.0004,
"step": 2363
},
{
"epoch": 139.0,
"eval_accuracy": 73.33333333333333,
"eval_average_metrics": 74.73814060272393,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7272727272727273, \"recall\": 0.7272727272727273, \"f1-score\": 0.7272727272727273, \"support\": 11.0}, \"challenge\": {\"precision\": 0.696969696969697, \"recall\": 0.6388888888888888, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7714285714285715, \"recall\": 0.6923076923076923, \"f1-score\": 0.7297297297297296, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7857676140284836, \"recall\": 0.7979737276612278, \"f1-score\": 0.7869864182364182, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7477583286278938, \"recall\": 0.7333333333333333, \"f1-score\": 0.7358725392058726, \"support\": 135.0}}",
"eval_f1_macro": 78.69864182364182,
"eval_f1_micro": 73.33333333333333,
"eval_f1_weighted": 73.58725392058726,
"eval_loss": 1.1620222330093384,
"eval_runtime": 2.2656,
"eval_samples_per_second": 59.586,
"step": 2363
},
{
"epoch": 140.0,
"learning_rate": 9.999999999999999e-05,
"loss": 0.0044,
"step": 2380
},
{
"epoch": 140.0,
"eval_accuracy": 71.85185185185186,
"eval_average_metrics": 72.40867003367005,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7272727272727273, \"recall\": 0.7272727272727273, \"f1-score\": 0.7272727272727273, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.5833333333333334, \"f1-score\": 0.6363636363636365, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7368421052631579, \"recall\": 0.717948717948718, \"f1-score\": 0.7272727272727273, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7788792530568847, \"recall\": 0.7317344114219114, \"f1-score\": 0.7391414141414141, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7357835666607596, \"recall\": 0.7185185185185186, \"f1-score\": 0.7201683501683502, \"support\": 135.0}}",
"eval_f1_macro": 73.91414141414143,
"eval_f1_micro": 71.85185185185186,
"eval_f1_weighted": 72.01683501683502,
"eval_loss": 1.2089694738388062,
"eval_runtime": 2.2372,
"eval_samples_per_second": 60.342,
"step": 2380
},
{
"epoch": 141.0,
"learning_rate": 9.833333333333333e-05,
"loss": 0.0005,
"step": 2397
},
{
"epoch": 141.0,
"eval_accuracy": 72.5925925925926,
"eval_average_metrics": 73.0998434515311,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.8181818181818182, \"f1-score\": 0.7826086956521738, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.5833333333333334, \"f1-score\": 0.6461538461538462, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7567567567567568, \"recall\": 0.717948717948718, \"f1-score\": 0.736842105263158, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.780597947085016, \"recall\": 0.7430980477855478, \"f1-score\": 0.7443116919947583, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7462899681290487, \"recall\": 0.725925925925926, \"f1-score\": 0.7278301942146337, \"support\": 135.0}}",
"eval_f1_macro": 74.43116919947583,
"eval_f1_micro": 72.59259259259261,
"eval_f1_weighted": 72.78301942146337,
"eval_loss": 1.2010650634765625,
"eval_runtime": 2.2981,
"eval_samples_per_second": 58.744,
"step": 2397
},
{
"epoch": 142.0,
"learning_rate": 9.666666666666667e-05,
"loss": 0.0001,
"step": 2414
},
{
"epoch": 142.0,
"eval_accuracy": 72.5925925925926,
"eval_average_metrics": 74.19280926182283,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.5833333333333334, \"f1-score\": 0.6461538461538462, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.6666666666666666, \"f1-score\": 0.7123287671232877, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7850638100067613, \"recall\": 0.8105514277389277, \"f1-score\": 0.7877778983036839, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7492036661407858, \"recall\": 0.725925925925926, \"f1-score\": 0.7280826203173772, \"support\": 135.0}}",
"eval_f1_macro": 78.77778983036839,
"eval_f1_micro": 72.59259259259261,
"eval_f1_weighted": 72.80826203173773,
"eval_loss": 1.2088888883590698,
"eval_runtime": 2.1385,
"eval_samples_per_second": 63.128,
"step": 2414
},
{
"epoch": 143.0,
"learning_rate": 9.499999999999999e-05,
"loss": 0.001,
"step": 2431
},
{
"epoch": 143.0,
"eval_accuracy": 74.07407407407408,
"eval_average_metrics": 75.53753615871489,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.6388888888888888, \"f1-score\": 0.6865671641791045, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.6666666666666666, \"f1-score\": 0.7123287671232877, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.793443807141625, \"recall\": 0.8174958721833722, \"f1-score\": 0.797105878846315, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7597858193999876, \"recall\": 0.7407407407407407, \"f1-score\": 0.7429140860207989, \"support\": 135.0}}",
"eval_f1_macro": 79.7105878846315,
"eval_f1_micro": 74.07407407407408,
"eval_f1_weighted": 74.29140860207988,
"eval_loss": 1.1494632959365845,
"eval_runtime": 2.0877,
"eval_samples_per_second": 64.663,
"step": 2431
},
{
"epoch": 144.0,
"learning_rate": 9.333333333333333e-05,
"loss": 0.0003,
"step": 2448
},
{
"epoch": 144.0,
"eval_accuracy": 74.07407407407408,
"eval_average_metrics": 75.5783020422968,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7142857142857143, \"recall\": 0.6944444444444444, \"f1-score\": 0.7042253521126761, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6410256410256411, \"f1-score\": 0.6944444444444444, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.7968614718614719, \"recall\": 0.8134226884226884, \"f1-score\": 0.7990512962136822, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7549655282988617, \"recall\": 0.7407407407407407, \"f1-score\": 0.742599303996708, \"support\": 135.0}}",
"eval_f1_macro": 79.90512962136823,
"eval_f1_micro": 74.07407407407408,
"eval_f1_weighted": 74.25993039967081,
"eval_loss": 1.1091960668563843,
"eval_runtime": 2.1685,
"eval_samples_per_second": 62.254,
"step": 2448
},
{
"epoch": 145.0,
"learning_rate": 9.166666666666667e-05,
"loss": 0.0002,
"step": 2465
},
{
"epoch": 145.0,
"eval_accuracy": 74.81481481481481,
"eval_average_metrics": 76.29216423947909,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6923076923076923, \"recall\": 0.75, \"f1-score\": 0.7199999999999999, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8, \"recall\": 0.6153846153846154, \"f1-score\": 0.6956521739130435, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 16.0}, \"accuracy\": 0.7481481481481481, \"macro avg\": {\"precision\": 0.8073717948717949, \"recall\": 0.8171620046620047, \"f1-score\": 0.8057793565410674, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7656030389363723, \"recall\": 0.7481481481481481, \"f1-score\": 0.7496109167418001, \"support\": 135.0}}",
"eval_f1_macro": 80.57793565410674,
"eval_f1_micro": 74.81481481481481,
"eval_f1_weighted": 74.96109167418001,
"eval_loss": 1.1161596775054932,
"eval_runtime": 2.1314,
"eval_samples_per_second": 63.338,
"step": 2465
},
{
"epoch": 146.0,
"learning_rate": 8.999999999999999e-05,
"loss": 0.0001,
"step": 2482
},
{
"epoch": 146.0,
"eval_accuracy": 75.55555555555556,
"eval_average_metrics": 76.97719093283912,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7105263157894737, \"recall\": 0.75, \"f1-score\": 0.7297297297297298, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8, \"recall\": 0.6153846153846154, \"f1-score\": 0.6956521739130435, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6190476190476191, \"recall\": 0.8125, \"f1-score\": 0.7027027027027026, \"support\": 16.0}, \"accuracy\": 0.7555555555555555, \"macro avg\": {\"precision\": 0.8120300751879699, \"recall\": 0.8249745046620047, \"f1-score\": 0.8115000772617881, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7727188341223429, \"recall\": 0.7555555555555555, \"f1-score\": 0.7564764489406657, \"support\": 135.0}}",
"eval_f1_macro": 81.15000772617881,
"eval_f1_micro": 75.55555555555556,
"eval_f1_weighted": 75.64764489406657,
"eval_loss": 1.1192935705184937,
"eval_runtime": 2.1466,
"eval_samples_per_second": 62.891,
"step": 2482
},
{
"epoch": 147.0,
"learning_rate": 8.833333333333333e-05,
"loss": 0.0004,
"step": 2499
},
{
"epoch": 147.0,
"eval_accuracy": 74.07407407407408,
"eval_average_metrics": 75.58963733628248,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7058823529411765, \"recall\": 0.6666666666666666, \"f1-score\": 0.6857142857142857, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.6666666666666666, \"f1-score\": 0.7123287671232877, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.7987710818593171, \"recall\": 0.809604458041958, \"f1-score\": 0.7987620895510332, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7557962734433322, \"recall\": 0.7407407407407407, \"f1-score\": 0.7433419224187846, \"support\": 135.0}}",
"eval_f1_macro": 79.87620895510332,
"eval_f1_micro": 74.07407407407408,
"eval_f1_weighted": 74.33419224187845,
"eval_loss": 1.14371657371521,
"eval_runtime": 2.2241,
"eval_samples_per_second": 60.698,
"step": 2499
},
{
"epoch": 148.0,
"learning_rate": 8.666666666666665e-05,
"loss": 0.0006,
"step": 2516
},
{
"epoch": 148.0,
"eval_accuracy": 72.5925925925926,
"eval_average_metrics": 73.05647965158836,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.8181818181818182, \"f1-score\": 0.7826086956521738, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.5833333333333334, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7368421052631579, \"recall\": 0.717948717948718, \"f1-score\": 0.7272727272727273, \"support\": 39.0}, \"question\": {\"precision\": 0.5384615384615384, \"recall\": 0.7777777777777778, \"f1-score\": 0.6363636363636364, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7792046221322537, \"recall\": 0.7430980477855478, \"f1-score\": 0.7418917435221783, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7518493527265457, \"recall\": 0.725925925925926, \"f1-score\": 0.7285155906895038, \"support\": 135.0}}",
"eval_f1_macro": 74.18917435221783,
"eval_f1_micro": 72.59259259259261,
"eval_f1_weighted": 72.85155906895038,
"eval_loss": 1.2226479053497314,
"eval_runtime": 2.2006,
"eval_samples_per_second": 61.347,
"step": 2516
},
{
"epoch": 149.0,
"learning_rate": 8.499999999999999e-05,
"loss": 0.0001,
"step": 2533
},
{
"epoch": 149.0,
"eval_accuracy": 72.5925925925926,
"eval_average_metrics": 73.12022510871195,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7586206896551724, \"recall\": 0.6111111111111112, \"f1-score\": 0.676923076923077, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7297297297297297, \"recall\": 0.6923076923076923, \"f1-score\": 0.7105263157894737, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7809471629514733, \"recall\": 0.7433651418026419, \"f1-score\": 0.7449589518668466, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7465120614545901, \"recall\": 0.725925925925926, \"f1-score\": 0.7279982006297796, \"support\": 135.0}}",
"eval_f1_macro": 74.49589518668466,
"eval_f1_micro": 72.59259259259261,
"eval_f1_weighted": 72.79982006297794,
"eval_loss": 1.2157655954360962,
"eval_runtime": 2.1514,
"eval_samples_per_second": 62.749,
"step": 2533
},
{
"epoch": 150.0,
"learning_rate": 8.333333333333333e-05,
"loss": 0.0009,
"step": 2550
},
{
"epoch": 150.0,
"eval_accuracy": 74.81481481481481,
"eval_average_metrics": 76.0197940520429,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7222222222222222, \"recall\": 0.7222222222222222, \"f1-score\": 0.7222222222222222, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.6666666666666666, \"f1-score\": 0.7123287671232877, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6190476190476191, \"recall\": 0.8125, \"f1-score\": 0.7027027027027026, \"support\": 16.0}, \"accuracy\": 0.7481481481481481, \"macro avg\": {\"precision\": 0.7942298714357537, \"recall\": 0.8026600135975136, \"f1-score\": 0.7949344892838044, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7581013031993423, \"recall\": 0.7481481481481481, \"f1-score\": 0.7495609765016158, \"support\": 135.0}}",
"eval_f1_macro": 79.49344892838043,
"eval_f1_micro": 74.81481481481481,
"eval_f1_weighted": 74.95609765016158,
"eval_loss": 1.1733694076538086,
"eval_runtime": 2.2975,
"eval_samples_per_second": 58.759,
"step": 2550
},
{
"epoch": 151.0,
"learning_rate": 8.166666666666665e-05,
"loss": 0.0008,
"step": 2567
},
{
"epoch": 151.0,
"eval_accuracy": 72.5925925925926,
"eval_average_metrics": 73.8567648637555,
"eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7272727272727273, \"recall\": 0.6666666666666666, \"f1-score\": 0.6956521739130435, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.625, \"recall\": 0.5555555555555556, \"f1-score\": 0.5882352941176471, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7776136152494848, \"recall\": 0.7818266802641802, \"f1-score\": 0.7749303779482808, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7380173905536225, \"recall\": 0.725925925925926, \"f1-score\": 0.7274883647500868, \"support\": 135.0}}",
"eval_f1_macro": 77.49303779482808,
"eval_f1_micro": 72.59259259259261,
"eval_f1_weighted": 72.74883647500869,
"eval_loss": 1.1846721172332764,
"eval_runtime": 2.2205,
"eval_samples_per_second": 60.796,
"step": 2567
},
{
"epoch": 152.0,
"learning_rate": 7.999999999999999e-05,
"loss": 0.0001,
"step": 2584
},
{
"epoch": 152.0,
"eval_accuracy": 74.07407407407408,
"eval_average_metrics": 75.57891605906845,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7272727272727273, \"recall\": 0.6666666666666666, \"f1-score\": 0.6956521739130435, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.6666666666666666, \"f1-score\": 0.7027027027027027, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.8020374552983248, \"recall\": 0.8070792055167055, \"f1-score\": 0.7986914184004976, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.755954576824142, \"recall\": 0.7407407407407407, \"f1-score\": 0.7429837424807587, \"support\": 135.0}}",
"eval_f1_macro": 79.86914184004976,
"eval_f1_micro": 74.07407407407408,
"eval_f1_weighted": 74.29837424807587,
"eval_loss": 1.188364863395691,
"eval_runtime": 2.1962,
"eval_samples_per_second": 61.471,
"step": 2584
},
{
"epoch": 153.0,
"learning_rate": 7.833333333333333e-05,
"loss": 0.0001,
"step": 2601
},
{
"epoch": 153.0,
"eval_accuracy": 72.5925925925926,
"eval_average_metrics": 73.21584842556148,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.8181818181818182, \"recall\": 0.8181818181818182, \"f1-score\": 0.8181818181818182, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7272727272727273, \"recall\": 0.6666666666666666, \"f1-score\": 0.6956521739130435, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6744186046511628, \"recall\": 0.7435897435897436, \"f1-score\": 0.7073170731707318, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.55, \"recall\": 0.6875, \"f1-score\": 0.6111111111111112, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.8045674770965469, \"recall\": 0.7175905691530691, \"f1-score\": 0.7470421497111972, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7428455093571373, \"recall\": 0.725925925925926, \"f1-score\": 0.7297399354594096, \"support\": 135.0}}",
"eval_f1_macro": 74.70421497111973,
"eval_f1_micro": 72.59259259259261,
"eval_f1_weighted": 72.97399354594096,
"eval_loss": 1.2244882583618164,
"eval_runtime": 2.2128,
"eval_samples_per_second": 61.008,
"step": 2601
},
{
"epoch": 154.0,
"learning_rate": 7.666666666666666e-05,
"loss": 0.0039,
"step": 2618
},
{
"epoch": 154.0,
"eval_accuracy": 75.55555555555556,
"eval_average_metrics": 76.95897870855806,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7647058823529411, \"recall\": 0.7222222222222222, \"f1-score\": 0.7428571428571428, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7105263157894737, \"recall\": 0.6923076923076923, \"f1-score\": 0.7012987012987013, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.631578947368421, \"recall\": 0.75, \"f1-score\": 0.6857142857142857, \"support\": 16.0}, \"accuracy\": 0.7555555555555555, \"macro avg\": {\"precision\": 0.8151371074745688, \"recall\": 0.8136897824397824, \"f1-score\": 0.8096817403891035, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.766684030337281, \"recall\": 0.7555555555555555, \"f1-score\": 0.7575662968421079, \"support\": 135.0}}",
"eval_f1_macro": 80.96817403891035,
"eval_f1_micro": 75.55555555555556,
"eval_f1_weighted": 75.75662968421078,
"eval_loss": 1.2217811346054077,
"eval_runtime": 2.2436,
"eval_samples_per_second": 60.172,
"step": 2618
},
{
"epoch": 155.0,
"learning_rate": 7.5e-05,
"loss": 0.0003,
"step": 2635
},
{
"epoch": 155.0,
"eval_accuracy": 75.55555555555556,
"eval_average_metrics": 77.04653690264276,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7575757575757576, \"recall\": 0.6944444444444444, \"f1-score\": 0.7246376811594203, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.6666666666666666, \"f1-score\": 0.7027027027027027, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7555555555555555, \"macro avg\": {\"precision\": 0.8132034632034633, \"recall\": 0.8244403166278166, \"f1-score\": 0.8132795190869966, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7693025493025493, \"recall\": 0.7555555555555555, \"f1-score\": 0.7574708459076027, \"support\": 135.0}}",
"eval_f1_macro": 81.32795190869966,
"eval_f1_micro": 75.55555555555556,
"eval_f1_weighted": 75.74708459076027,
"eval_loss": 1.2007545232772827,
"eval_runtime": 2.1766,
"eval_samples_per_second": 62.023,
"step": 2635
},
{
"epoch": 156.0,
"learning_rate": 7.333333333333332e-05,
"loss": 0.0002,
"step": 2652
},
{
"epoch": 156.0,
"eval_accuracy": 75.55555555555556,
"eval_average_metrics": 77.04477692135627,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7352941176470589, \"recall\": 0.6944444444444444, \"f1-score\": 0.7142857142857144, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.6666666666666666, \"f1-score\": 0.7123287671232877, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7555555555555555, \"macro avg\": {\"precision\": 0.8131493506493507, \"recall\": 0.8244403166278166, \"f1-score\": 0.8131887812803564, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7696726367314601, \"recall\": 0.7555555555555555, \"f1-score\": 0.7574911844627835, \"support\": 135.0}}",
"eval_f1_macro": 81.31887812803565,
"eval_f1_micro": 75.55555555555556,
"eval_f1_weighted": 75.74911844627835,
"eval_loss": 1.1942009925842285,
"eval_runtime": 2.3622,
"eval_samples_per_second": 57.15,
"step": 2652
},
{
"epoch": 157.0,
"learning_rate": 7.166666666666667e-05,
"loss": 0.0022,
"step": 2669
},
{
"epoch": 157.0,
"eval_accuracy": 73.33333333333333,
"eval_average_metrics": 74.09230915854165,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.6388888888888888, \"f1-score\": 0.6865671641791045, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.8025137608806963, \"recall\": 0.7549958721833722, \"f1-score\": 0.7614065508371455, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7533344711839336, \"recall\": 0.7333333333333333, \"f1-score\": 0.735619148837854, \"support\": 135.0}}",
"eval_f1_macro": 76.14065508371455,
"eval_f1_micro": 73.33333333333333,
"eval_f1_weighted": 73.56191488378542,
"eval_loss": 1.217727541923523,
"eval_runtime": 2.2066,
"eval_samples_per_second": 61.18,
"step": 2669
},
{
"epoch": 158.0,
"learning_rate": 7e-05,
"loss": 0.0,
"step": 2686
},
{
"epoch": 158.0,
"eval_accuracy": 71.85185185185186,
"eval_average_metrics": 72.62519417356738,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.6111111111111112, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7910440066690066, \"recall\": 0.7419082653457654, \"f1-score\": 0.7472856934535829, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7411591678258346, \"recall\": 0.7185185185185186, \"f1-score\": 0.7206850364520756, \"support\": 135.0}}",
"eval_f1_macro": 74.72856934535828,
"eval_f1_micro": 71.85185185185186,
"eval_f1_weighted": 72.06850364520756,
"eval_loss": 1.2388432025909424,
"eval_runtime": 2.153,
"eval_samples_per_second": 62.703,
"step": 2686
},
{
"epoch": 159.0,
"learning_rate": 6.833333333333333e-05,
"loss": 0.0006,
"step": 2703
},
{
"epoch": 159.0,
"eval_accuracy": 73.33333333333333,
"eval_average_metrics": 74.02792297951471,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.8040257260365956, \"recall\": 0.7488527097902098, \"f1-score\": 0.7588762507291247, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7526372335067987, \"recall\": 0.7333333333333333, \"f1-score\": 0.7355740017847974, \"support\": 135.0}}",
"eval_f1_macro": 75.88762507291247,
"eval_f1_micro": 73.33333333333333,
"eval_f1_weighted": 73.55740017847972,
"eval_loss": 1.221247673034668,
"eval_runtime": 2.273,
"eval_samples_per_second": 59.392,
"step": 2703
},
{
"epoch": 160.0,
"learning_rate": 6.666666666666666e-05,
"loss": 0.0028,
"step": 2720
},
{
"epoch": 160.0,
"eval_accuracy": 74.81481481481481,
"eval_average_metrics": 76.18622883029525,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7428571428571429, \"recall\": 0.7222222222222222, \"f1-score\": 0.732394366197183, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7741935483870968, \"recall\": 0.6153846153846154, \"f1-score\": 0.6857142857142857, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 0.8461538461538461, \"recall\": 0.8461538461538461, \"f1-score\": 0.8461538461538461, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7481481481481481, \"macro avg\": {\"precision\": 0.7960499178241114, \"recall\": 0.8215022824397824, \"f1-score\": 0.8033114500290621, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7596155099380906, \"recall\": 0.7481481481481481, \"f1-score\": 0.7478414068864512, \"support\": 135.0}}",
"eval_f1_macro": 80.33114500290621,
"eval_f1_micro": 74.81481481481481,
"eval_f1_weighted": 74.78414068864512,
"eval_loss": 1.1919903755187988,
"eval_runtime": 2.2082,
"eval_samples_per_second": 61.136,
"step": 2720
},
{
"epoch": 161.0,
"learning_rate": 6.5e-05,
"loss": 0.0005,
"step": 2737
},
{
"epoch": 161.0,
"eval_accuracy": 72.5925925925926,
"eval_average_metrics": 73.24915577342048,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.6111111111111112, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7860111205219901, \"recall\": 0.75152364996115, \"f1-score\": 0.7518137254901962, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7415646457675443, \"recall\": 0.725925925925926, \"f1-score\": 0.7263006535947714, \"support\": 135.0}}",
"eval_f1_macro": 75.1813725490196,
"eval_f1_micro": 72.59259259259261,
"eval_f1_weighted": 72.63006535947713,
"eval_loss": 1.2137112617492676,
"eval_runtime": 2.2438,
"eval_samples_per_second": 60.166,
"step": 2737
},
{
"epoch": 162.0,
"learning_rate": 6.333333333333333e-05,
"loss": 0.0003,
"step": 2754
},
{
"epoch": 162.0,
"eval_accuracy": 73.33333333333333,
"eval_average_metrics": 74.01724012020507,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.6388888888888888, \"f1-score\": 0.6865671641791045, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.797503056005861, \"recall\": 0.7549958721833722, \"f1-score\": 0.7588846210125841, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7518832439356048, \"recall\": 0.7333333333333333, \"f1-score\": 0.7351383171289522, \"support\": 135.0}}",
"eval_f1_macro": 75.88846210125841,
"eval_f1_micro": 73.33333333333333,
"eval_f1_weighted": 73.51383171289521,
"eval_loss": 1.2145878076553345,
"eval_runtime": 2.142,
"eval_samples_per_second": 63.026,
"step": 2754
},
{
"epoch": 163.0,
"learning_rate": 6.166666666666666e-05,
"loss": 0.0007,
"step": 2771
},
{
"epoch": 163.0,
"eval_accuracy": 74.07407407407408,
"eval_average_metrics": 74.5749463235295,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7027027027027027, \"recall\": 0.7222222222222222, \"f1-score\": 0.7123287671232876, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6410256410256411, \"f1-score\": 0.7042253521126761, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.797979519073269, \"recall\": 0.7508437742812742, \"f1-score\": 0.7594312360087268, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.755945511362178, \"recall\": 0.7407407407407407, \"f1-score\": 0.7420851354509721, \"support\": 135.0}}",
"eval_f1_macro": 75.94312360087267,
"eval_f1_micro": 74.07407407407408,
"eval_f1_weighted": 74.2085135450972,
"eval_loss": 1.2377903461456299,
"eval_runtime": 2.1826,
"eval_samples_per_second": 61.852,
"step": 2771
},
{
"epoch": 164.0,
"learning_rate": 5.9999999999999995e-05,
"loss": 0.0012,
"step": 2788
},
{
"epoch": 164.0,
"eval_accuracy": 73.33333333333333,
"eval_average_metrics": 73.9473371413507,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7058823529411765, \"recall\": 0.6666666666666666, \"f1-score\": 0.6857142857142857, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.6666666666666666, \"f1-score\": 0.7123287671232877, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7955596194087242, \"recall\": 0.747104458041958, \"f1-score\": 0.7549024404282261, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7527513312679553, \"recall\": 0.7333333333333333, \"f1-score\": 0.7363243785591355, \"support\": 135.0}}",
"eval_f1_macro": 75.49024404282261,
"eval_f1_micro": 73.33333333333333,
"eval_f1_weighted": 73.63243785591355,
"eval_loss": 1.2320207357406616,
"eval_runtime": 2.1945,
"eval_samples_per_second": 61.519,
"step": 2788
},
{
"epoch": 165.0,
"learning_rate": 5.8333333333333326e-05,
"loss": 0.0002,
"step": 2805
},
{
"epoch": 165.0,
"eval_accuracy": 72.5925925925926,
"eval_average_metrics": 73.35054466230937,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.6111111111111112, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7934839466089466, \"recall\": 0.75152364996115, \"f1-score\": 0.7543137254901962, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7467981401314735, \"recall\": 0.725925925925926, \"f1-score\": 0.7278562091503267, \"support\": 135.0}}",
"eval_f1_macro": 75.43137254901961,
"eval_f1_micro": 72.59259259259261,
"eval_f1_weighted": 72.78562091503268,
"eval_loss": 1.2451300621032715,
"eval_runtime": 2.1694,
"eval_samples_per_second": 62.23,
"step": 2805
},
{
"epoch": 166.0,
"learning_rate": 5.666666666666666e-05,
"loss": 0.0004,
"step": 2822
},
{
"epoch": 166.0,
"eval_accuracy": 72.5925925925926,
"eval_average_metrics": 73.29745663811057,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.6111111111111112, \"f1-score\": 0.6470588235294118, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.6666666666666666, \"f1-score\": 0.7123287671232877, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.56, \"recall\": 0.875, \"f1-score\": 0.6829268292682927, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7926096513780337, \"recall\": 0.7479725135975136, \"f1-score\": 0.7521030279803201, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7472310129957188, \"recall\": 0.725925925925926, \"f1-score\": 0.7279433856922508, \"support\": 135.0}}",
"eval_f1_macro": 75.21030279803202,
"eval_f1_micro": 72.59259259259261,
"eval_f1_weighted": 72.79433856922508,
"eval_loss": 1.2745343446731567,
"eval_runtime": 2.2435,
"eval_samples_per_second": 60.174,
"step": 2822
},
{
"epoch": 167.0,
"learning_rate": 5.499999999999999e-05,
"loss": 0.0001,
"step": 2839
},
{
"epoch": 167.0,
"eval_accuracy": 73.33333333333333,
"eval_average_metrics": 73.96219025098748,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.696969696969697, \"recall\": 0.6388888888888888, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7878787878787878, \"recall\": 0.6666666666666666, \"f1-score\": 0.7222222222222221, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.56, \"recall\": 0.875, \"f1-score\": 0.6829268292682927, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7966899766899767, \"recall\": 0.7514447358197358, \"f1-score\": 0.7557906902598437, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7564506604506605, \"recall\": 0.7333333333333333, \"f1-score\": 0.7360302531129886, \"support\": 135.0}}",
"eval_f1_macro": 75.57906902598438,
"eval_f1_micro": 73.33333333333333,
"eval_f1_weighted": 73.60302531129886,
"eval_loss": 1.2781518697738647,
"eval_runtime": 2.2619,
"eval_samples_per_second": 59.684,
"step": 2839
},
{
"epoch": 168.0,
"learning_rate": 5.333333333333333e-05,
"loss": 0.0002,
"step": 2856
},
{
"epoch": 168.0,
"eval_accuracy": 71.85185185185186,
"eval_average_metrics": 72.56451583121837,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.56, \"recall\": 0.875, \"f1-score\": 0.6829268292682927, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7900713712810488, \"recall\": 0.738357128982129, \"f1-score\": 0.7450481168947851, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.740871934119246, \"recall\": 0.7185185185185186, \"f1-score\": 0.7204954793169126, \"support\": 135.0}}",
"eval_f1_macro": 74.50481168947852,
"eval_f1_micro": 71.85185185185186,
"eval_f1_weighted": 72.04954793169127,
"eval_loss": 1.305252194404602,
"eval_runtime": 2.1319,
"eval_samples_per_second": 63.325,
"step": 2856
},
{
"epoch": 169.0,
"learning_rate": 5.1666666666666664e-05,
"loss": 0.0001,
"step": 2873
},
{
"epoch": 169.0,
"eval_accuracy": 71.11111111111111,
"eval_average_metrics": 71.89929243227385,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.785339764674442, \"recall\": 0.730544628982129, \"f1-score\": 0.7397919123590555, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7330601223074342, \"recall\": 0.7111111111111111, \"f1-score\": 0.7139575627096765, \"support\": 135.0}}",
"eval_f1_macro": 73.97919123590555,
"eval_f1_micro": 71.11111111111111,
"eval_f1_weighted": 71.39575627096765,
"eval_loss": 1.3155133724212646,
"eval_runtime": 2.2317,
"eval_samples_per_second": 60.491,
"step": 2873
},
{
"epoch": 170.0,
"learning_rate": 4.9999999999999996e-05,
"loss": 0.0,
"step": 2890
},
{
"epoch": 170.0,
"eval_accuracy": 71.11111111111111,
"eval_average_metrics": 71.89929243227385,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.785339764674442, \"recall\": 0.730544628982129, \"f1-score\": 0.7397919123590555, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7330601223074342, \"recall\": 0.7111111111111111, \"f1-score\": 0.7139575627096765, \"support\": 135.0}}",
"eval_f1_macro": 73.97919123590555,
"eval_f1_micro": 71.11111111111111,
"eval_f1_weighted": 71.39575627096765,
"eval_loss": 1.3125479221343994,
"eval_runtime": 2.1736,
"eval_samples_per_second": 62.109,
"step": 2890
},
{
"epoch": 171.0,
"learning_rate": 4.8333333333333334e-05,
"loss": 0.0001,
"step": 2907
},
{
"epoch": 171.0,
"eval_accuracy": 71.11111111111111,
"eval_average_metrics": 71.89929243227385,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.785339764674442, \"recall\": 0.730544628982129, \"f1-score\": 0.7397919123590555, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7330601223074342, \"recall\": 0.7111111111111111, \"f1-score\": 0.7139575627096765, \"support\": 135.0}}",
"eval_f1_macro": 73.97919123590555,
"eval_f1_micro": 71.11111111111111,
"eval_f1_weighted": 71.39575627096765,
"eval_loss": 1.3118623495101929,
"eval_runtime": 2.1217,
"eval_samples_per_second": 63.628,
"step": 2907
},
{
"epoch": 172.0,
"learning_rate": 4.6666666666666665e-05,
"loss": 0.0,
"step": 2924
},
{
"epoch": 172.0,
"eval_accuracy": 71.11111111111111,
"eval_average_metrics": 71.89929243227385,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.785339764674442, \"recall\": 0.730544628982129, \"f1-score\": 0.7397919123590555, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7330601223074342, \"recall\": 0.7111111111111111, \"f1-score\": 0.7139575627096765, \"support\": 135.0}}",
"eval_f1_macro": 73.97919123590555,
"eval_f1_micro": 71.11111111111111,
"eval_f1_weighted": 71.39575627096765,
"eval_loss": 1.312318205833435,
"eval_runtime": 2.0073,
"eval_samples_per_second": 67.255,
"step": 2924
},
{
"epoch": 173.0,
"learning_rate": 4.4999999999999996e-05,
"loss": 0.0,
"step": 2941
},
{
"epoch": 173.0,
"eval_accuracy": 71.11111111111111,
"eval_average_metrics": 71.89929243227385,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.785339764674442, \"recall\": 0.730544628982129, \"f1-score\": 0.7397919123590555, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7330601223074342, \"recall\": 0.7111111111111111, \"f1-score\": 0.7139575627096765, \"support\": 135.0}}",
"eval_f1_macro": 73.97919123590555,
"eval_f1_micro": 71.11111111111111,
"eval_f1_weighted": 71.39575627096765,
"eval_loss": 1.3142341375350952,
"eval_runtime": 2.155,
"eval_samples_per_second": 62.645,
"step": 2941
},
{
"epoch": 174.0,
"learning_rate": 4.333333333333333e-05,
"loss": 0.0001,
"step": 2958
},
{
"epoch": 174.0,
"eval_accuracy": 73.33333333333333,
"eval_average_metrics": 73.9747073753716,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6756756756756757, \"recall\": 0.6944444444444444, \"f1-score\": 0.684931506849315, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6410256410256411, \"f1-score\": 0.7042253521126761, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5833333333333334, \"recall\": 0.875, \"f1-score\": 0.7000000000000001, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.8040708376645876, \"recall\": 0.7455686674436675, \"f1-score\": 0.7566759148589195, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7558651279484612, \"recall\": 0.7333333333333333, \"f1-score\": 0.7356457134892781, \"support\": 135.0}}",
"eval_f1_macro": 75.66759148589195,
"eval_f1_micro": 73.33333333333333,
"eval_f1_weighted": 73.5645713489278,
"eval_loss": 1.2926682233810425,
"eval_runtime": 2.0992,
"eval_samples_per_second": 64.311,
"step": 2958
},
{
"epoch": 175.0,
"learning_rate": 4.1666666666666665e-05,
"loss": 0.0001,
"step": 2975
},
{
"epoch": 175.0,
"eval_accuracy": 73.33333333333333,
"eval_average_metrics": 73.9747073753716,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6756756756756757, \"recall\": 0.6944444444444444, \"f1-score\": 0.684931506849315, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6410256410256411, \"f1-score\": 0.7042253521126761, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5833333333333334, \"recall\": 0.875, \"f1-score\": 0.7000000000000001, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.8040708376645876, \"recall\": 0.7455686674436675, \"f1-score\": 0.7566759148589195, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7558651279484612, \"recall\": 0.7333333333333333, \"f1-score\": 0.7356457134892781, \"support\": 135.0}}",
"eval_f1_macro": 75.66759148589195,
"eval_f1_micro": 73.33333333333333,
"eval_f1_weighted": 73.5645713489278,
"eval_loss": 1.2941668033599854,
"eval_runtime": 2.1293,
"eval_samples_per_second": 63.402,
"step": 2975
},
{
"epoch": 176.0,
"learning_rate": 3.9999999999999996e-05,
"loss": 0.0002,
"step": 2992
},
{
"epoch": 176.0,
"eval_accuracy": 71.85185185185186,
"eval_average_metrics": 72.61480524125062,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.8181818181818182, \"f1-score\": 0.7826086956521738, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7105263157894737, \"recall\": 0.6923076923076923, \"f1-score\": 0.7012987012987013, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7935292547718269, \"recall\": 0.7337497571872572, \"f1-score\": 0.7460040211884412, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7400211319000227, \"recall\": 0.7185185185185186, \"f1-score\": 0.7215511514245466, \"support\": 135.0}}",
"eval_f1_macro": 74.60040211884412,
"eval_f1_micro": 71.85185185185186,
"eval_f1_weighted": 72.15511514245466,
"eval_loss": 1.3083090782165527,
"eval_runtime": 2.2303,
"eval_samples_per_second": 60.529,
"step": 2992
},
{
"epoch": 177.0,
"learning_rate": 3.833333333333333e-05,
"loss": 0.0002,
"step": 3009
},
{
"epoch": 177.0,
"eval_accuracy": 71.85185185185186,
"eval_average_metrics": 72.60305283815627,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.8181818181818182, \"f1-score\": 0.7826086956521738, \"support\": 11.0}, \"challenge\": {\"precision\": 0.71875, \"recall\": 0.6388888888888888, \"f1-score\": 0.676470588235294, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6923076923076923, \"recall\": 0.6923076923076923, \"f1-score\": 0.6923076923076923, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5454545454545454, \"recall\": 0.75, \"f1-score\": 0.631578947368421, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7928594842657343, \"recall\": 0.7294094794094794, \"f1-score\": 0.7450467847762234, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7376262626262626, \"recall\": 0.7185185185185186, \"f1-score\": 0.7220382917129899, \"support\": 135.0}}",
"eval_f1_macro": 74.50467847762235,
"eval_f1_micro": 71.85185185185186,
"eval_f1_weighted": 72.203829171299,
"eval_loss": 1.3025906085968018,
"eval_runtime": 2.2827,
"eval_samples_per_second": 59.141,
"step": 3009
},
{
"epoch": 178.0,
"learning_rate": 3.666666666666666e-05,
"loss": 0.0001,
"step": 3026
},
{
"epoch": 178.0,
"eval_accuracy": 72.5925925925926,
"eval_average_metrics": 73.32833124918041,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.8181818181818182, \"recall\": 0.8181818181818182, \"f1-score\": 0.8181818181818182, \"support\": 11.0}, \"challenge\": {\"precision\": 0.71875, \"recall\": 0.6388888888888888, \"f1-score\": 0.676470588235294, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7, \"recall\": 0.717948717948718, \"f1-score\": 0.708860759493671, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5454545454545454, \"recall\": 0.75, \"f1-score\": 0.631578947368421, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.80234375, \"recall\": 0.7326146076146076, \"f1-score\": 0.7515625584906763, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7454040404040404, \"recall\": 0.725925925925926, \"f1-score\": 0.7297188396246881, \"support\": 135.0}}",
"eval_f1_macro": 75.15625584906763,
"eval_f1_micro": 72.59259259259261,
"eval_f1_weighted": 72.97188396246881,
"eval_loss": 1.2811334133148193,
"eval_runtime": 2.3185,
"eval_samples_per_second": 58.228,
"step": 3026
},
{
"epoch": 179.0,
"learning_rate": 3.5e-05,
"loss": 0.002,
"step": 3043
},
{
"epoch": 179.0,
"eval_accuracy": 71.11111111111111,
"eval_average_metrics": 71.94066572336109,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.8181818181818182, \"f1-score\": 0.7826086956521738, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6923076923076923, \"recall\": 0.6923076923076923, \"f1-score\": 0.6923076923076923, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5217391304347826, \"recall\": 0.75, \"f1-score\": 0.6153846153846153, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7887609848076188, \"recall\": 0.7259372571872572, \"f1-score\": 0.740553221987642, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7323961919333588, \"recall\": 0.7111111111111111, \"f1-score\": 0.7148511847245799, \"support\": 135.0}}",
"eval_f1_macro": 74.05532219876419,
"eval_f1_micro": 71.11111111111111,
"eval_f1_weighted": 71.48511847245798,
"eval_loss": 1.3122072219848633,
"eval_runtime": 2.2232,
"eval_samples_per_second": 60.724,
"step": 3043
},
{
"epoch": 180.0,
"learning_rate": 3.333333333333333e-05,
"loss": 0.0001,
"step": 3060
},
{
"epoch": 180.0,
"eval_accuracy": 71.85185185185186,
"eval_average_metrics": 72.62519417356738,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.6111111111111112, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7910440066690066, \"recall\": 0.7419082653457654, \"f1-score\": 0.7472856934535829, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7411591678258346, \"recall\": 0.7185185185185186, \"f1-score\": 0.7206850364520756, \"support\": 135.0}}",
"eval_f1_macro": 74.72856934535828,
"eval_f1_micro": 71.85185185185186,
"eval_f1_weighted": 72.06850364520756,
"eval_loss": 1.3237799406051636,
"eval_runtime": 2.2054,
"eval_samples_per_second": 61.213,
"step": 3060
},
{
"epoch": 181.0,
"learning_rate": 3.1666666666666666e-05,
"loss": 0.0001,
"step": 3077
},
{
"epoch": 181.0,
"eval_accuracy": 71.85185185185186,
"eval_average_metrics": 72.62519417356738,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.6111111111111112, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7910440066690066, \"recall\": 0.7419082653457654, \"f1-score\": 0.7472856934535829, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7411591678258346, \"recall\": 0.7185185185185186, \"f1-score\": 0.7206850364520756, \"support\": 135.0}}",
"eval_f1_macro": 74.72856934535828,
"eval_f1_micro": 71.85185185185186,
"eval_f1_weighted": 72.06850364520756,
"eval_loss": 1.3265109062194824,
"eval_runtime": 2.3696,
"eval_samples_per_second": 56.971,
"step": 3077
},
{
"epoch": 182.0,
"learning_rate": 2.9999999999999997e-05,
"loss": 0.0,
"step": 3094
},
{
"epoch": 182.0,
"eval_accuracy": 71.85185185185186,
"eval_average_metrics": 72.62519417356738,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.6111111111111112, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7910440066690066, \"recall\": 0.7419082653457654, \"f1-score\": 0.7472856934535829, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7411591678258346, \"recall\": 0.7185185185185186, \"f1-score\": 0.7206850364520756, \"support\": 135.0}}",
"eval_f1_macro": 74.72856934535828,
"eval_f1_micro": 71.85185185185186,
"eval_f1_weighted": 72.06850364520756,
"eval_loss": 1.3261035680770874,
"eval_runtime": 2.2698,
"eval_samples_per_second": 59.477,
"step": 3094
},
{
"epoch": 183.0,
"learning_rate": 2.833333333333333e-05,
"loss": 0.0009,
"step": 3111
},
{
"epoch": 183.0,
"eval_accuracy": 70.37037037037037,
"eval_average_metrics": 71.22624100757768,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6842105263157895, \"recall\": 0.6666666666666666, \"f1-score\": 0.6753246753246753, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5217391304347826, \"recall\": 0.75, \"f1-score\": 0.6153846153846153, \"support\": 16.0}, \"accuracy\": 0.7037037037037037, \"macro avg\": {\"precision\": 0.7805373005970924, \"recall\": 0.722732128982129, \"f1-score\": 0.7343542579082432, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7253561559459545, \"recall\": 0.7037037037037037, \"f1-score\": 0.7072879749874571, \"support\": 135.0}}",
"eval_f1_macro": 73.43542579082431,
"eval_f1_micro": 70.37037037037037,
"eval_f1_weighted": 70.7287974987457,
"eval_loss": 1.2959227561950684,
"eval_runtime": 2.1932,
"eval_samples_per_second": 61.554,
"step": 3111
},
{
"epoch": 184.0,
"learning_rate": 2.6666666666666667e-05,
"loss": 0.0,
"step": 3128
},
{
"epoch": 184.0,
"eval_accuracy": 71.85185185185186,
"eval_average_metrics": 72.55073789502072,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7272727272727273, \"recall\": 0.6666666666666666, \"f1-score\": 0.6956521739130435, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6842105263157895, \"recall\": 0.6666666666666666, \"f1-score\": 0.6753246753246753, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.75, \"f1-score\": 0.6486486486486486, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7889478942110522, \"recall\": 0.7296765734265734, \"f1-score\": 0.7433792315665717, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7359373569899886, \"recall\": 0.7185185185185186, \"f1-score\": 0.7216132471972199, \"support\": 135.0}}",
"eval_f1_macro": 74.33792315665717,
"eval_f1_micro": 71.85185185185186,
"eval_f1_weighted": 72.161324719722,
"eval_loss": 1.2773268222808838,
"eval_runtime": 2.157,
"eval_samples_per_second": 62.586,
"step": 3128
},
{
"epoch": 185.0,
"learning_rate": 2.4999999999999998e-05,
"loss": 0.0007,
"step": 3145
},
{
"epoch": 185.0,
"eval_accuracy": 71.85185185185186,
"eval_average_metrics": 72.5453732812465,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7058823529411765, \"recall\": 0.6666666666666666, \"f1-score\": 0.6857142857142857, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.75, \"f1-score\": 0.6486486486486486, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7885856194679723, \"recall\": 0.7296765734265734, \"f1-score\": 0.7432477269156164, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7355754414577943, \"recall\": 0.7185185185185186, \"f1-score\": 0.7215301672972063, \"support\": 135.0}}",
"eval_f1_macro": 74.32477269156163,
"eval_f1_micro": 71.85185185185186,
"eval_f1_weighted": 72.15301672972065,
"eval_loss": 1.2745593786239624,
"eval_runtime": 2.2478,
"eval_samples_per_second": 60.059,
"step": 3145
},
{
"epoch": 186.0,
"learning_rate": 2.3333333333333332e-05,
"loss": 0.0001,
"step": 3162
},
{
"epoch": 186.0,
"eval_accuracy": 71.85185185185186,
"eval_average_metrics": 72.5453732812465,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7058823529411765, \"recall\": 0.6666666666666666, \"f1-score\": 0.6857142857142857, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.75, \"f1-score\": 0.6486486486486486, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7885856194679723, \"recall\": 0.7296765734265734, \"f1-score\": 0.7432477269156164, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7355754414577943, \"recall\": 0.7185185185185186, \"f1-score\": 0.7215301672972063, \"support\": 135.0}}",
"eval_f1_macro": 74.32477269156163,
"eval_f1_micro": 71.85185185185186,
"eval_f1_weighted": 72.15301672972065,
"eval_loss": 1.273395299911499,
"eval_runtime": 2.1937,
"eval_samples_per_second": 61.539,
"step": 3162
},
{
"epoch": 187.0,
"learning_rate": 2.1666666666666664e-05,
"loss": 0.002,
"step": 3179
},
{
"epoch": 187.0,
"eval_accuracy": 71.11111111111111,
"eval_average_metrics": 71.89929243227385,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.785339764674442, \"recall\": 0.730544628982129, \"f1-score\": 0.7397919123590555, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7330601223074342, \"recall\": 0.7111111111111111, \"f1-score\": 0.7139575627096765, \"support\": 135.0}}",
"eval_f1_macro": 73.97919123590555,
"eval_f1_micro": 71.11111111111111,
"eval_f1_weighted": 71.39575627096765,
"eval_loss": 1.2829253673553467,
"eval_runtime": 2.2014,
"eval_samples_per_second": 61.325,
"step": 3179
},
{
"epoch": 188.0,
"learning_rate": 1.9999999999999998e-05,
"loss": 0.0001,
"step": 3196
},
{
"epoch": 188.0,
"eval_accuracy": 71.85185185185186,
"eval_average_metrics": 72.62519417356738,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.6111111111111112, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7910440066690066, \"recall\": 0.7419082653457654, \"f1-score\": 0.7472856934535829, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7411591678258346, \"recall\": 0.7185185185185186, \"f1-score\": 0.7206850364520756, \"support\": 135.0}}",
"eval_f1_macro": 74.72856934535828,
"eval_f1_micro": 71.85185185185186,
"eval_f1_weighted": 72.06850364520756,
"eval_loss": 1.3078842163085938,
"eval_runtime": 2.2833,
"eval_samples_per_second": 59.124,
"step": 3196
},
{
"epoch": 189.0,
"learning_rate": 1.833333333333333e-05,
"loss": 0.0,
"step": 3213
},
{
"epoch": 189.0,
"eval_accuracy": 71.85185185185186,
"eval_average_metrics": 72.62216036599705,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7905269573616347, \"recall\": 0.7419082653457654, \"f1-score\": 0.7471822632362485, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.740489896403875, \"recall\": 0.7185185185185186, \"f1-score\": 0.7206671143665966, \"support\": 135.0}}",
"eval_f1_macro": 74.71822632362485,
"eval_f1_micro": 71.85185185185186,
"eval_f1_weighted": 72.06671143665966,
"eval_loss": 1.2980728149414062,
"eval_runtime": 2.2356,
"eval_samples_per_second": 60.387,
"step": 3213
},
{
"epoch": 190.0,
"learning_rate": 1.6666666666666664e-05,
"loss": 0.0,
"step": 3230
},
{
"epoch": 190.0,
"eval_accuracy": 71.85185185185186,
"eval_average_metrics": 72.62216036599705,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7905269573616347, \"recall\": 0.7419082653457654, \"f1-score\": 0.7471822632362485, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.740489896403875, \"recall\": 0.7185185185185186, \"f1-score\": 0.7206671143665966, \"support\": 135.0}}",
"eval_f1_macro": 74.71822632362485,
"eval_f1_micro": 71.85185185185186,
"eval_f1_weighted": 72.06671143665966,
"eval_loss": 1.2942575216293335,
"eval_runtime": 2.1696,
"eval_samples_per_second": 62.223,
"step": 3230
},
{
"epoch": 191.0,
"learning_rate": 1.4999999999999999e-05,
"loss": 0.0001,
"step": 3247
},
{
"epoch": 191.0,
"eval_accuracy": 71.11111111111111,
"eval_average_metrics": 71.974518342938,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.6111111111111112, \"f1-score\": 0.6470588235294118, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7867627164502164, \"recall\": 0.7387031371406372, \"f1-score\": 0.7437678567314118, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7322831489498156, \"recall\": 0.7111111111111111, \"f1-score\": 0.7129906547638859, \"support\": 135.0}}",
"eval_f1_macro": 74.37678567314117,
"eval_f1_micro": 71.11111111111111,
"eval_f1_weighted": 71.29906547638859,
"eval_loss": 1.293574571609497,
"eval_runtime": 2.1704,
"eval_samples_per_second": 62.2,
"step": 3247
},
{
"epoch": 192.0,
"learning_rate": 1.3333333333333333e-05,
"loss": 0.0,
"step": 3264
},
{
"epoch": 192.0,
"eval_accuracy": 71.85185185185186,
"eval_average_metrics": 72.62216036599705,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7905269573616347, \"recall\": 0.7419082653457654, \"f1-score\": 0.7471822632362485, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.740489896403875, \"recall\": 0.7185185185185186, \"f1-score\": 0.7206671143665966, \"support\": 135.0}}",
"eval_f1_macro": 74.71822632362485,
"eval_f1_micro": 71.85185185185186,
"eval_f1_weighted": 72.06671143665966,
"eval_loss": 1.2941248416900635,
"eval_runtime": 2.1999,
"eval_samples_per_second": 61.367,
"step": 3264
},
{
"epoch": 193.0,
"learning_rate": 1.1666666666666666e-05,
"loss": 0.0001,
"step": 3281
},
{
"epoch": 193.0,
"eval_accuracy": 71.11111111111111,
"eval_average_metrics": 71.89912419879386,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.6111111111111112, \"f1-score\": 0.6470588235294118, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7850075271950272, \"recall\": 0.730544628982129, \"f1-score\": 0.739725063938619, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7327851161184494, \"recall\": 0.7111111111111111, \"f1-score\": 0.714017681790913, \"support\": 135.0}}",
"eval_f1_macro": 73.9725063938619,
"eval_f1_micro": 71.11111111111111,
"eval_f1_weighted": 71.40176817909129,
"eval_loss": 1.2996220588684082,
"eval_runtime": 2.3185,
"eval_samples_per_second": 58.228,
"step": 3281
},
{
"epoch": 194.0,
"learning_rate": 9.999999999999999e-06,
"loss": 0.0001,
"step": 3298
},
{
"epoch": 194.0,
"eval_accuracy": 71.11111111111111,
"eval_average_metrics": 71.89912419879386,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.6111111111111112, \"f1-score\": 0.6470588235294118, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7850075271950272, \"recall\": 0.730544628982129, \"f1-score\": 0.739725063938619, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7327851161184494, \"recall\": 0.7111111111111111, \"f1-score\": 0.714017681790913, \"support\": 135.0}}",
"eval_f1_macro": 73.9725063938619,
"eval_f1_micro": 71.11111111111111,
"eval_f1_weighted": 71.40176817909129,
"eval_loss": 1.3008208274841309,
"eval_runtime": 2.2664,
"eval_samples_per_second": 59.567,
"step": 3298
},
{
"epoch": 195.0,
"learning_rate": 8.333333333333332e-06,
"loss": 0.0001,
"step": 3315
},
{
"epoch": 195.0,
"eval_accuracy": 70.37037037037037,
"eval_average_metrics": 71.2542231596473,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.6111111111111112, \"f1-score\": 0.6376811594202899, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7037037037037037, \"macro avg\": {\"precision\": 0.781411297036297, \"recall\": 0.7273395007770008, \"f1-score\": 0.7363456487177715, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7249367916034583, \"recall\": 0.7037037037037037, \"f1-score\": 0.7064158702607126, \"support\": 135.0}}",
"eval_f1_macro": 73.63456487177716,
"eval_f1_micro": 70.37037037037037,
"eval_f1_weighted": 70.64158702607126,
"eval_loss": 1.2944679260253906,
"eval_runtime": 2.2091,
"eval_samples_per_second": 61.112,
"step": 3315
},
{
"epoch": 196.0,
"learning_rate": 6.666666666666667e-06,
"loss": 0.0154,
"step": 3332
},
{
"epoch": 196.0,
"eval_accuracy": 70.37037037037037,
"eval_average_metrics": 71.2542231596473,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.6111111111111112, \"f1-score\": 0.6376811594202899, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7037037037037037, \"macro avg\": {\"precision\": 0.781411297036297, \"recall\": 0.7273395007770008, \"f1-score\": 0.7363456487177715, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7249367916034583, \"recall\": 0.7037037037037037, \"f1-score\": 0.7064158702607126, \"support\": 135.0}}",
"eval_f1_macro": 73.63456487177716,
"eval_f1_micro": 70.37037037037037,
"eval_f1_weighted": 70.64158702607126,
"eval_loss": 1.2910572290420532,
"eval_runtime": 2.3432,
"eval_samples_per_second": 57.614,
"step": 3332
},
{
"epoch": 197.0,
"learning_rate": 4.9999999999999996e-06,
"loss": 0.0003,
"step": 3349
},
{
"epoch": 197.0,
"eval_accuracy": 71.11111111111111,
"eval_average_metrics": 71.91662203626719,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6764705882352942, \"recall\": 0.6388888888888888, \"f1-score\": 0.6571428571428571, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7855806278120856, \"recall\": 0.730811722999223, \"f1-score\": 0.7408616942664257, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7303423676825213, \"recall\": 0.7111111111111111, \"f1-score\": 0.7135809649620393, \"support\": 135.0}}",
"eval_f1_macro": 74.08616942664258,
"eval_f1_micro": 71.11111111111111,
"eval_f1_weighted": 71.35809649620393,
"eval_loss": 1.287298560142517,
"eval_runtime": 2.18,
"eval_samples_per_second": 61.926,
"step": 3349
},
{
"epoch": 198.0,
"learning_rate": 3.3333333333333333e-06,
"loss": 0.0001,
"step": 3366
},
{
"epoch": 198.0,
"eval_accuracy": 71.11111111111111,
"eval_average_metrics": 71.91662203626719,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6764705882352942, \"recall\": 0.6388888888888888, \"f1-score\": 0.6571428571428571, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7855806278120856, \"recall\": 0.730811722999223, \"f1-score\": 0.7408616942664257, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7303423676825213, \"recall\": 0.7111111111111111, \"f1-score\": 0.7135809649620393, \"support\": 135.0}}",
"eval_f1_macro": 74.08616942664258,
"eval_f1_micro": 71.11111111111111,
"eval_f1_weighted": 71.35809649620393,
"eval_loss": 1.2868810892105103,
"eval_runtime": 2.1185,
"eval_samples_per_second": 63.724,
"step": 3366
},
{
"epoch": 199.0,
"learning_rate": 1.6666666666666667e-06,
"loss": 0.0002,
"step": 3383
},
{
"epoch": 199.0,
"eval_accuracy": 71.11111111111111,
"eval_average_metrics": 71.91662203626719,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6764705882352942, \"recall\": 0.6388888888888888, \"f1-score\": 0.6571428571428571, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7855806278120856, \"recall\": 0.730811722999223, \"f1-score\": 0.7408616942664257, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7303423676825213, \"recall\": 0.7111111111111111, \"f1-score\": 0.7135809649620393, \"support\": 135.0}}",
"eval_f1_macro": 74.08616942664258,
"eval_f1_micro": 71.11111111111111,
"eval_f1_weighted": 71.35809649620393,
"eval_loss": 1.2866387367248535,
"eval_runtime": 2.1869,
"eval_samples_per_second": 61.732,
"step": 3383
},
{
"epoch": 200.0,
"learning_rate": 0.0,
"loss": 0.0081,
"step": 3400
},
{
"epoch": 200.0,
"eval_accuracy": 71.11111111111111,
"eval_average_metrics": 71.91662203626719,
"eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6764705882352942, \"recall\": 0.6388888888888888, \"f1-score\": 0.6571428571428571, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7855806278120856, \"recall\": 0.730811722999223, \"f1-score\": 0.7408616942664257, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7303423676825213, \"recall\": 0.7111111111111111, \"f1-score\": 0.7135809649620393, \"support\": 135.0}}",
"eval_f1_macro": 74.08616942664258,
"eval_f1_micro": 71.11111111111111,
"eval_f1_weighted": 71.35809649620393,
"eval_loss": 1.2865550518035889,
"eval_runtime": 2.1903,
"eval_samples_per_second": 61.635,
"step": 3400
},
{
"epoch": 200.0,
"step": 3400,
"total_flos": 1.9410334099963904e+16,
"train_runtime": 2245.6861,
"train_samples_per_second": 1.514
}
],
"max_steps": 3400,
"num_train_epochs": 200,
"total_flos": 1.9410334099963904e+16,
"trial_name": null,
"trial_params": null
}