{ "best_metric": 78.31433183733813, "best_model_checkpoint": "/mounts/work/faeze/attempt/new_setting_outputs/source_adapter/crisis_9/3407/checkpoint-512", "epoch": 200.0, "global_step": 3200, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 1.0, "learning_rate": 1.4999999999999999e-05, "loss": 9.307, "step": 16 }, { "epoch": 1.0, "eval_accuracy": 0.0, "eval_average_metrics": 0.0, "eval_classification_report": "{\"\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"\\\"media\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"How Dar\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"-\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \".\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"0:\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \":\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": D\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": Lots\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": apparently\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": good\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": please\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": very\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"????\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Anders\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Emers\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Can use passport\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"False\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"No Friday Im\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"No they have\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"PPC 2019\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"SCP-04\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Shawn\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"True\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a friend\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"canada is under\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 37.0}, \"instagram\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"vac\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"weather\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.0, \"macro avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}}", "eval_f1_macro": 0.0, "eval_f1_micro": 0.0, "eval_f1_weighted": 0.0, "eval_loss": 9.491299629211426, "eval_runtime": 2.2405, "eval_samples_per_second": 51.774, "step": 16 }, { "epoch": 2.0, "learning_rate": 2.9999999999999997e-05, "loss": 8.9068, "step": 32 }, { "epoch": 2.0, "eval_accuracy": 0.0, "eval_average_metrics": 0.0, "eval_classification_report": "{\"\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"\\\"media\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"How Dar\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"-\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": Lots\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": apparently\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": good\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": please\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": very\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Anders\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Emers\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Can use passport\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Climate change is\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Ddx\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"False\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"No Friday Im\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"No they have\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"PPC 2019\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Shawn\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"True\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a friend\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"canada is under\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 37.0}, \"instagram\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"negative\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"weather\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.0, \"macro avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}}", "eval_f1_macro": 0.0, "eval_f1_micro": 0.0, "eval_f1_weighted": 0.0, "eval_loss": 9.213845252990723, "eval_runtime": 2.3697, "eval_samples_per_second": 48.951, "step": 32 }, { "epoch": 3.0, "learning_rate": 4.4999999999999996e-05, "loss": 8.1759, "step": 48 }, { "epoch": 3.0, "eval_accuracy": 0.0, "eval_average_metrics": 0.0, "eval_classification_report": "{\"\\\"How Dar\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"-\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": Lots\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": good\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": please\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": very\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Can use passport\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Climate change is\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Ddx\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"False\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"No Friday Im\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"PPC 2019\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"True\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a friend\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"canada is under\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 31.0}, \"entail\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 37.0}, \"instagram\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"negative\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_e\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"weather\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.0, \"macro avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}}", "eval_f1_macro": 0.0, "eval_f1_micro": 0.0, "eval_f1_weighted": 0.0, "eval_loss": 8.574390411376953, "eval_runtime": 2.1785, "eval_samples_per_second": 53.247, "step": 48 }, { "epoch": 4.0, "learning_rate": 5.9999999999999995e-05, "loss": 6.8531, "step": 64 }, { "epoch": 4.0, "eval_accuracy": 0.0, "eval_average_metrics": 0.0, "eval_classification_report": "{\"\\\"How Dar\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"Reality\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"????\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Australian\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Ddx\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"False\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"SCP-04\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"True\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a friend\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a real\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"canada\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 31.0}, \"classification: good\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"classification: very\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 37.0}, \"negative\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_e\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_en\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"panne\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"please show us\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"vac\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"weather\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.0, \"macro avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}}", "eval_f1_macro": 0.0, "eval_f1_micro": 0.0, "eval_f1_weighted": 0.0, "eval_loss": 6.66550874710083, "eval_runtime": 2.3739, "eval_samples_per_second": 48.865, "step": 64 }, { "epoch": 5.0, "learning_rate": 7.5e-05, "loss": 4.9379, "step": 80 }, { "epoch": 5.0, "eval_accuracy": 0.0, "eval_average_metrics": 0.0, "eval_classification_report": "{\"\\\"Reality\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"-54\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \".\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"0\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"1\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \":D\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"?\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"????\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Australian\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Bruh\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Ddx\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Florida\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"India\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"PPC\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"SCP-04\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"True\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"canada\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 31.0}, \"classification\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"classification: good\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"dal\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"eder\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"entail\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 37.0}, \"instagram\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"microwave\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"name\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"negative\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_e\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_en\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"trud\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"u\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"vac\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"weather\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.0, \"macro avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}}", "eval_f1_macro": 0.0, "eval_f1_micro": 0.0, "eval_f1_weighted": 0.0, "eval_loss": 3.4962198734283447, "eval_runtime": 2.0892, "eval_samples_per_second": 55.523, "step": 80 }, { "epoch": 6.0, "learning_rate": 8.999999999999999e-05, "loss": 2.6265, "step": 96 }, { "epoch": 6.0, "eval_accuracy": 12.068965517241379, "eval_average_metrics": 10.647060714963407, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"challenge\": {\"precision\": 0.2857142857142857, \"recall\": 0.3870967741935484, \"f1-score\": 0.3287671232876712, \"support\": 31.0}, \"entail\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 37.0}, \"negative\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_en\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.10526315789473684, \"recall\": 0.2222222222222222, \"f1-score\": 0.14285714285714285, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.1206896551724138, \"macro avg\": {\"precision\": 0.03007518796992481, \"recall\": 0.04687069203198235, \"f1-score\": 0.03627878970344724, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.0845216489499611, \"recall\": 0.1206896551724138, \"f1-score\": 0.09894392334165598, \"support\": 116.0}}", "eval_f1_macro": 5.240269623831267, "eval_f1_micro": 15.384615384615385, "eval_f1_weighted": 9.894392334165598, "eval_loss": 1.4142650365829468, "eval_runtime": 1.9052, "eval_samples_per_second": 60.887, "step": 96 }, { "epoch": 7.0, "learning_rate": 0.00010499999999999999, "loss": 1.1522, "step": 112 }, { "epoch": 7.0, "eval_accuracy": 31.896551724137932, "eval_average_metrics": 21.324233716475096, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.3333333333333333, \"recall\": 1.0, \"f1-score\": 0.5, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"accuracy\": 0.31896551724137934, \"macro avg\": {\"precision\": 0.041666666666666664, \"recall\": 0.125, \"f1-score\": 0.0625, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.10632183908045977, \"recall\": 0.31896551724137934, \"f1-score\": 0.15948275862068967, \"support\": 116.0}}", "eval_f1_macro": 5.555555555555555, "eval_f1_micro": 31.896551724137932, "eval_f1_weighted": 15.948275862068966, "eval_loss": 0.7661405205726624, "eval_runtime": 2.2446, "eval_samples_per_second": 51.68, "step": 112 }, { "epoch": 8.0, "learning_rate": 0.00011999999999999999, "loss": 0.7426, "step": 128 }, { "epoch": 8.0, "eval_accuracy": 38.793103448275865, "eval_average_metrics": 30.810095419319552, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6, \"recall\": 0.3333333333333333, \"f1-score\": 0.42857142857142855, \"support\": 9.0}, \"challenge\": {\"precision\": 0.3611111111111111, \"recall\": 0.41935483870967744, \"f1-score\": 0.3880597014925373, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.4027777777777778, \"recall\": 0.7837837837837838, \"f1-score\": 0.5321100917431194, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"accuracy\": 0.3879310344827586, \"macro avg\": {\"precision\": 0.17048611111111112, \"recall\": 0.19205899447834932, \"f1-score\": 0.16859265272588567, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.2715277777777778, \"recall\": 0.3879310344827586, \"f1-score\": 0.3066816120509218, \"support\": 116.0}}", "eval_f1_macro": 14.986013575634283, "eval_f1_micro": 38.793103448275865, "eval_f1_weighted": 30.668161205092183, "eval_loss": 0.6721057891845703, "eval_runtime": 2.0959, "eval_samples_per_second": 55.347, "step": 128 }, { "epoch": 9.0, "learning_rate": 0.000135, "loss": 0.7745, "step": 144 }, { "epoch": 9.0, "eval_accuracy": 31.03448275862069, "eval_average_metrics": 22.98618128359508, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.3333333333333333, \"f1-score\": 0.46153846153846156, \"support\": 9.0}, \"challenge\": {\"precision\": 0.28440366972477066, \"recall\": 1.0, \"f1-score\": 0.44285714285714284, \"support\": 31.0}, \"informing statement\": {\"precision\": 1.0, \"recall\": 0.05405405405405406, \"f1-score\": 0.10256410256410257, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"accuracy\": 0.3103448275862069, \"macro avg\": {\"precision\": 0.25430045871559637, \"recall\": 0.17342342342342343, \"f1-score\": 0.12586996336996337, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.45315960139196454, \"recall\": 0.3103448275862069, \"f1-score\": 0.18687318428697738, \"support\": 116.0}}", "eval_f1_macro": 11.188441188441187, "eval_f1_micro": 31.03448275862069, "eval_f1_weighted": 18.68731842869774, "eval_loss": 0.7225409150123596, "eval_runtime": 2.094, "eval_samples_per_second": 55.396, "step": 144 }, { "epoch": 10.0, "learning_rate": 0.00015, "loss": 0.725, "step": 160 }, { "epoch": 10.0, "eval_accuracy": 33.62068965517241, "eval_average_metrics": 26.30665126493825, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.3333333333333333, \"f1-score\": 0.46153846153846156, \"support\": 9.0}, \"challenge\": {\"precision\": 0.3118279569892473, \"recall\": 0.9354838709677419, \"f1-score\": 0.46774193548387094, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.4666666666666667, \"recall\": 0.1891891891891892, \"f1-score\": 0.2692307692307693, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"accuracy\": 0.33620689655172414, \"macro avg\": {\"precision\": 0.19106182795698923, \"recall\": 0.18225079918628306, \"f1-score\": 0.14981389578163773, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.29037356321839075, \"recall\": 0.33620689655172414, \"f1-score\": 0.24668435013262602, \"support\": 116.0}}", "eval_f1_macro": 13.316790736145576, "eval_f1_micro": 33.62068965517241, "eval_f1_weighted": 24.668435013262602, "eval_loss": 0.6647224426269531, "eval_runtime": 2.1313, "eval_samples_per_second": 54.427, "step": 160 }, { "epoch": 11.0, "learning_rate": 0.000165, "loss": 0.7576, "step": 176 }, { "epoch": 11.0, "eval_accuracy": 33.62068965517241, "eval_average_metrics": 25.33956739840334, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.5, \"recall\": 0.5555555555555556, \"f1-score\": 0.5263157894736842, \"support\": 9.0}, \"challenge\": {\"precision\": 0.3163265306122449, \"recall\": 1.0, \"f1-score\": 0.4806201550387597, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.5, \"recall\": 0.08108108108108109, \"f1-score\": 0.13953488372093023, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"accuracy\": 0.33620689655172414, \"macro avg\": {\"precision\": 0.16454081632653061, \"recall\": 0.2045795795795796, \"f1-score\": 0.14330885352917178, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.28281140042223785, \"recall\": 0.33620689655172414, \"f1-score\": 0.21378325525119937, \"support\": 116.0}}", "eval_f1_macro": 12.738564758148602, "eval_f1_micro": 33.62068965517241, "eval_f1_weighted": 21.378325525119937, "eval_loss": 0.6687206029891968, "eval_runtime": 2.1419, "eval_samples_per_second": 54.157, "step": 176 }, { "epoch": 12.0, "learning_rate": 0.00017999999999999998, "loss": 0.6978, "step": 192 }, { "epoch": 12.0, "eval_accuracy": 36.206896551724135, "eval_average_metrics": 28.099909132802715, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6, \"recall\": 0.3333333333333333, \"f1-score\": 0.42857142857142855, \"support\": 9.0}, \"challenge\": {\"precision\": 0.5, \"recall\": 0.16129032258064516, \"f1-score\": 0.24390243902439024, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.3695652173913043, \"recall\": 0.918918918918919, \"f1-score\": 0.5271317829457364, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"accuracy\": 0.3620689655172414, \"macro avg\": {\"precision\": 0.18369565217391304, \"recall\": 0.17669282185411217, \"f1-score\": 0.14995070631769442, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.2980509745127436, \"recall\": 0.3620689655172414, \"f1-score\": 0.2665689175507862, \"support\": 116.0}}", "eval_f1_macro": 13.328951672683944, "eval_f1_micro": 36.20689655172414, "eval_f1_weighted": 26.656891755078625, "eval_loss": 0.6495026350021362, "eval_runtime": 2.1678, "eval_samples_per_second": 53.51, "step": 192 }, { "epoch": 13.0, "learning_rate": 0.000195, "loss": 0.7123, "step": 208 }, { "epoch": 13.0, "eval_accuracy": 36.206896551724135, "eval_average_metrics": 31.268398801131113, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.07407407407407407, \"recall\": 0.16666666666666666, \"f1-score\": 0.10256410256410256, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.4146341463414634, \"recall\": 0.5483870967741935, \"f1-score\": 0.4722222222222222, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.47058823529411764, \"recall\": 0.43243243243243246, \"f1-score\": 0.4507042253521127, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"accuracy\": 0.3620689655172414, \"macro avg\": {\"precision\": 0.19945751150916147, \"recall\": 0.2406579967063838, \"f1-score\": 0.2156863187673047, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.31794469705714945, \"recall\": 0.3620689655172414, \"f1-score\": 0.3348768487731576, \"support\": 116.0}}", "eval_f1_macro": 19.172117223760416, "eval_f1_micro": 36.20689655172414, "eval_f1_weighted": 33.48768487731576, "eval_loss": 0.6175509691238403, "eval_runtime": 2.2362, "eval_samples_per_second": 51.874, "step": 208 }, { "epoch": 14.0, "learning_rate": 0.00020999999999999998, "loss": 0.6115, "step": 224 }, { "epoch": 14.0, "eval_accuracy": 45.689655172413794, "eval_average_metrics": 37.07872659990352, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.4262295081967213, \"recall\": 0.8387096774193549, \"f1-score\": 0.5652173913043479, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.5, \"recall\": 0.5405405405405406, \"f1-score\": 0.5194805194805195, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"accuracy\": 0.45689655172413796, \"macro avg\": {\"precision\": 0.1953241430700447, \"recall\": 0.2696284994672091, \"f1-score\": 0.22308723884810844, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.3227619610463025, \"recall\": 0.45689655172413796, \"f1-score\": 0.3710561926828794, \"support\": 116.0}}", "eval_f1_macro": 19.829976786498527, "eval_f1_micro": 45.689655172413794, "eval_f1_weighted": 37.105619268287946, "eval_loss": 0.5956699252128601, "eval_runtime": 2.0913, "eval_samples_per_second": 55.468, "step": 224 }, { "epoch": 15.0, "learning_rate": 0.000225, "loss": 0.632, "step": 240 }, { "epoch": 15.0, "eval_accuracy": 43.103448275862064, "eval_average_metrics": 35.833906806036026, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.5333333333333333, \"recall\": 0.8888888888888888, \"f1-score\": 0.6666666666666667, \"support\": 9.0}, \"challenge\": {\"precision\": 0.3888888888888889, \"recall\": 0.9032258064516129, \"f1-score\": 0.5436893203883495, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.5217391304347826, \"recall\": 0.32432432432432434, \"f1-score\": 0.4, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.1111111111111111, \"f1-score\": 0.19999999999999998, \"support\": 9.0}, \"request\": {\"precision\": 0.25, \"recall\": 0.125, \"f1-score\": 0.16666666666666666, \"support\": 8.0}, \"accuracy\": 0.43103448275862066, \"macro avg\": {\"precision\": 0.3367451690821256, \"recall\": 0.2940687663469922, \"f1-score\": 0.24712783171521036, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.4065508912210561, \"recall\": 0.43103448275862066, \"f1-score\": 0.3516181229773463, \"support\": 116.0}}", "eval_f1_macro": 21.966918374685367, "eval_f1_micro": 43.103448275862064, "eval_f1_weighted": 35.161812297734635, "eval_loss": 0.5922260880470276, "eval_runtime": 2.1966, "eval_samples_per_second": 52.809, "step": 240 }, { "epoch": 16.0, "learning_rate": 0.00023999999999999998, "loss": 0.613, "step": 256 }, { "epoch": 16.0, "eval_accuracy": 46.55172413793103, "eval_average_metrics": 37.72845991532228, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"challenge\": {\"precision\": 0.5238095238095238, \"recall\": 0.7096774193548387, \"f1-score\": 0.6027397260273972, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.4, \"recall\": 0.6486486486486487, \"f1-score\": 0.49484536082474234, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"accuracy\": 0.46551724137931033, \"macro avg\": {\"precision\": 0.19239926739926738, \"recall\": 0.28090186961154706, \"f1-score\": 0.22810722676560835, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.31531514462548943, \"recall\": 0.46551724137931033, \"f1-score\": 0.37534193450706316, \"support\": 116.0}}", "eval_f1_macro": 20.27619793472074, "eval_f1_micro": 46.55172413793103, "eval_f1_weighted": 37.53419345070632, "eval_loss": 0.5560970306396484, "eval_runtime": 2.0931, "eval_samples_per_second": 55.419, "step": 256 }, { "epoch": 17.0, "learning_rate": 0.00025499999999999996, "loss": 0.544, "step": 272 }, { "epoch": 17.0, "eval_accuracy": 52.58620689655172, "eval_average_metrics": 45.3769094880828, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"challenge\": {\"precision\": 0.5217391304347826, \"recall\": 0.7741935483870968, \"f1-score\": 0.6233766233766235, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.4897959183673469, \"recall\": 0.6486486486486487, \"f1-score\": 0.5581395348837208, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5, \"recall\": 0.5555555555555556, \"f1-score\": 0.5263157894736842, \"support\": 9.0}, \"request\": {\"precision\": 0.3333333333333333, \"recall\": 0.125, \"f1-score\": 0.18181818181818182, \"support\": 8.0}, \"accuracy\": 0.5258620689655172, \"macro avg\": {\"precision\": 0.33998354776693285, \"recall\": 0.3601469412961349, \"f1-score\": 0.3391474426646145, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.42532783353221343, \"recall\": 0.5258620689655172, \"f1-score\": 0.46188784811262035, \"support\": 116.0}}", "eval_f1_macro": 30.14643934796573, "eval_f1_micro": 52.58620689655172, "eval_f1_weighted": 46.18878481126203, "eval_loss": 0.5054384469985962, "eval_runtime": 2.1595, "eval_samples_per_second": 53.717, "step": 272 }, { "epoch": 18.0, "learning_rate": 0.00027, "loss": 0.5119, "step": 288 }, { "epoch": 18.0, "eval_accuracy": 52.58620689655172, "eval_average_metrics": 46.69862925181969, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.5333333333333333, \"recall\": 0.8888888888888888, \"f1-score\": 0.6666666666666667, \"support\": 9.0}, \"challenge\": {\"precision\": 0.5609756097560976, \"recall\": 0.7419354838709677, \"f1-score\": 0.638888888888889, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.5277777777777778, \"recall\": 0.5135135135135135, \"f1-score\": 0.5205479452054794, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.5555555555555556, \"f1-score\": 0.7142857142857143, \"support\": 9.0}, \"request\": {\"precision\": 0.375, \"recall\": 0.75, \"f1-score\": 0.5, \"support\": 8.0}, \"accuracy\": 0.5258620689655172, \"macro avg\": {\"precision\": 0.37463584010840106, \"recall\": 0.4312366802286157, \"f1-score\": 0.3800486518808437, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.46308639379497246, \"recall\": 0.5258620689655172, \"f1-score\": 0.47840000824767004, \"support\": 116.0}}", "eval_f1_macro": 33.78210238940833, "eval_f1_micro": 52.58620689655172, "eval_f1_weighted": 47.840000824767, "eval_loss": 0.5165005922317505, "eval_runtime": 2.2159, "eval_samples_per_second": 52.349, "step": 288 }, { "epoch": 19.0, "learning_rate": 0.000285, "loss": 0.4492, "step": 304 }, { "epoch": 19.0, "eval_accuracy": 60.3448275862069, "eval_average_metrics": 55.316034256647285, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5384615384615384, \"recall\": 0.5833333333333334, \"f1-score\": 0.5599999999999999, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.5, \"recall\": 0.967741935483871, \"f1-score\": 0.6593406593406593, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.35135135135135137, \"f1-score\": 0.48148148148148157, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.603448275862069, \"macro avg\": {\"precision\": 0.5309514831573654, \"recall\": 0.5242616608544028, \"f1-score\": 0.5012169312169312, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.6142728974879077, \"recall\": 0.603448275862069, \"f1-score\": 0.5602186574600367, \"support\": 116.0}}", "eval_f1_macro": 44.55261610817167, "eval_f1_micro": 60.3448275862069, "eval_f1_weighted": 56.02186574600368, "eval_loss": 0.4782406687736511, "eval_runtime": 2.2036, "eval_samples_per_second": 52.641, "step": 304 }, { "epoch": 20.0, "learning_rate": 0.0003, "loss": 0.3965, "step": 320 }, { "epoch": 20.0, "eval_accuracy": 66.37931034482759, "eval_average_metrics": 61.760117229791256, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.5, \"recall\": 0.2222222222222222, \"f1-score\": 0.30769230769230765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.75, \"recall\": 0.5, \"f1-score\": 0.6, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6486486486486487, \"recall\": 0.7741935483870968, \"f1-score\": 0.7058823529411764, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7058823529411765, \"recall\": 0.6486486486486487, \"f1-score\": 0.676056338028169, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5333333333333333, \"recall\": 0.8888888888888888, \"f1-score\": 0.6666666666666667, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.6637931034482759, \"macro avg\": {\"precision\": 0.5797330418653948, \"recall\": 0.5684802746294682, \"f1-score\": 0.5540610176898495, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.665451912358606, \"recall\": 0.6637931034482759, \"f1-score\": 0.6503197999041211, \"support\": 116.0}}", "eval_f1_macro": 49.24986823909773, "eval_f1_micro": 66.37931034482759, "eval_f1_weighted": 65.03197999041211, "eval_loss": 0.39811432361602783, "eval_runtime": 2.0694, "eval_samples_per_second": 56.054, "step": 320 }, { "epoch": 21.0, "learning_rate": 0.00029833333333333334, "loss": 0.3235, "step": 336 }, { "epoch": 21.0, "eval_accuracy": 68.10344827586206, "eval_average_metrics": 64.5612969778262, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.4444444444444444, \"f1-score\": 0.5333333333333333, \"support\": 9.0}, \"accusation\": {\"precision\": 0.7272727272727273, \"recall\": 0.6666666666666666, \"f1-score\": 0.6956521739130435, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.5490196078431373, \"recall\": 0.9032258064516129, \"f1-score\": 0.6829268292682927, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.85, \"recall\": 0.4594594594594595, \"f1-score\": 0.5964912280701754, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.6810344827586207, \"macro avg\": {\"precision\": 0.650707176810118, \"recall\": 0.6251967693499951, \"f1-score\": 0.6198996519223119, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7246036502375245, \"recall\": 0.6810344827586207, \"f1-score\": 0.6693610007759736, \"support\": 116.0}}", "eval_f1_macro": 55.10219128198328, "eval_f1_micro": 68.10344827586206, "eval_f1_weighted": 66.93610007759737, "eval_loss": 0.4031699597835541, "eval_runtime": 2.1347, "eval_samples_per_second": 54.34, "step": 336 }, { "epoch": 22.0, "learning_rate": 0.00029666666666666665, "loss": 0.2945, "step": 352 }, { "epoch": 22.0, "eval_accuracy": 69.82758620689656, "eval_average_metrics": 66.08601331128551, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.5, \"recall\": 0.5555555555555556, \"f1-score\": 0.5263157894736842, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8333333333333334, \"f1-score\": 0.7407407407407408, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.7096774193548387, \"f1-score\": 0.6875, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7666666666666667, \"recall\": 0.6216216216216216, \"f1-score\": 0.6865671641791045, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.6242424242424243, \"recall\": 0.6403707134553909, \"f1-score\": 0.6215324031911326, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7075757575757575, \"recall\": 0.6982758620689655, \"f1-score\": 0.6944155610324821, \"support\": 116.0}}", "eval_f1_macro": 55.24732472810068, "eval_f1_micro": 69.82758620689656, "eval_f1_weighted": 69.44155610324822, "eval_loss": 0.3617841899394989, "eval_runtime": 2.0239, "eval_samples_per_second": 57.314, "step": 352 }, { "epoch": 23.0, "learning_rate": 0.00029499999999999996, "loss": 0.3049, "step": 368 }, { "epoch": 23.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 71.50408533829969, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.7142857142857143, \"recall\": 0.5555555555555556, \"f1-score\": 0.6250000000000001, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8333333333333334, \"f1-score\": 0.7407407407407408, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.7741935483870968, \"f1-score\": 0.7164179104477612, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7692307692307693, \"recall\": 0.5405405405405406, \"f1-score\": 0.6349206349206349, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7720872877122877, \"recall\": 0.8084389833381769, \"f1-score\": 0.7817480060017373, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7320725251759735, \"recall\": 0.7241379310344828, \"f1-score\": 0.717000435017033, \"support\": 116.0}}", "eval_f1_macro": 69.4887116445989, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 71.7000435017033, "eval_loss": 0.3993362486362457, "eval_runtime": 2.1528, "eval_samples_per_second": 53.884, "step": 368 }, { "epoch": 24.0, "learning_rate": 0.00029333333333333327, "loss": 0.2268, "step": 384 }, { "epoch": 24.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 72.61508757752452, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.7142857142857143, \"recall\": 0.5555555555555556, \"f1-score\": 0.6250000000000001, \"support\": 9.0}, \"accusation\": {\"precision\": 0.9, \"recall\": 0.75, \"f1-score\": 0.8181818181818182, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6, \"recall\": 0.8709677419354839, \"f1-score\": 0.7105263157894737, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.8, \"recall\": 0.5405405405405406, \"f1-score\": 0.6451612903225806, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.875, \"f1-score\": 0.9333333333333333, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.8212301587301587, \"recall\": 0.7962302019761698, \"f1-score\": 0.7978745510526071, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7623152709359605, \"recall\": 0.7327586206896551, \"f1-score\": 0.7298644385637975, \"support\": 116.0}}", "eval_f1_macro": 70.9221823157873, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 72.98644385637975, "eval_loss": 0.3922426402568817, "eval_runtime": 1.9795, "eval_samples_per_second": 58.601, "step": 384 }, { "epoch": 25.0, "learning_rate": 0.00029166666666666664, "loss": 0.2401, "step": 400 }, { "epoch": 25.0, "eval_accuracy": 75.86206896551724, "eval_average_metrics": 74.81572695778885, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.7142857142857143, \"recall\": 0.5555555555555556, \"f1-score\": 0.6250000000000001, \"support\": 9.0}, \"accusation\": {\"precision\": 0.9, \"recall\": 0.75, \"f1-score\": 0.8181818181818182, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6428571428571429, \"recall\": 0.8709677419354839, \"f1-score\": 0.7397260273972603, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.8214285714285714, \"recall\": 0.6216216216216216, \"f1-score\": 0.7076923076923075, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.875, \"f1-score\": 0.9333333333333333, \"support\": 8.0}, \"accuracy\": 0.7586206896551724, \"macro avg\": {\"precision\": 0.8257305194805196, \"recall\": 0.8063653371113049, \"f1-score\": 0.8082548437203267, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7784090909090909, \"recall\": 0.7586206896551724, \"f1-score\": 0.7569389490275861, \"support\": 116.0}}", "eval_f1_macro": 71.84487499736237, "eval_f1_micro": 75.86206896551724, "eval_f1_weighted": 75.69389490275861, "eval_loss": 0.38488852977752686, "eval_runtime": 2.103, "eval_samples_per_second": 55.158, "step": 400 }, { "epoch": 26.0, "learning_rate": 0.00029, "loss": 0.1774, "step": 416 }, { "epoch": 26.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 69.79597725103807, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.625, \"recall\": 0.5555555555555556, \"f1-score\": 0.5882352941176471, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8333333333333334, \"f1-score\": 0.7407407407407408, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.5333333333333333, \"recall\": 0.8888888888888888, \"f1-score\": 0.6666666666666667, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.7419354838709677, \"f1-score\": 0.7419354838709677, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7307692307692307, \"recall\": 0.5135135135135135, \"f1-score\": 0.6031746031746031, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6923076923076923, \"recall\": 1.0, \"f1-score\": 0.8181818181818181, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.875, \"f1-score\": 0.9333333333333333, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7487515508684863, \"recall\": 0.8010283468952824, \"f1-score\": 0.7615334925107221, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7215019893899203, \"recall\": 0.7068965517241379, \"f1-score\": 0.701127326583716, \"support\": 116.0}}", "eval_f1_macro": 67.69186600095307, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.1127326583716, "eval_loss": 0.4694823622703552, "eval_runtime": 2.2038, "eval_samples_per_second": 52.636, "step": 416 }, { "epoch": 27.0, "learning_rate": 0.0002883333333333333, "loss": 0.1601, "step": 432 }, { "epoch": 27.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 71.6286036100653, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.7142857142857143, \"recall\": 0.5555555555555556, \"f1-score\": 0.6250000000000001, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.7741935483870968, \"f1-score\": 0.7164179104477612, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7777777777777778, \"recall\": 0.5675675675675675, \"f1-score\": 0.65625, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.875, \"f1-score\": 0.9333333333333333, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7842797725150666, \"recall\": 0.7979284728276663, \"f1-score\": 0.7828452099190978, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7385973616196739, \"recall\": 0.7241379310344828, \"f1-score\": 0.7210058735166711, \"support\": 116.0}}", "eval_f1_macro": 69.58624088169758, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.10058735166709, "eval_loss": 0.4516548216342926, "eval_runtime": 2.1426, "eval_samples_per_second": 54.14, "step": 432 }, { "epoch": 28.0, "learning_rate": 0.0002866666666666667, "loss": 0.1426, "step": 448 }, { "epoch": 28.0, "eval_accuracy": 75.86206896551724, "eval_average_metrics": 75.02176780786866, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.7741935483870968, \"f1-score\": 0.7164179104477612, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.8, \"recall\": 0.6486486486486487, \"f1-score\": 0.7164179104477612, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.875, \"f1-score\": 0.9333333333333333, \"support\": 8.0}, \"accuracy\": 0.7586206896551724, \"macro avg\": {\"precision\": 0.8166666666666667, \"recall\": 0.8254247190739126, \"f1-score\": 0.8164155862659283, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7682471264367817, \"recall\": 0.7586206896551724, \"f1-score\": 0.75792658965691, \"support\": 116.0}}", "eval_f1_macro": 72.57027433474917, "eval_f1_micro": 75.86206896551724, "eval_f1_weighted": 75.79265896569099, "eval_loss": 0.4462680220603943, "eval_runtime": 2.0575, "eval_samples_per_second": 56.379, "step": 448 }, { "epoch": 29.0, "learning_rate": 0.000285, "loss": 0.1153, "step": 464 }, { "epoch": 29.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 69.98010595460308, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.7142857142857143, \"recall\": 0.5555555555555556, \"f1-score\": 0.6250000000000001, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6428571428571429, \"recall\": 0.75, \"f1-score\": 0.6923076923076924, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.5, \"recall\": 0.8888888888888888, \"f1-score\": 0.64, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6410256410256411, \"recall\": 0.8064516129032258, \"f1-score\": 0.7142857142857142, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.8260869565217391, \"recall\": 0.5135135135135135, \"f1-score\": 0.6333333333333333, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.875, \"f1-score\": 0.9333333333333333, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7766430429473907, \"recall\": 0.784787307468759, \"f1-score\": 0.7658936202686203, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7420678397065203, \"recall\": 0.7068965517241379, \"f1-score\": 0.704616805608185, \"support\": 116.0}}", "eval_f1_macro": 68.07943291276625, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.4616805608185, "eval_loss": 0.5256942510604858, "eval_runtime": 2.1565, "eval_samples_per_second": 53.791, "step": 464 }, { "epoch": 30.0, "learning_rate": 0.0002833333333333333, "loss": 0.0824, "step": 480 }, { "epoch": 30.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 70.1937005308774, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.6451612903225806, \"f1-score\": 0.6557377049180327, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7777777777777778, \"recall\": 0.5675675675675675, \"f1-score\": 0.65625, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.875, \"f1-score\": 0.9333333333333333, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7606060606060606, \"recall\": 0.8095772183473796, \"f1-score\": 0.775683926773902, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7230581678857542, \"recall\": 0.7068965517241379, \"f1-score\": 0.7044580939877959, \"support\": 116.0}}", "eval_f1_macro": 68.9496823799024, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.4458093987796, "eval_loss": 0.5310544371604919, "eval_runtime": 2.0193, "eval_samples_per_second": 57.447, "step": 480 }, { "epoch": 31.0, "learning_rate": 0.00028166666666666666, "loss": 0.0898, "step": 496 }, { "epoch": 31.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 71.05355939177225, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.5714285714285714, \"recall\": 0.8888888888888888, \"f1-score\": 0.6956521739130435, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.5675675675675675, \"f1-score\": 0.6461538461538462, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.875, \"f1-score\": 0.9333333333333333, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7776785714285714, \"recall\": 0.8072250678097452, \"f1-score\": 0.7837074960829841, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7310806650246304, \"recall\": 0.7155172413793104, \"f1-score\": 0.7144790075051725, \"support\": 116.0}}", "eval_f1_macro": 69.6628885407097, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.44790075051725, "eval_loss": 0.5583775043487549, "eval_runtime": 2.0569, "eval_samples_per_second": 56.394, "step": 496 }, { "epoch": 32.0, "learning_rate": 0.00028, "loss": 0.0709, "step": 512 }, { "epoch": 32.0, "eval_accuracy": 78.44827586206897, "eval_average_metrics": 77.31764307402186, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.9, \"recall\": 0.75, \"f1-score\": 0.8181818181818182, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6923076923076923, \"recall\": 0.8709677419354839, \"f1-score\": 0.7714285714285716, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.8275862068965517, \"recall\": 0.6486486486486487, \"f1-score\": 0.7272727272727273, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.875, \"f1-score\": 0.9333333333333333, \"support\": 8.0}, \"accuracy\": 0.7844827586206896, \"macro avg\": {\"precision\": 0.8420700707338638, \"recall\": 0.8375214932674611, \"f1-score\": 0.8331714982643775, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7994157596268179, \"recall\": 0.7844827586206896, \"f1-score\": 0.7831433183733814, \"support\": 116.0}}", "eval_f1_macro": 74.05968873461136, "eval_f1_micro": 78.44827586206897, "eval_f1_weighted": 78.31433183733813, "eval_loss": 0.5467008948326111, "eval_runtime": 1.9594, "eval_samples_per_second": 59.203, "step": 512 }, { "epoch": 33.0, "learning_rate": 0.00027833333333333334, "loss": 0.084, "step": 528 }, { "epoch": 33.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 72.73042654857636, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6428571428571429, \"recall\": 0.75, \"f1-score\": 0.6923076923076924, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6216216216216216, \"recall\": 0.7419354838709677, \"f1-score\": 0.676470588235294, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.8148148148148148, \"recall\": 0.5945945945945946, \"f1-score\": 0.6875, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.875, \"f1-score\": 0.9333333333333333, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7994950307450308, \"recall\": 0.8146357042526398, \"f1-score\": 0.8005958937218612, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7498508976957252, \"recall\": 0.7327586206896551, \"f1-score\": 0.7320590261443122, \"support\": 116.0}}", "eval_f1_macro": 71.16407944194322, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.20590261443122, "eval_loss": 0.5764071345329285, "eval_runtime": 2.116, "eval_samples_per_second": 54.82, "step": 528 }, { "epoch": 34.0, "learning_rate": 0.00027666666666666665, "loss": 0.0838, "step": 544 }, { "epoch": 34.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 70.64634709659882, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6923076923076923, \"recall\": 0.75, \"f1-score\": 0.7199999999999999, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6486486486486487, \"recall\": 0.7741935483870968, \"f1-score\": 0.7058823529411764, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7777777777777778, \"recall\": 0.5675675675675675, \"f1-score\": 0.65625, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7754478254478254, \"recall\": 0.7840395839387775, \"f1-score\": 0.7681058539306411, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7323897806656428, \"recall\": 0.7155172413793104, \"f1-score\": 0.7120586420558731, \"support\": 116.0}}", "eval_f1_macro": 68.27607590494587, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.2058642055873, "eval_loss": 0.6548714637756348, "eval_runtime": 2.138, "eval_samples_per_second": 54.256, "step": 544 }, { "epoch": 35.0, "learning_rate": 0.00027499999999999996, "loss": 0.0924, "step": 560 }, { "epoch": 35.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 72.46959003253342, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6428571428571429, \"recall\": 0.75, \"f1-score\": 0.6923076923076924, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6451612903225806, \"recall\": 0.6451612903225806, \"f1-score\": 0.6451612903225806, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.8333333333333334, \"recall\": 0.6756756756756757, \"f1-score\": 0.746268656716418, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 9.0}, \"request\": {\"precision\": 0.7777777777777778, \"recall\": 0.875, \"f1-score\": 0.823529411764706, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7707245263696876, \"recall\": 0.8126740651942264, \"f1-score\": 0.7875938450480724, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7428434592227696, \"recall\": 0.7327586206896551, \"f1-score\": 0.7331829421015178, \"support\": 116.0}}", "eval_f1_macro": 70.00834178205089, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.31829421015178, "eval_loss": 0.5695762038230896, "eval_runtime": 2.0764, "eval_samples_per_second": 55.866, "step": 560 }, { "epoch": 36.0, "learning_rate": 0.00027333333333333333, "loss": 0.0452, "step": 576 }, { "epoch": 36.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 70.22480821008993, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.7142857142857143, \"recall\": 0.5555555555555556, \"f1-score\": 0.6250000000000001, \"support\": 9.0}, \"accusation\": {\"precision\": 0.43478260869565216, \"recall\": 0.8333333333333334, \"f1-score\": 0.5714285714285714, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7037037037037037, \"recall\": 0.6129032258064516, \"f1-score\": 0.6551724137931035, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.8, \"recall\": 0.6486486486486487, \"f1-score\": 0.7164179104477612, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.875, \"f1-score\": 0.9333333333333333, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7736419578810884, \"recall\": 0.7879023176402209, \"f1-score\": 0.7685900812569252, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7404142265086794, \"recall\": 0.7068965517241379, \"f1-score\": 0.7120080416158326, \"support\": 116.0}}", "eval_f1_macro": 68.3191183339489, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 71.20080416158326, "eval_loss": 0.6615910530090332, "eval_runtime": 2.0721, "eval_samples_per_second": 55.982, "step": 576 }, { "epoch": 37.0, "learning_rate": 0.00027166666666666664, "loss": 0.0662, "step": 592 }, { "epoch": 37.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 70.7853828197613, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.7142857142857143, \"recall\": 0.5555555555555556, \"f1-score\": 0.6250000000000001, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5, \"recall\": 0.8333333333333334, \"f1-score\": 0.625, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6774193548387096, \"recall\": 0.6774193548387096, \"f1-score\": 0.6774193548387096, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.8214285714285714, \"recall\": 0.6216216216216216, \"f1-score\": 0.7076923076923075, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.774747765675185, \"recall\": 0.7908523442797636, \"f1-score\": 0.7692699101973295, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7429746977160769, \"recall\": 0.7155172413793104, \"f1-score\": 0.7165853543008714, \"support\": 116.0}}", "eval_f1_macro": 68.37954757309596, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.65853543008714, "eval_loss": 0.686450719833374, "eval_runtime": 1.9539, "eval_samples_per_second": 59.368, "step": 592 }, { "epoch": 38.0, "learning_rate": 0.00027, "loss": 0.0408, "step": 608 }, { "epoch": 38.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 71.66157684555884, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6571428571428571, \"recall\": 0.7419354838709677, \"f1-score\": 0.6969696969696969, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.8148148148148148, \"recall\": 0.5945945945945946, \"f1-score\": 0.6875, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7836632696007696, \"recall\": 0.7990107042526398, \"f1-score\": 0.7815321014769544, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7446828714070093, \"recall\": 0.7241379310344828, \"f1-score\": 0.7234920104405398, \"support\": 116.0}}", "eval_f1_macro": 69.46952013128484, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.34920104405398, "eval_loss": 0.7136876583099365, "eval_runtime": 2.1177, "eval_samples_per_second": 54.777, "step": 608 }, { "epoch": 39.0, "learning_rate": 0.0002683333333333333, "loss": 0.0315, "step": 624 }, { "epoch": 39.0, "eval_accuracy": 68.10344827586206, "eval_average_metrics": 67.42979960880837, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.4444444444444444, \"f1-score\": 0.5333333333333333, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6052631578947368, \"recall\": 0.7419354838709677, \"f1-score\": 0.6666666666666666, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7777777777777778, \"recall\": 0.5675675675675675, \"f1-score\": 0.65625, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"accuracy\": 0.6810344827586207, \"macro avg\": {\"precision\": 0.7563779239766082, \"recall\": 0.7383406592075947, \"f1-score\": 0.7353737797113752, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7091096995361968, \"recall\": 0.6810344827586207, \"f1-score\": 0.6814574368694261, \"support\": 116.0}}", "eval_f1_macro": 65.36655819656669, "eval_f1_micro": 68.10344827586206, "eval_f1_weighted": 68.14574368694261, "eval_loss": 0.8272783756256104, "eval_runtime": 2.1248, "eval_samples_per_second": 54.594, "step": 624 }, { "epoch": 40.0, "learning_rate": 0.0002666666666666666, "loss": 0.0537, "step": 640 }, { "epoch": 40.0, "eval_accuracy": 69.82758620689656, "eval_average_metrics": 69.24541954934139, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6363636363636364, \"recall\": 0.6774193548387096, \"f1-score\": 0.65625, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.8076923076923077, \"recall\": 0.5675675675675675, \"f1-score\": 0.6666666666666666, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7550152972027973, \"recall\": 0.787567809745229, \"f1-score\": 0.7602210426634769, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7212412587412587, \"recall\": 0.6982758620689655, \"f1-score\": 0.6975130199126343, \"support\": 116.0}}", "eval_f1_macro": 67.57520379230904, "eval_f1_micro": 69.82758620689656, "eval_f1_weighted": 69.75130199126343, "eval_loss": 0.724757969379425, "eval_runtime": 2.3811, "eval_samples_per_second": 48.716, "step": 640 }, { "epoch": 41.0, "learning_rate": 0.000265, "loss": 0.0215, "step": 656 }, { "epoch": 41.0, "eval_accuracy": 69.82758620689656, "eval_average_metrics": 69.33402692747292, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6388888888888888, \"recall\": 0.7419354838709677, \"f1-score\": 0.6865671641791045, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7407407407407407, \"recall\": 0.5405405405405406, \"f1-score\": 0.625, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7615477693602694, \"recall\": 0.7801011697181053, \"f1-score\": 0.7655244436082891, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7106916869847903, \"recall\": 0.6982758620689655, \"f1-score\": 0.6963431808647288, \"support\": 116.0}}", "eval_f1_macro": 68.0466172096257, "eval_f1_micro": 69.82758620689656, "eval_f1_weighted": 69.63431808647287, "eval_loss": 0.7391438484191895, "eval_runtime": 2.157, "eval_samples_per_second": 53.778, "step": 656 }, { "epoch": 42.0, "learning_rate": 0.0002633333333333333, "loss": 0.0267, "step": 672 }, { "epoch": 42.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 71.60599983270731, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6923076923076923, \"recall\": 0.75, \"f1-score\": 0.7199999999999999, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6388888888888888, \"recall\": 0.7419354838709677, \"f1-score\": 0.6865671641791045, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7857142857142857, \"recall\": 0.5945945945945946, \"f1-score\": 0.676923076923077, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7761127761127761, \"recall\": 0.7990107042526398, \"f1-score\": 0.781409669493515, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7340987556504798, \"recall\": 0.7241379310344828, \"f1-score\": 0.7213777583562028, \"support\": 116.0}}", "eval_f1_macro": 69.45863728831245, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.13777583562027, "eval_loss": 0.730360209941864, "eval_runtime": 2.2556, "eval_samples_per_second": 51.428, "step": 672 }, { "epoch": 43.0, "learning_rate": 0.00026166666666666667, "loss": 0.0238, "step": 688 }, { "epoch": 43.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 71.03986860621411, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.43478260869565216, \"recall\": 0.8333333333333334, \"f1-score\": 0.5714285714285714, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.5806451612903226, \"f1-score\": 0.6545454545454547, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7878787878787878, \"recall\": 0.7027027027027027, \"f1-score\": 0.7428571428571429, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.875, \"f1-score\": 0.9333333333333333, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7809892402283706, \"recall\": 0.7906268163324615, \"f1-score\": 0.7746169679993209, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7508291308891009, \"recall\": 0.7155172413793104, \"f1-score\": 0.7220118454905474, \"support\": 116.0}}", "eval_f1_macro": 68.85484159993965, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 72.20118454905474, "eval_loss": 0.7982239127159119, "eval_runtime": 2.2527, "eval_samples_per_second": 51.494, "step": 688 }, { "epoch": 44.0, "learning_rate": 0.00026, "loss": 0.0209, "step": 704 }, { "epoch": 44.0, "eval_accuracy": 67.24137931034483, "eval_average_metrics": 66.37757277211017, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.625, \"recall\": 0.8333333333333334, \"f1-score\": 0.7142857142857143, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6451612903225806, \"recall\": 0.6451612903225806, \"f1-score\": 0.6451612903225806, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7333333333333333, \"recall\": 0.5945945945945946, \"f1-score\": 0.6567164179104478, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.6724137931034483, \"macro avg\": {\"precision\": 0.7129368279569892, \"recall\": 0.7400389300590913, \"f1-score\": 0.7194389741807576, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.6818965517241379, \"recall\": 0.6724137931034483, \"f1-score\": 0.6707740142946146, \"support\": 116.0}}", "eval_f1_macro": 63.95013103828957, "eval_f1_micro": 67.24137931034483, "eval_f1_weighted": 67.07740142946146, "eval_loss": 0.788184404373169, "eval_runtime": 2.2001, "eval_samples_per_second": 52.725, "step": 704 }, { "epoch": 45.0, "learning_rate": 0.00025833333333333334, "loss": 0.0349, "step": 720 }, { "epoch": 45.0, "eval_accuracy": 69.82758620689656, "eval_average_metrics": 68.73739006091948, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6428571428571429, \"recall\": 0.75, \"f1-score\": 0.6923076923076924, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6571428571428571, \"recall\": 0.7419354838709677, \"f1-score\": 0.6969696969696969, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7586206896551724, \"recall\": 0.5945945945945946, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 8.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7613339964633068, \"recall\": 0.7538718153637509, \"f1-score\": 0.7397088205911735, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7193374253570448, \"recall\": 0.6982758620689655, \"f1-score\": 0.6954249266622493, \"support\": 116.0}}", "eval_f1_macro": 65.75189516365988, "eval_f1_micro": 69.82758620689656, "eval_f1_weighted": 69.54249266622493, "eval_loss": 0.9080767035484314, "eval_runtime": 2.3303, "eval_samples_per_second": 49.778, "step": 720 }, { "epoch": 46.0, "learning_rate": 0.00025666666666666665, "loss": 0.0085, "step": 736 }, { "epoch": 46.0, "eval_accuracy": 68.96551724137932, "eval_average_metrics": 68.39529379047518, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.7142857142857143, \"recall\": 0.5555555555555556, \"f1-score\": 0.6250000000000001, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6470588235294118, \"recall\": 0.7096774193548387, \"f1-score\": 0.6769230769230768, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7333333333333333, \"recall\": 0.5945945945945946, \"f1-score\": 0.6567164179104478, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"accuracy\": 0.6896551724137931, \"macro avg\": {\"precision\": 0.761999207577768, \"recall\": 0.7515756684103458, \"f1-score\": 0.7481724665655658, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7097975669396616, \"recall\": 0.6896551724137931, \"f1-score\": 0.6914592142886957, \"support\": 116.0}}", "eval_f1_macro": 66.50421925027251, "eval_f1_micro": 68.96551724137932, "eval_f1_weighted": 69.14592142886957, "eval_loss": 0.9103646278381348, "eval_runtime": 2.3131, "eval_samples_per_second": 50.149, "step": 736 }, { "epoch": 47.0, "learning_rate": 0.00025499999999999996, "loss": 0.0252, "step": 752 }, { "epoch": 47.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 69.88211421894832, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6428571428571429, \"recall\": 0.75, \"f1-score\": 0.6923076923076924, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6486486486486487, \"recall\": 0.7741935483870968, \"f1-score\": 0.7058823529411764, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.5675675675675675, \"f1-score\": 0.6461538461538462, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7606265356265356, \"recall\": 0.7718868061609997, \"f1-score\": 0.7615414384377233, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7164380121276674, \"recall\": 0.7068965517241379, \"f1-score\": 0.7045657422539028, \"support\": 116.0}}", "eval_f1_macro": 67.69257230557541, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.45657422539028, "eval_loss": 0.8836216330528259, "eval_runtime": 2.3628, "eval_samples_per_second": 49.094, "step": 752 }, { "epoch": 48.0, "learning_rate": 0.00025333333333333333, "loss": 0.0068, "step": 768 }, { "epoch": 48.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 70.60514862150394, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6774193548387096, \"recall\": 0.6774193548387096, \"f1-score\": 0.6774193548387096, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.7027027027027027, \"f1-score\": 0.7222222222222223, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7785762288786482, \"recall\": 0.7610569238593432, \"f1-score\": 0.7609549816618129, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7289716748768472, \"recall\": 0.7155172413793104, \"f1-score\": 0.7167670339577035, \"support\": 116.0}}", "eval_f1_macro": 67.64044281438335, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.67670339577035, "eval_loss": 0.7966245412826538, "eval_runtime": 2.368, "eval_samples_per_second": 48.987, "step": 768 }, { "epoch": 49.0, "learning_rate": 0.00025166666666666664, "loss": 0.0115, "step": 784 }, { "epoch": 49.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 71.30009730584443, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5, \"recall\": 0.9166666666666666, \"f1-score\": 0.6470588235294118, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7586206896551724, \"recall\": 0.7096774193548387, \"f1-score\": 0.7333333333333333, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.8461538461538461, \"recall\": 0.5945945945945946, \"f1-score\": 0.6984126984126985, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 1.0, \"f1-score\": 0.8571428571428571, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7630968169761274, \"recall\": 0.800186779521457, \"f1-score\": 0.7647525676937441, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7585482789109425, \"recall\": 0.7241379310344828, \"f1-score\": 0.7239479699925947, \"support\": 116.0}}", "eval_f1_macro": 67.9780060172217, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.39479699925948, "eval_loss": 0.9425663352012634, "eval_runtime": 2.4101, "eval_samples_per_second": 48.13, "step": 784 }, { "epoch": 50.0, "learning_rate": 0.00025, "loss": 0.0116, "step": 800 }, { "epoch": 50.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 70.36753700765975, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7741935483870968, \"recall\": 0.6486486486486487, \"f1-score\": 0.7058823529411764, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"request\": {\"precision\": 0.7142857142857143, \"recall\": 0.625, \"f1-score\": 0.6666666666666666, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.74224677652097, \"recall\": 0.778605722658142, \"f1-score\": 0.7523584009875137, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7290652432476681, \"recall\": 0.7155172413793104, \"f1-score\": 0.7149039744477569, \"support\": 116.0}}", "eval_f1_macro": 66.87630231000122, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.49039744477568, "eval_loss": 0.9411255121231079, "eval_runtime": 2.3319, "eval_samples_per_second": 49.744, "step": 800 }, { "epoch": 51.0, "learning_rate": 0.0002483333333333333, "loss": 0.0087, "step": 816 }, { "epoch": 51.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 69.62433585996804, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6571428571428571, \"recall\": 0.7419354838709677, \"f1-score\": 0.6969696969696969, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7931034482758621, \"recall\": 0.6216216216216216, \"f1-score\": 0.696969696969697, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7493363437328955, \"recall\": 0.7589863048532404, \"f1-score\": 0.7476551226551227, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7205424381405356, \"recall\": 0.7068965517241379, \"f1-score\": 0.706597999701448, \"support\": 116.0}}", "eval_f1_macro": 66.45823312489979, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.6597999701448, "eval_loss": 0.9394950866699219, "eval_runtime": 2.3963, "eval_samples_per_second": 48.408, "step": 816 }, { "epoch": 52.0, "learning_rate": 0.0002466666666666666, "loss": 0.0126, "step": 832 }, { "epoch": 52.0, "eval_accuracy": 69.82758620689656, "eval_average_metrics": 69.02533360824437, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6428571428571429, \"recall\": 0.75, \"f1-score\": 0.6923076923076924, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6216216216216216, \"recall\": 0.7419354838709677, \"f1-score\": 0.676470588235294, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7419354838709677, \"recall\": 0.6216216216216216, \"f1-score\": 0.676470588235294, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7626389022558377, \"recall\": 0.7450974159643514, \"f1-score\": 0.7486182934712347, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7108198989803551, \"recall\": 0.6982758620689655, \"f1-score\": 0.6990231371063014, \"support\": 116.0}}", "eval_f1_macro": 66.54384830855419, "eval_f1_micro": 69.82758620689656, "eval_f1_weighted": 69.90231371063015, "eval_loss": 0.9984409809112549, "eval_runtime": 2.2343, "eval_samples_per_second": 51.918, "step": 832 }, { "epoch": 53.0, "learning_rate": 0.000245, "loss": 0.0192, "step": 848 }, { "epoch": 53.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 69.02199924752503, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5238095238095238, \"recall\": 0.9166666666666666, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.6774193548387096, \"f1-score\": 0.7, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7857142857142857, \"recall\": 0.5945945945945946, \"f1-score\": 0.676923076923077, \"support\": 37.0}, \"question\": {\"precision\": 0.5, \"recall\": 1.0, \"f1-score\": 0.6666666666666666, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.6971460292580982, \"recall\": 0.7840017436791631, \"f1-score\": 0.7197805032944352, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7313778819278225, \"recall\": 0.7068965517241379, \"f1-score\": 0.7072819746354496, \"support\": 116.0}}", "eval_f1_macro": 63.98048918172757, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.72819746354496, "eval_loss": 1.050161600112915, "eval_runtime": 2.3204, "eval_samples_per_second": 49.991, "step": 848 }, { "epoch": 54.0, "learning_rate": 0.0002433333333333333, "loss": 0.0124, "step": 864 }, { "epoch": 54.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 72.27556400444958, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6923076923076923, \"recall\": 0.75, \"f1-score\": 0.7199999999999999, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.7096774193548387, \"f1-score\": 0.7096774193548389, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.7027027027027027, \"f1-score\": 0.7123287671232876, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7841271072117846, \"recall\": 0.7807141819238593, \"f1-score\": 0.7791653391360963, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7371729085091153, \"recall\": 0.7327586206896551, \"f1-score\": 0.7329139062332538, \"support\": 116.0}}", "eval_f1_macro": 69.2591412565419, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.29139062332538, "eval_loss": 0.9056476950645447, "eval_runtime": 2.3786, "eval_samples_per_second": 48.768, "step": 864 }, { "epoch": 55.0, "learning_rate": 0.00024166666666666664, "loss": 0.0034, "step": 880 }, { "epoch": 55.0, "eval_accuracy": 68.10344827586206, "eval_average_metrics": 67.43628999917863, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.5, \"recall\": 0.6666666666666666, \"f1-score\": 0.5714285714285715, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.6451612903225806, \"f1-score\": 0.6557377049180327, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.5675675675675675, \"f1-score\": 0.6461538461538462, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.6810344827586207, \"macro avg\": {\"precision\": 0.7209686147186147, \"recall\": 0.769646662791824, \"f1-score\": 0.7377364438839848, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.6941315868040006, \"recall\": 0.6810344827586207, \"f1-score\": 0.6796169065530286, \"support\": 116.0}}", "eval_f1_macro": 65.57657278968753, "eval_f1_micro": 68.10344827586206, "eval_f1_weighted": 67.96169065530286, "eval_loss": 1.0662262439727783, "eval_runtime": 2.2695, "eval_samples_per_second": 51.112, "step": 880 }, { "epoch": 56.0, "learning_rate": 0.00023999999999999998, "loss": 0.0158, "step": 896 }, { "epoch": 56.0, "eval_accuracy": 68.96551724137932, "eval_average_metrics": 68.23178854815569, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.46153846153846156, \"recall\": 0.6666666666666666, \"f1-score\": 0.5454545454545455, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5789473684210527, \"recall\": 0.9166666666666666, \"f1-score\": 0.7096774193548387, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.72, \"recall\": 0.5806451612903226, \"f1-score\": 0.6428571428571428, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7777777777777778, \"recall\": 0.5675675675675675, \"f1-score\": 0.65625, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 1.0, \"f1-score\": 0.8571428571428571, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.6896551724137931, \"macro avg\": {\"precision\": 0.726509141443352, \"recall\": 0.796304368885014, \"f1-score\": 0.7466608408392683, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7138459931880984, \"recall\": 0.6896551724137931, \"f1-score\": 0.6862626719081808, \"support\": 116.0}}", "eval_f1_macro": 66.36985251904606, "eval_f1_micro": 68.96551724137932, "eval_f1_weighted": 68.62626719081808, "eval_loss": 1.1591931581497192, "eval_runtime": 2.2123, "eval_samples_per_second": 52.434, "step": 896 }, { "epoch": 57.0, "learning_rate": 0.0002383333333333333, "loss": 0.0376, "step": 912 }, { "epoch": 57.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 69.91520968255182, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6428571428571429, \"recall\": 0.75, \"f1-score\": 0.6923076923076924, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.7096774193548387, \"f1-score\": 0.6875, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7333333333333333, \"recall\": 0.5945945945945946, \"f1-score\": 0.6567164179104478, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7554924242424242, \"recall\": 0.7810895572992347, \"f1-score\": 0.7630389031330098, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7126455441110613, \"recall\": 0.7068965517241379, \"f1-score\": 0.7045584810688993, \"support\": 116.0}}", "eval_f1_macro": 67.82568027848977, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.45584810688995, "eval_loss": 0.9944426417350769, "eval_runtime": 2.3478, "eval_samples_per_second": 49.409, "step": 912 }, { "epoch": 58.0, "learning_rate": 0.00023666666666666663, "loss": 0.0072, "step": 928 }, { "epoch": 58.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 70.84391114732561, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6764705882352942, \"recall\": 0.7419354838709677, \"f1-score\": 0.7076923076923077, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7586206896551724, \"recall\": 0.5945945945945946, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.768480475392874, \"recall\": 0.786857926474862, \"f1-score\": 0.7733734198807728, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7264381593086916, \"recall\": 0.7155172413793104, \"f1-score\": 0.7152789232403838, \"support\": 116.0}}", "eval_f1_macro": 68.74430398940203, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.52789232403838, "eval_loss": 0.9998703002929688, "eval_runtime": 2.4048, "eval_samples_per_second": 48.237, "step": 928 }, { "epoch": 59.0, "learning_rate": 0.00023499999999999997, "loss": 0.0045, "step": 944 }, { "epoch": 59.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 71.44905544977263, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.7096774193548387, \"f1-score\": 0.7096774193548389, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7741935483870968, \"recall\": 0.6486486486486487, \"f1-score\": 0.7058823529411764, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7650225614439323, \"recall\": 0.7878463140559915, \"f1-score\": 0.7710386616446239, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.734502621960909, \"recall\": 0.7241379310344828, \"f1-score\": 0.7243186566822737, \"support\": 116.0}}", "eval_f1_macro": 68.53676992396657, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.43186566822737, "eval_loss": 0.9982974529266357, "eval_runtime": 2.2032, "eval_samples_per_second": 52.652, "step": 944 }, { "epoch": 60.0, "learning_rate": 0.0002333333333333333, "loss": 0.0095, "step": 960 }, { "epoch": 60.0, "eval_accuracy": 69.82758620689656, "eval_average_metrics": 69.03488330679734, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.5, \"recall\": 0.6666666666666666, \"f1-score\": 0.5714285714285715, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6896551724137931, \"recall\": 0.6451612903225806, \"f1-score\": 0.6666666666666667, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7666666666666667, \"recall\": 0.6216216216216216, \"f1-score\": 0.6865671641791045, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7350164203612479, \"recall\": 0.7764034195485807, \"f1-score\": 0.7494173867504583, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7112337919710096, \"recall\": 0.6982758620689655, \"f1-score\": 0.6986948199113326, \"support\": 116.0}}", "eval_f1_macro": 66.61487882226295, "eval_f1_micro": 69.82758620689656, "eval_f1_weighted": 69.86948199113327, "eval_loss": 0.9828293919563293, "eval_runtime": 2.1348, "eval_samples_per_second": 54.338, "step": 960 }, { "epoch": 61.0, "learning_rate": 0.00023166666666666667, "loss": 0.0037, "step": 976 }, { "epoch": 61.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 69.7600508509814, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6486486486486487, \"f1-score\": 0.6956521739130435, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7619791666666667, \"recall\": 0.7681890559914752, \"f1-score\": 0.7532445691144488, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7237068965517242, \"recall\": 0.7068965517241379, \"f1-score\": 0.7070582024892478, \"support\": 116.0}}", "eval_f1_macro": 66.95507281017323, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.70582024892478, "eval_loss": 1.0811266899108887, "eval_runtime": 2.041, "eval_samples_per_second": 56.836, "step": 976 }, { "epoch": 62.0, "learning_rate": 0.00023, "loss": 0.0035, "step": 992 }, { "epoch": 62.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 70.90889013651437, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6756756756756757, \"f1-score\": 0.6944444444444444, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7938311688311688, \"recall\": 0.7634469146565921, \"f1-score\": 0.7717174369747899, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7324857254814152, \"recall\": 0.7155172413793104, \"f1-score\": 0.7193500676132522, \"support\": 116.0}}", "eval_f1_macro": 68.59710550887021, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.93500676132521, "eval_loss": 1.030908465385437, "eval_runtime": 2.3005, "eval_samples_per_second": 50.424, "step": 992 }, { "epoch": 63.0, "learning_rate": 0.0002283333333333333, "loss": 0.0086, "step": 1008 }, { "epoch": 63.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 71.35518309105177, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6756756756756757, \"f1-score\": 0.7246376811594203, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7590061745852187, \"recall\": 0.7935768429720043, \"f1-score\": 0.766072320982366, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7432002359962553, \"recall\": 0.7241379310344828, \"f1-score\": 0.724978287366558, \"support\": 116.0}}", "eval_f1_macro": 68.09531742065475, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.49782873665578, "eval_loss": 1.0668530464172363, "eval_runtime": 2.4387, "eval_samples_per_second": 47.567, "step": 1008 }, { "epoch": 64.0, "learning_rate": 0.00022666666666666663, "loss": 0.0016, "step": 1024 }, { "epoch": 64.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 72.52369399223397, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.7096774193548387, \"f1-score\": 0.7096774193548389, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7666666666666667, \"recall\": 0.6216216216216216, \"f1-score\": 0.6865671641791045, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.875, \"f1-score\": 0.9333333333333333, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7851557558507274, \"recall\": 0.8105096023442797, \"f1-score\": 0.7911406834106129, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7446162947937796, \"recall\": 0.7327586206896551, \"f1-score\": 0.7321943552783927, \"support\": 116.0}}", "eval_f1_macro": 70.32361630316558, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.21943552783927, "eval_loss": 1.0520766973495483, "eval_runtime": 2.3594, "eval_samples_per_second": 49.164, "step": 1024 }, { "epoch": 65.0, "learning_rate": 0.000225, "loss": 0.0091, "step": 1040 }, { "epoch": 65.0, "eval_accuracy": 68.96551724137932, "eval_average_metrics": 68.31228885061847, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.5454545454545454, \"recall\": 0.6666666666666666, \"f1-score\": 0.6, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6774193548387096, \"recall\": 0.6774193548387096, \"f1-score\": 0.6774193548387096, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.5675675675675675, \"f1-score\": 0.6461538461538462, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"request\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"accuracy\": 0.6896551724137931, \"macro avg\": {\"precision\": 0.7318522930922124, \"recall\": 0.7754150319674513, \"f1-score\": 0.7476391104415299, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7018025078369906, \"recall\": 0.6896551724137931, \"f1-score\": 0.6886131110269041, \"support\": 116.0}}", "eval_f1_macro": 66.45680981702488, "eval_f1_micro": 68.96551724137932, "eval_f1_weighted": 68.86131110269041, "eval_loss": 1.0356357097625732, "eval_runtime": 2.3808, "eval_samples_per_second": 48.723, "step": 1040 }, { "epoch": 66.0, "learning_rate": 0.00022333333333333333, "loss": 0.0028, "step": 1056 }, { "epoch": 66.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 72.2270820572477, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6756756756756757, \"f1-score\": 0.7246376811594203, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7725114193404634, \"recall\": 0.7976091010365204, \"f1-score\": 0.776483943258395, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7479355868522958, \"recall\": 0.7327586206896551, \"f1-score\": 0.7333580913475801, \"support\": 116.0}}", "eval_f1_macro": 69.02079495630178, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.335809134758, "eval_loss": 1.0336496829986572, "eval_runtime": 2.2453, "eval_samples_per_second": 51.664, "step": 1056 }, { "epoch": 67.0, "learning_rate": 0.00022166666666666667, "loss": 0.0028, "step": 1072 }, { "epoch": 67.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 71.6024787948385, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6756756756756757, \"f1-score\": 0.7142857142857142, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.787405303030303, \"recall\": 0.7773358035454809, \"f1-score\": 0.7755018674136321, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7389465778474399, \"recall\": 0.7241379310344828, \"f1-score\": 0.7264882964680124, \"support\": 116.0}}", "eval_f1_macro": 68.93349932565619, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.64882964680125, "eval_loss": 1.002924919128418, "eval_runtime": 2.1768, "eval_samples_per_second": 53.289, "step": 1072 }, { "epoch": 68.0, "learning_rate": 0.00021999999999999995, "loss": 0.0037, "step": 1088 }, { "epoch": 68.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 72.42795900402467, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.7096774193548387, \"f1-score\": 0.7213114754098361, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6486486486486487, \"f1-score\": 0.6956521739130435, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.875, \"f1-score\": 0.9333333333333333, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7832516339869281, \"recall\": 0.7999990918337692, \"f1-score\": 0.7856896309617154, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7438979039891818, \"recall\": 0.7327586206896551, \"f1-score\": 0.7332103357045962, \"support\": 116.0}}", "eval_f1_macro": 69.83907830770802, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.32103357045962, "eval_loss": 1.0405433177947998, "eval_runtime": 2.4018, "eval_samples_per_second": 48.298, "step": 1088 }, { "epoch": 69.0, "learning_rate": 0.0002183333333333333, "loss": 0.0023, "step": 1104 }, { "epoch": 69.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 70.44974265216908, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.8, \"recall\": 0.6451612903225806, \"f1-score\": 0.7142857142857142, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.7027027027027027, \"f1-score\": 0.7123287671232876, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.76635101010101, \"recall\": 0.7674413324614937, \"f1-score\": 0.753787379019047, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7385013932427725, \"recall\": 0.7155172413793104, \"f1-score\": 0.7169219975334343, \"support\": 116.0}}", "eval_f1_macro": 67.00332257947085, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.69219975334343, "eval_loss": 1.0624644756317139, "eval_runtime": 2.39, "eval_samples_per_second": 48.535, "step": 1104 }, { "epoch": 70.0, "learning_rate": 0.00021666666666666666, "loss": 0.0023, "step": 1120 }, { "epoch": 70.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 69.99088604087092, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.5675675675675675, \"f1-score\": 0.6461538461538462, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7510416666666666, \"recall\": 0.7933361789208564, \"f1-score\": 0.7666375473941263, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7156250000000001, \"recall\": 0.7068965517241379, \"f1-score\": 0.704386740502893, \"support\": 116.0}}", "eval_f1_macro": 68.1455597683668, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.4386740502893, "eval_loss": 1.127900242805481, "eval_runtime": 2.2682, "eval_samples_per_second": 51.141, "step": 1120 }, { "epoch": 71.0, "learning_rate": 0.000215, "loss": 0.0277, "step": 1136 }, { "epoch": 71.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 73.17305223934821, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7586206896551724, \"recall\": 0.7096774193548387, \"f1-score\": 0.7333333333333333, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6756756756756757, \"f1-score\": 0.7142857142857142, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.875, \"f1-score\": 0.9333333333333333, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7873595232241277, \"recall\": 0.8033774702121476, \"f1-score\": 0.7895215557487363, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7530721374053893, \"recall\": 0.7413793103448276, \"f1-score\": 0.7423665304409517, \"support\": 116.0}}", "eval_f1_macro": 70.17969384433212, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.23665304409515, "eval_loss": 1.0505918264389038, "eval_runtime": 2.3583, "eval_samples_per_second": 49.188, "step": 1136 }, { "epoch": 72.0, "learning_rate": 0.00021333333333333333, "loss": 0.0013, "step": 1152 }, { "epoch": 72.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 70.18153002432922, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6470588235294118, \"recall\": 0.7096774193548387, \"f1-score\": 0.6769230769230768, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7419354838709677, \"recall\": 0.6216216216216216, \"f1-score\": 0.676470588235294, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.875, \"f1-score\": 0.9333333333333333, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.776607242970502, \"recall\": 0.7723151578998353, \"f1-score\": 0.7698744882568411, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7207992168957784, \"recall\": 0.7068965517241379, \"f1-score\": 0.7091352190743672, \"support\": 116.0}}", "eval_f1_macro": 68.43328784505256, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.91352190743672, "eval_loss": 1.0885376930236816, "eval_runtime": 2.3047, "eval_samples_per_second": 50.332, "step": 1152 }, { "epoch": 73.0, "learning_rate": 0.00021166666666666667, "loss": 0.0062, "step": 1168 }, { "epoch": 73.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 69.98285623150008, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.5675675675675675, \"f1-score\": 0.6461538461538462, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7474877450980392, \"recall\": 0.7997205875230069, \"f1-score\": 0.7668653226453315, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7164553752535497, \"recall\": 0.7068965517241379, \"f1-score\": 0.7038630812380989, \"support\": 116.0}}", "eval_f1_macro": 68.1658064573628, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.38630812380991, "eval_loss": 1.1306726932525635, "eval_runtime": 2.2798, "eval_samples_per_second": 50.881, "step": 1168 }, { "epoch": 74.0, "learning_rate": 0.00020999999999999998, "loss": 0.0045, "step": 1184 }, { "epoch": 74.0, "eval_accuracy": 75.0, "eval_average_metrics": 74.14612206166869, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.7297297297297297, \"f1-score\": 0.7397260273972601, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.875, \"f1-score\": 0.9333333333333333, \"support\": 8.0}, \"accuracy\": 0.75, \"macro avg\": {\"precision\": 0.8087121212121212, \"recall\": 0.8061019689043882, \"f1-score\": 0.8013653477451443, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7672849181469871, \"recall\": 0.75, \"f1-score\": 0.7535201289155082, \"support\": 116.0}}", "eval_f1_macro": 71.23247535512394, "eval_f1_micro": 75.0, "eval_f1_weighted": 75.35201289155083, "eval_loss": 1.092178225517273, "eval_runtime": 2.2337, "eval_samples_per_second": 51.932, "step": 1184 }, { "epoch": 75.0, "learning_rate": 0.00020833333333333332, "loss": 0.0082, "step": 1200 }, { "epoch": 75.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 71.86428236984912, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6470588235294118, \"recall\": 0.7096774193548387, \"f1-score\": 0.6769230769230768, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7666666666666667, \"recall\": 0.6216216216216216, \"f1-score\": 0.6865671641791045, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"request\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7881740196078432, \"recall\": 0.8000929356776131, \"f1-score\": 0.7892650616503777, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7359956051386072, \"recall\": 0.7241379310344828, \"f1-score\": 0.7247264890357742, \"support\": 116.0}}", "eval_f1_macro": 70.15689436892248, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.47264890357744, "eval_loss": 1.0961277484893799, "eval_runtime": 2.248, "eval_samples_per_second": 51.601, "step": 1200 }, { "epoch": 76.0, "learning_rate": 0.00020666666666666666, "loss": 0.004, "step": 1216 }, { "epoch": 76.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 70.9403461206404, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.47619047619047616, \"recall\": 0.8333333333333334, \"f1-score\": 0.6060606060606061, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7391304347826086, \"recall\": 0.5483870967741935, \"f1-score\": 0.6296296296296297, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.7297297297297297, \"f1-score\": 0.7397260273972601, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"request\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7783355684170902, \"recall\": 0.7899729366463237, \"f1-score\": 0.7746843858271134, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.740125878619132, \"recall\": 0.7155172413793104, \"f1-score\": 0.7179710191095614, \"support\": 116.0}}", "eval_f1_macro": 68.86083429574342, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.79710191095613, "eval_loss": 1.1594845056533813, "eval_runtime": 2.0779, "eval_samples_per_second": 55.827, "step": 1216 }, { "epoch": 77.0, "learning_rate": 0.000205, "loss": 0.0064, "step": 1232 }, { "epoch": 77.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 69.73375044753716, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.6451612903225806, \"f1-score\": 0.689655172413793, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6486486486486487, \"f1-score\": 0.6956521739130435, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7420965608465608, \"recall\": 0.776309575704737, \"f1-score\": 0.7521734433535112, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7201445904032111, \"recall\": 0.7068965517241379, \"f1-score\": 0.7069582981389781, \"support\": 116.0}}", "eval_f1_macro": 66.85986163142323, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.69582981389783, "eval_loss": 1.1205899715423584, "eval_runtime": 2.1939, "eval_samples_per_second": 52.875, "step": 1232 }, { "epoch": 78.0, "learning_rate": 0.00020333333333333333, "loss": 0.0179, "step": 1248 }, { "epoch": 78.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 70.08500699570796, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.696969696969697, \"recall\": 0.6216216216216216, \"f1-score\": 0.6571428571428571, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7792585784313726, \"recall\": 0.767106824566502, \"f1-score\": 0.766470668727261, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7199265090048558, \"recall\": 0.7068965517241379, \"f1-score\": 0.7082999152891438, \"support\": 116.0}}", "eval_f1_macro": 68.13072610908988, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.82999152891438, "eval_loss": 1.1212310791015625, "eval_runtime": 2.3495, "eval_samples_per_second": 49.372, "step": 1248 }, { "epoch": 79.0, "learning_rate": 0.00020166666666666667, "loss": 0.002, "step": 1264 }, { "epoch": 79.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 71.4467055723817, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.8095238095238095, \"recall\": 0.5483870967741935, \"f1-score\": 0.6538461538461537, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7297297297297297, \"recall\": 0.7297297297297297, \"f1-score\": 0.7297297297297297, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 0.7777777777777778, \"recall\": 0.875, \"f1-score\": 0.823529411764706, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7618511368511369, \"recall\": 0.8038618255352126, \"f1-score\": 0.7739040349334467, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7412424740010948, \"recall\": 0.7241379310344828, \"f1-score\": 0.7216776631076833, \"support\": 116.0}}", "eval_f1_macro": 68.79146977186194, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.16776631076833, "eval_loss": 1.1796600818634033, "eval_runtime": 2.3002, "eval_samples_per_second": 50.43, "step": 1264 }, { "epoch": 80.0, "learning_rate": 0.00019999999999999998, "loss": 0.0039, "step": 1280 }, { "epoch": 80.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 69.91079996424841, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7333333333333333, \"recall\": 0.5945945945945946, \"f1-score\": 0.6567164179104478, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7565718901196843, \"recall\": 0.7776173350770126, \"f1-score\": 0.7615964881045013, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7158540728203506, \"recall\": 0.7068965517241379, \"f1-score\": 0.7056642390287705, \"support\": 116.0}}", "eval_f1_macro": 67.69746560928901, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.56642390287705, "eval_loss": 1.1252835988998413, "eval_runtime": 2.5769, "eval_samples_per_second": 45.014, "step": 1280 }, { "epoch": 81.0, "learning_rate": 0.00019833333333333332, "loss": 0.0033, "step": 1296 }, { "epoch": 81.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 72.81511711278446, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.8333333333333334, \"recall\": 0.6451612903225806, \"f1-score\": 0.7272727272727272, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7567567567567568, \"recall\": 0.7567567567567568, \"f1-score\": 0.7567567567567567, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.78619812994813, \"recall\": 0.7880869781071393, \"f1-score\": 0.7734642543466073, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7654780564263323, \"recall\": 0.7413793103448276, \"f1-score\": 0.7423222821802943, \"support\": 116.0}}", "eval_f1_macro": 68.75237816414287, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.23222821802943, "eval_loss": 1.1057864427566528, "eval_runtime": 2.4718, "eval_samples_per_second": 46.929, "step": 1296 }, { "epoch": 82.0, "learning_rate": 0.00019666666666666666, "loss": 0.0009, "step": 1312 }, { "epoch": 82.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 71.35367532053087, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.6774193548387096, \"f1-score\": 0.7, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.7027027027027027, \"f1-score\": 0.7222222222222223, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7801992505466137, \"recall\": 0.77147359052601, \"f1-score\": 0.766136972495732, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7385352276247219, \"recall\": 0.7241379310344828, \"f1-score\": 0.7248605085338408, \"support\": 116.0}}", "eval_f1_macro": 68.10106422184285, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.48605085338409, "eval_loss": 1.0622020959854126, "eval_runtime": 2.3973, "eval_samples_per_second": 48.389, "step": 1312 }, { "epoch": 83.0, "learning_rate": 0.000195, "loss": 0.0018, "step": 1328 }, { "epoch": 83.0, "eval_accuracy": 68.10344827586206, "eval_average_metrics": 67.65207391031592, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.47619047619047616, \"recall\": 0.8333333333333334, \"f1-score\": 0.6060606060606061, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.6451612903225806, \"f1-score\": 0.6557377049180327, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.5675675675675675, \"f1-score\": 0.6461538461538462, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.6810344827586207, \"macro avg\": {\"precision\": 0.7420454545454545, \"recall\": 0.7661744405696018, \"f1-score\": 0.7444924716539445, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7040117928048962, \"recall\": 0.6810344827586207, \"f1-score\": 0.6822429049807779, \"support\": 116.0}}", "eval_f1_macro": 66.17710859146173, "eval_f1_micro": 68.10344827586206, "eval_f1_weighted": 68.22429049807779, "eval_loss": 1.2004237174987793, "eval_runtime": 2.4489, "eval_samples_per_second": 47.369, "step": 1328 }, { "epoch": 84.0, "learning_rate": 0.00019333333333333333, "loss": 0.0012, "step": 1344 }, { "epoch": 84.0, "eval_accuracy": 68.96551724137932, "eval_average_metrics": 68.13078144338341, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.5675675675675675, \"f1-score\": 0.6461538461538462, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.6896551724137931, \"macro avg\": {\"precision\": 0.7366228070175438, \"recall\": 0.7684705875230069, \"f1-score\": 0.7404771647095969, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7084694494857835, \"recall\": 0.6896551724137931, \"f1-score\": 0.6877189887214419, \"support\": 116.0}}", "eval_f1_macro": 65.82019241863084, "eval_f1_micro": 68.96551724137932, "eval_f1_weighted": 68.77189887214419, "eval_loss": 1.1874313354492188, "eval_runtime": 2.4341, "eval_samples_per_second": 47.656, "step": 1344 }, { "epoch": 85.0, "learning_rate": 0.00019166666666666665, "loss": 0.0044, "step": 1360 }, { "epoch": 85.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 72.94211721069689, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.8076923076923077, \"recall\": 0.6774193548387096, \"f1-score\": 0.7368421052631579, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.717948717948718, \"recall\": 0.7567567567567568, \"f1-score\": 0.736842105263158, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7986411056263998, \"recall\": 0.7782303472827666, \"f1-score\": 0.778417674250118, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7598160375797294, \"recall\": 0.7413793103448276, \"f1-score\": 0.7429992461825603, \"support\": 116.0}}", "eval_f1_macro": 69.19268215556605, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.29992461825603, "eval_loss": 1.1297322511672974, "eval_runtime": 2.3474, "eval_samples_per_second": 49.416, "step": 1360 }, { "epoch": 86.0, "learning_rate": 0.00018999999999999998, "loss": 0.0179, "step": 1376 }, { "epoch": 86.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 70.03168438592736, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6388888888888888, \"recall\": 0.7419354838709677, \"f1-score\": 0.6865671641791045, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7419354838709677, \"recall\": 0.6216216216216216, \"f1-score\": 0.676470588235294, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.78548183447377, \"recall\": 0.7589863048532404, \"f1-score\": 0.7633472906958421, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7257279412590869, \"recall\": 0.7068965517241379, \"f1-score\": 0.7089433469258483, \"support\": 116.0}}", "eval_f1_macro": 67.85309250629707, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.89433469258482, "eval_loss": 1.117130994796753, "eval_runtime": 2.394, "eval_samples_per_second": 48.455, "step": 1376 }, { "epoch": 87.0, "learning_rate": 0.00018833333333333332, "loss": 0.0013, "step": 1392 }, { "epoch": 87.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 70.59785642558124, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.8, \"recall\": 0.5161290322580645, \"f1-score\": 0.6274509803921569, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7368421052631579, \"recall\": 0.7567567567567568, \"f1-score\": 0.7466666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7662280701754387, \"recall\": 0.7875829458490748, \"f1-score\": 0.7640013458694174, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7411978221415607, \"recall\": 0.7155172413793104, \"f1-score\": 0.7137674668251471, \"support\": 116.0}}", "eval_f1_macro": 67.91123074394821, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.3767466825147, "eval_loss": 1.1777045726776123, "eval_runtime": 2.2717, "eval_samples_per_second": 51.064, "step": 1392 }, { "epoch": 88.0, "learning_rate": 0.00018666666666666666, "loss": 0.0047, "step": 1408 }, { "epoch": 88.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 70.77918320692154, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6571428571428571, \"recall\": 0.7419354838709677, \"f1-score\": 0.6969696969696969, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7857142857142857, \"recall\": 0.5945945945945946, \"f1-score\": 0.676923076923077, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"request\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7657738095238095, \"recall\": 0.786857926474862, \"f1-score\": 0.7709575351119469, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7286022167487685, \"recall\": 0.7155172413793104, \"f1-score\": 0.714837258752066, \"support\": 116.0}}", "eval_f1_macro": 68.5295586766175, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.4837258752066, "eval_loss": 1.106935739517212, "eval_runtime": 2.2634, "eval_samples_per_second": 51.251, "step": 1408 }, { "epoch": 89.0, "learning_rate": 0.000185, "loss": 0.0042, "step": 1424 }, { "epoch": 89.0, "eval_accuracy": 69.82758620689656, "eval_average_metrics": 69.00121160981696, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.47619047619047616, \"recall\": 0.8333333333333334, \"f1-score\": 0.6060606060606061, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7619047619047619, \"recall\": 0.5161290322580645, \"f1-score\": 0.6153846153846153, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.7027027027027027, \"f1-score\": 0.7323943661971832, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7408263305322129, \"recall\": 0.780826189092318, \"f1-score\": 0.7486785700091854, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7248671882546122, \"recall\": 0.6982758620689655, \"f1-score\": 0.6980046780243607, \"support\": 116.0}}", "eval_f1_macro": 66.5492062230387, "eval_f1_micro": 69.82758620689656, "eval_f1_weighted": 69.80046780243605, "eval_loss": 1.2019020318984985, "eval_runtime": 2.4101, "eval_samples_per_second": 48.131, "step": 1424 }, { "epoch": 90.0, "learning_rate": 0.00018333333333333334, "loss": 0.0019, "step": 1440 }, { "epoch": 90.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 72.61043030355066, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7931034482758621, \"recall\": 0.6216216216216216, \"f1-score\": 0.696969696969697, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"request\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7909782088122606, \"recall\": 0.8105096023442797, \"f1-score\": 0.7932515809354045, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7506378814902894, \"recall\": 0.7327586206896551, \"f1-score\": 0.7337874543756896, \"support\": 116.0}}", "eval_f1_macro": 70.51125163870262, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.37874543756897, "eval_loss": 1.147979974746704, "eval_runtime": 2.2001, "eval_samples_per_second": 52.726, "step": 1440 }, { "epoch": 91.0, "learning_rate": 0.00018166666666666665, "loss": 0.0084, "step": 1456 }, { "epoch": 91.0, "eval_accuracy": 69.82758620689656, "eval_average_metrics": 69.01145238292938, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7307692307692307, \"recall\": 0.6129032258064516, \"f1-score\": 0.6666666666666667, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7419354838709677, \"recall\": 0.6216216216216216, \"f1-score\": 0.676470588235294, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7349178151597506, \"recall\": 0.7827878281507313, \"f1-score\": 0.7504109538552264, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7118510614339314, \"recall\": 0.6982758620689655, \"f1-score\": 0.6968744121968207, \"support\": 116.0}}", "eval_f1_macro": 66.70319589824234, "eval_f1_micro": 69.82758620689656, "eval_f1_weighted": 69.68744121968207, "eval_loss": 1.1545029878616333, "eval_runtime": 2.3358, "eval_samples_per_second": 49.662, "step": 1456 }, { "epoch": 92.0, "learning_rate": 0.00017999999999999998, "loss": 0.0013, "step": 1472 }, { "epoch": 92.0, "eval_accuracy": 69.82758620689656, "eval_average_metrics": 69.01145238292938, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7307692307692307, \"recall\": 0.6129032258064516, \"f1-score\": 0.6666666666666667, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7419354838709677, \"recall\": 0.6216216216216216, \"f1-score\": 0.676470588235294, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7349178151597506, \"recall\": 0.7827878281507313, \"f1-score\": 0.7504109538552264, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7118510614339314, \"recall\": 0.6982758620689655, \"f1-score\": 0.6968744121968207, \"support\": 116.0}}", "eval_f1_macro": 66.70319589824234, "eval_f1_micro": 69.82758620689656, "eval_f1_weighted": 69.68744121968207, "eval_loss": 1.1468178033828735, "eval_runtime": 2.2416, "eval_samples_per_second": 51.749, "step": 1472 }, { "epoch": 93.0, "learning_rate": 0.00017833333333333332, "loss": 0.0014, "step": 1488 }, { "epoch": 93.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 70.0316545956725, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.7096774193548387, \"f1-score\": 0.6875, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.5675675675675675, \"f1-score\": 0.6461538461538462, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7582831996434938, \"recall\": 0.7898639566986341, \"f1-score\": 0.7682870793333388, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7168341631323376, \"recall\": 0.7068965517241379, \"f1-score\": 0.7045512320823231, \"support\": 116.0}}", "eval_f1_macro": 68.29218482963012, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.45512320823231, "eval_loss": 1.1516327857971191, "eval_runtime": 2.4305, "eval_samples_per_second": 47.727, "step": 1488 }, { "epoch": 94.0, "learning_rate": 0.00017666666666666666, "loss": 0.002, "step": 1504 }, { "epoch": 94.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 69.98519551620497, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.6451612903225806, \"f1-score\": 0.689655172413793, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.696969696969697, \"recall\": 0.6216216216216216, \"f1-score\": 0.6571428571428571, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7504970043572985, \"recall\": 0.7885561973263586, \"f1-score\": 0.7650281414743756, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7132494655821227, \"recall\": 0.7068965517241379, \"f1-score\": 0.7055897025560334, \"support\": 116.0}}", "eval_f1_macro": 68.00250146438894, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.55897025560334, "eval_loss": 1.162688136100769, "eval_runtime": 2.1458, "eval_samples_per_second": 54.059, "step": 1504 }, { "epoch": 95.0, "learning_rate": 0.000175, "loss": 0.0055, "step": 1520 }, { "epoch": 95.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 71.0756914574162, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.625, \"recall\": 0.8333333333333334, \"f1-score\": 0.7142857142857143, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6470588235294118, \"recall\": 0.7096774193548387, \"f1-score\": 0.6769230769230768, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7333333333333333, \"recall\": 0.5945945945945946, \"f1-score\": 0.6567164179104478, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"request\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7833444741532977, \"recall\": 0.7932423350770126, \"f1-score\": 0.7837480040810814, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7255985309484294, \"recall\": 0.7155172413793104, \"f1-score\": 0.7153282830215106, \"support\": 116.0}}", "eval_f1_macro": 69.66648925165167, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.53282830215106, "eval_loss": 1.1537128686904907, "eval_runtime": 2.3897, "eval_samples_per_second": 48.543, "step": 1520 }, { "epoch": 96.0, "learning_rate": 0.0001733333333333333, "loss": 0.0049, "step": 1536 }, { "epoch": 96.0, "eval_accuracy": 69.82758620689656, "eval_average_metrics": 68.8282167495548, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7619047619047619, \"recall\": 0.5161290322580645, \"f1-score\": 0.6153846153846153, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.717948717948718, \"recall\": 0.7567567567567568, \"f1-score\": 0.736842105263158, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.75531661321135, \"recall\": 0.758069056960186, \"f1-score\": 0.74253787384093, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.722639761931958, \"recall\": 0.6982758620689655, \"f1-score\": 0.6965432802078783, \"support\": 116.0}}", "eval_f1_macro": 66.00336656363822, "eval_f1_micro": 69.82758620689656, "eval_f1_weighted": 69.65432802078783, "eval_loss": 1.1960561275482178, "eval_runtime": 2.4859, "eval_samples_per_second": 46.663, "step": 1536 }, { "epoch": 97.0, "learning_rate": 0.00017166666666666665, "loss": 0.0142, "step": 1552 }, { "epoch": 97.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 70.00234719378676, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6774193548387096, \"recall\": 0.6774193548387096, \"f1-score\": 0.6774193548387096, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7058823529411765, \"recall\": 0.6486486486486487, \"f1-score\": 0.676056338028169, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7808486908937573, \"recall\": 0.7647168337692531, \"f1-score\": 0.7633916095426883, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7211529596164484, \"recall\": 0.7068965517241379, \"f1-score\": 0.7077304647096941, \"support\": 116.0}}", "eval_f1_macro": 67.85703195935007, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.77304647096942, "eval_loss": 1.1356544494628906, "eval_runtime": 2.2827, "eval_samples_per_second": 50.817, "step": 1552 }, { "epoch": 98.0, "learning_rate": 0.00016999999999999999, "loss": 0.0011, "step": 1568 }, { "epoch": 98.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 70.02228110967859, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5, \"recall\": 0.8333333333333334, \"f1-score\": 0.625, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.72, \"recall\": 0.5806451612903226, \"f1-score\": 0.6428571428571428, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6486486486486487, \"f1-score\": 0.6956521739130435, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7577489177489177, \"recall\": 0.7960228373534826, \"f1-score\": 0.7664120539520156, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7244894760412003, \"recall\": 0.7068965517241379, \"f1-score\": 0.7058429818704093, \"support\": 116.0}}", "eval_f1_macro": 68.12551590684583, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.58429818704093, "eval_loss": 1.159824252128601, "eval_runtime": 2.3383, "eval_samples_per_second": 49.609, "step": 1568 }, { "epoch": 99.0, "learning_rate": 0.00016833333333333332, "loss": 0.001, "step": 1584 }, { "epoch": 99.0, "eval_accuracy": 69.82758620689656, "eval_average_metrics": 68.93085179723111, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.8095238095238095, \"recall\": 0.5483870967741935, \"f1-score\": 0.6538461538461537, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.6756756756756757, \"recall\": 0.6756756756756757, \"f1-score\": 0.6756756756756757, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 1.0, \"f1-score\": 0.8571428571428571, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7446777134277135, \"recall\": 0.7797439576673447, \"f1-score\": 0.7495235620235621, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7170566502463054, \"recall\": 0.6982758620689655, \"f1-score\": 0.694439181508147, \"support\": 116.0}}", "eval_f1_macro": 66.62431662431662, "eval_f1_micro": 69.82758620689656, "eval_f1_weighted": 69.4439181508147, "eval_loss": 1.1499238014221191, "eval_runtime": 2.5501, "eval_samples_per_second": 45.488, "step": 1584 }, { "epoch": 100.0, "learning_rate": 0.00016666666666666666, "loss": 0.007, "step": 1600 }, { "epoch": 100.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 71.4194327673127, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.625, \"recall\": 0.8333333333333334, \"f1-score\": 0.7142857142857143, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7058823529411765, \"recall\": 0.6486486486486487, \"f1-score\": 0.676056338028169, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7686330213903743, \"recall\": 0.7924946115470308, \"f1-score\": 0.7730349110281496, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7297237076648841, \"recall\": 0.7241379310344828, \"f1-score\": 0.7213593054874093, \"support\": 116.0}}", "eval_f1_macro": 68.71421431361331, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.13593054874093, "eval_loss": 1.0965989828109741, "eval_runtime": 2.3288, "eval_samples_per_second": 49.81, "step": 1600 }, { "epoch": 101.0, "learning_rate": 0.000165, "loss": 0.0002, "step": 1616 }, { "epoch": 101.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 71.41445986904613, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7352941176470589, \"recall\": 0.6756756756756757, \"f1-score\": 0.7042253521126761, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7724673202614379, \"recall\": 0.7819841010365204, \"f1-score\": 0.7692147679474433, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7374070317782286, \"recall\": 0.7241379310344828, \"f1-score\": 0.7245560722951524, \"support\": 116.0}}", "eval_f1_macro": 68.37464603977274, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.45560722951524, "eval_loss": 1.1273354291915894, "eval_runtime": 2.3586, "eval_samples_per_second": 49.182, "step": 1616 }, { "epoch": 102.0, "learning_rate": 0.0001633333333333333, "loss": 0.0002, "step": 1632 }, { "epoch": 102.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 72.15508814394651, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.7027027027027027, \"f1-score\": 0.7222222222222223, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7768849206349207, \"recall\": 0.7853624794148988, \"f1-score\": 0.7729985672429891, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7472427476737821, \"recall\": 0.7327586206896551, \"f1-score\": 0.7335764468292271, \"support\": 116.0}}", "eval_f1_macro": 68.71098375493236, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.35764468292271, "eval_loss": 1.1395922899246216, "eval_runtime": 2.2846, "eval_samples_per_second": 50.775, "step": 1632 }, { "epoch": 103.0, "learning_rate": 0.00016166666666666665, "loss": 0.0023, "step": 1648 }, { "epoch": 103.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 70.42410021572994, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.8260869565217391, \"recall\": 0.6129032258064516, \"f1-score\": 0.7037037037037037, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6756756756756757, \"f1-score\": 0.6944444444444444, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 1.0, \"f1-score\": 0.8571428571428571, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7473259956156375, \"recall\": 0.787808473796377, \"f1-score\": 0.756589450157962, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7320065282484809, \"recall\": 0.7155172413793104, \"f1-score\": 0.7134055701746107, \"support\": 116.0}}", "eval_f1_macro": 67.25239556959663, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.34055701746107, "eval_loss": 1.1940408945083618, "eval_runtime": 2.2464, "eval_samples_per_second": 51.637, "step": 1648 }, { "epoch": 104.0, "learning_rate": 0.00015999999999999999, "loss": 0.0078, "step": 1664 }, { "epoch": 104.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 70.8781299077741, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7419354838709677, \"recall\": 0.6216216216216216, \"f1-score\": 0.676470588235294, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7823500704045059, \"recall\": 0.7809957134553909, \"f1-score\": 0.7730421335200748, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7330855381640977, \"recall\": 0.7155172413793104, \"f1-score\": 0.7169421504233877, \"support\": 116.0}}", "eval_f1_macro": 68.71485631289553, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.69421504233877, "eval_loss": 1.1880056858062744, "eval_runtime": 2.3241, "eval_samples_per_second": 49.913, "step": 1664 }, { "epoch": 105.0, "learning_rate": 0.00015833333333333332, "loss": 0.0007, "step": 1680 }, { "epoch": 105.0, "eval_accuracy": 68.10344827586206, "eval_average_metrics": 67.81830142186269, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.44, \"recall\": 0.9166666666666666, \"f1-score\": 0.5945945945945945, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.68, \"recall\": 0.5483870967741935, \"f1-score\": 0.6071428571428571, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6486486486486487, \"f1-score\": 0.6956521739130435, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.6810344827586207, \"macro avg\": {\"precision\": 0.7641883116883117, \"recall\": 0.7607405792889663, \"f1-score\": 0.747283997323959, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7193484102104792, \"recall\": 0.6810344827586207, \"f1-score\": 0.6864106492915251, \"support\": 116.0}}", "eval_f1_macro": 66.42524420657413, "eval_f1_micro": 68.10344827586206, "eval_f1_weighted": 68.6410649291525, "eval_loss": 1.272948980331421, "eval_runtime": 2.2187, "eval_samples_per_second": 52.283, "step": 1680 }, { "epoch": 106.0, "learning_rate": 0.00015666666666666666, "loss": 0.0006, "step": 1696 }, { "epoch": 106.0, "eval_accuracy": 68.96551724137932, "eval_average_metrics": 68.14354636082183, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7096774193548387, \"recall\": 0.5945945945945946, \"f1-score\": 0.6470588235294118, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.6896551724137931, \"macro avg\": {\"precision\": 0.7390253339850115, \"recall\": 0.7579600770124963, \"f1-score\": 0.7403022684483629, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7009838372602555, \"recall\": 0.6896551724137931, \"f1-score\": 0.6883850487622973, \"support\": 116.0}}", "eval_f1_macro": 65.80464608429892, "eval_f1_micro": 68.96551724137932, "eval_f1_weighted": 68.83850487622973, "eval_loss": 1.1888731718063354, "eval_runtime": 2.3283, "eval_samples_per_second": 49.821, "step": 1696 }, { "epoch": 107.0, "learning_rate": 0.000155, "loss": 0.0001, "step": 1712 }, { "epoch": 107.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 71.35838392975816, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.84, \"recall\": 0.6774193548387096, \"f1-score\": 0.75, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.6842105263157895, \"recall\": 0.7027027027027027, \"f1-score\": 0.6933333333333334, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.778831240031898, \"recall\": 0.774945812748232, \"f1-score\": 0.7654010450333979, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7448383105098169, \"recall\": 0.7241379310344828, \"f1-score\": 0.7257030106472298, \"support\": 116.0}}", "eval_f1_macro": 68.03564844741315, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.57030106472297, "eval_loss": 1.1652153730392456, "eval_runtime": 2.2279, "eval_samples_per_second": 52.067, "step": 1712 }, { "epoch": 108.0, "learning_rate": 0.0001533333333333333, "loss": 0.0011, "step": 1728 }, { "epoch": 108.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 69.89276815025954, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.71875, \"recall\": 0.6216216216216216, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7675398284313726, \"recall\": 0.7752273442797636, \"f1-score\": 0.7605942362755609, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7202808062880325, \"recall\": 0.7068965517241379, \"f1-score\": 0.7058338569838294, \"support\": 116.0}}", "eval_f1_macro": 67.60837655782763, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.58338569838294, "eval_loss": 1.1787962913513184, "eval_runtime": 2.3198, "eval_samples_per_second": 50.005, "step": 1728 }, { "epoch": 109.0, "learning_rate": 0.00015166666666666665, "loss": 0.0003, "step": 1744 }, { "epoch": 109.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 70.57402020858294, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.5, \"recall\": 0.6666666666666666, \"f1-score\": 0.5714285714285715, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7916666666666666, \"recall\": 0.6129032258064516, \"f1-score\": 0.6909090909090909, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7352941176470589, \"recall\": 0.6756756756756757, \"f1-score\": 0.7042253521126761, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7625055704099822, \"recall\": 0.787808473796377, \"f1-score\": 0.7609192147499579, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7385345595918618, \"recall\": 0.7155172413793104, \"f1-score\": 0.7155536902514009, \"support\": 116.0}}", "eval_f1_macro": 67.63726353332959, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.5553690251401, "eval_loss": 1.215308427810669, "eval_runtime": 2.4262, "eval_samples_per_second": 47.812, "step": 1744 }, { "epoch": 110.0, "learning_rate": 0.00015, "loss": 0.0007, "step": 1760 }, { "epoch": 110.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 70.4562231169158, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7916666666666666, \"recall\": 0.6129032258064516, \"f1-score\": 0.6909090909090909, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.7027027027027027, \"f1-score\": 0.7222222222222223, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7605023448773449, \"recall\": 0.7772979632858665, \"f1-score\": 0.7553140572877415, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7382715703836394, \"recall\": 0.7155172413793104, \"f1-score\": 0.7158241687733521, \"support\": 116.0}}", "eval_f1_macro": 67.1390273144659, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.5824168773352, "eval_loss": 1.2416104078292847, "eval_runtime": 2.2686, "eval_samples_per_second": 51.133, "step": 1760 }, { "epoch": 111.0, "learning_rate": 0.00014833333333333332, "loss": 0.0006, "step": 1776 }, { "epoch": 111.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 70.4444570130092, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.8, \"recall\": 0.6451612903225806, \"f1-score\": 0.7142857142857142, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.7027027027027027, \"f1-score\": 0.7123287671232876, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7635101010101011, \"recall\": 0.7674413324614937, \"f1-score\": 0.7536473229966381, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.73673807035876, \"recall\": 0.7155172413793104, \"f1-score\": 0.7168350662091805, \"support\": 116.0}}", "eval_f1_macro": 66.99087315525671, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.68350662091805, "eval_loss": 1.234352946281433, "eval_runtime": 2.3796, "eval_samples_per_second": 48.747, "step": 1776 }, { "epoch": 112.0, "learning_rate": 0.00014666666666666664, "loss": 0.0005, "step": 1792 }, { "epoch": 112.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 71.34661390855857, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.7096774193548387, \"f1-score\": 0.7213114754098361, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6756756756756757, \"f1-score\": 0.7142857142857142, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7791035353535354, \"recall\": 0.7721274702121477, \"f1-score\": 0.7649352802115073, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7423066875653083, \"recall\": 0.7241379310344828, \"f1-score\": 0.7256462229742598, \"support\": 116.0}}", "eval_f1_macro": 67.99424712991176, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.56462229742598, "eval_loss": 1.2403353452682495, "eval_runtime": 2.3489, "eval_samples_per_second": 49.385, "step": 1792 }, { "epoch": 113.0, "learning_rate": 0.000145, "loss": 0.0002, "step": 1808 }, { "epoch": 113.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 70.37924043117656, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.7096774193548387, \"f1-score\": 0.7213114754098361, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7741935483870968, \"recall\": 0.6486486486486487, \"f1-score\": 0.7058823529411764, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7512565167807104, \"recall\": 0.7687490918337693, \"f1-score\": 0.7517535702805713, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7304703037044528, \"recall\": 0.7155172413793104, \"f1-score\": 0.7159097386834893, \"support\": 116.0}}", "eval_f1_macro": 66.82253958049523, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.59097386834891, "eval_loss": 1.2404866218566895, "eval_runtime": 2.2047, "eval_samples_per_second": 52.614, "step": 1808 }, { "epoch": 114.0, "learning_rate": 0.00014333333333333334, "loss": 0.0001, "step": 1824 }, { "epoch": 114.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 70.44727017147791, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5, \"recall\": 0.8333333333333334, \"f1-score\": 0.625, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7727272727272727, \"recall\": 0.5483870967741935, \"f1-score\": 0.6415094339622641, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7941176470588235, \"recall\": 0.7297297297297297, \"f1-score\": 0.7605633802816901, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 1.0, \"f1-score\": 0.8571428571428571, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7604389483065954, \"recall\": 0.7865007144241015, \"f1-score\": 0.7558662687363454, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7455767103079475, \"recall\": 0.7155172413793104, \"f1-score\": 0.7149751963348554, \"support\": 116.0}}", "eval_f1_macro": 67.18811277656404, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.49751963348554, "eval_loss": 1.2582147121429443, "eval_runtime": 2.2366, "eval_samples_per_second": 51.864, "step": 1824 }, { "epoch": 115.0, "learning_rate": 0.00014166666666666665, "loss": 0.0014, "step": 1840 }, { "epoch": 115.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 71.7426864751411, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6428571428571429, \"recall\": 0.75, \"f1-score\": 0.6923076923076924, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6388888888888888, \"recall\": 0.7419354838709677, \"f1-score\": 0.6865671641791045, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7857142857142857, \"recall\": 0.5945945945945946, \"f1-score\": 0.676923076923077, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7810019841269841, \"recall\": 0.8007468153637509, \"f1-score\": 0.7859342421430876, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7357006020799124, \"recall\": 0.7241379310344828, \"f1-score\": 0.7228233816983779, \"support\": 116.0}}", "eval_f1_macro": 69.86082152383003, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.2823381698378, "eval_loss": 1.202860713005066, "eval_runtime": 2.3445, "eval_samples_per_second": 49.477, "step": 1840 }, { "epoch": 116.0, "learning_rate": 0.00014, "loss": 0.0003, "step": 1856 }, { "epoch": 116.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 70.5549298094937, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7916666666666666, \"recall\": 0.6129032258064516, \"f1-score\": 0.6909090909090909, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.7027027027027027, \"f1-score\": 0.7027027027027027, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6923076923076923, \"recall\": 1.0, \"f1-score\": 0.8181818181818181, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.770730466042966, \"recall\": 0.7807701855080886, \"f1-score\": 0.7614585798409328, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7351072060123783, \"recall\": 0.7155172413793104, \"f1-score\": 0.7143106386514095, \"support\": 116.0}}", "eval_f1_macro": 67.6852070969718, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.43106386514096, "eval_loss": 1.2238714694976807, "eval_runtime": 2.1973, "eval_samples_per_second": 52.793, "step": 1856 }, { "epoch": 117.0, "learning_rate": 0.00013833333333333333, "loss": 0.0028, "step": 1872 }, { "epoch": 117.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 70.59935701207709, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6486486486486487, \"f1-score\": 0.6956521739130435, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 1.0, \"f1-score\": 0.8571428571428571, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7718248663101605, \"recall\": 0.778605722658142, \"f1-score\": 0.7632609894756968, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7304835884196939, \"recall\": 0.7155172413793104, \"f1-score\": 0.7144855848571772, \"support\": 116.0}}", "eval_f1_macro": 67.84542128672861, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.44855848571771, "eval_loss": 1.1936085224151611, "eval_runtime": 2.2205, "eval_samples_per_second": 52.241, "step": 1872 }, { "epoch": 118.0, "learning_rate": 0.00013666666666666666, "loss": 0.0033, "step": 1888 }, { "epoch": 118.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 70.58604104216032, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7741935483870968, \"recall\": 0.6486486486486487, \"f1-score\": 0.7058823529411764, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 1.0, \"f1-score\": 0.8571428571428571, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7707640925382861, \"recall\": 0.778605722658142, \"f1-score\": 0.7616661986358225, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.734819833485017, \"recall\": 0.7155172413793104, \"f1-score\": 0.715370537918172, \"support\": 116.0}}", "eval_f1_macro": 67.703662100962, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.5370537918172, "eval_loss": 1.1629191637039185, "eval_runtime": 2.2141, "eval_samples_per_second": 52.391, "step": 1888 }, { "epoch": 119.0, "learning_rate": 0.000135, "loss": 0.0002, "step": 1904 }, { "epoch": 119.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 69.70267547622983, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7037037037037037, \"recall\": 0.6129032258064516, \"f1-score\": 0.6551724137931035, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6756756756756757, \"f1-score\": 0.7142857142857142, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 1.0, \"f1-score\": 0.8571428571428571, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7544847989700931, \"recall\": 0.7739195849074881, \"f1-score\": 0.7521332342793323, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7222518457051925, \"recall\": 0.7068965517241379, \"f1-score\": 0.7057510406859552, \"support\": 116.0}}", "eval_f1_macro": 66.8562874914962, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.57510406859552, "eval_loss": 1.1673834323883057, "eval_runtime": 2.2399, "eval_samples_per_second": 51.788, "step": 1904 }, { "epoch": 120.0, "learning_rate": 0.0001333333333333333, "loss": 0.0008, "step": 1920 }, { "epoch": 120.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 69.87312178530907, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7272727272727273, \"recall\": 0.6486486486486487, \"f1-score\": 0.6857142857142857, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.77896148989899, \"recall\": 0.7548602029448803, \"f1-score\": 0.7561795230912878, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7262114681295715, \"recall\": 0.7068965517241379, \"f1-score\": 0.7089721918829424, \"support\": 116.0}}", "eval_f1_macro": 67.21595760811448, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.89721918829423, "eval_loss": 1.1733537912368774, "eval_runtime": 2.3365, "eval_samples_per_second": 49.648, "step": 1920 }, { "epoch": 121.0, "learning_rate": 0.00013166666666666665, "loss": 0.0007, "step": 1936 }, { "epoch": 121.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 71.45106394557037, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.6774193548387096, \"f1-score\": 0.7, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.7027027027027027, \"f1-score\": 0.7222222222222223, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7872253943374633, \"recall\": 0.77147359052601, \"f1-score\": 0.7691113624937154, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7420511167389884, \"recall\": 0.7241379310344828, \"f1-score\": 0.7261121513149911, \"support\": 116.0}}", "eval_f1_macro": 68.36545444388584, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.61121513149911, "eval_loss": 1.1702574491500854, "eval_runtime": 2.2206, "eval_samples_per_second": 52.239, "step": 1936 }, { "epoch": 122.0, "learning_rate": 0.00013, "loss": 0.0003, "step": 1952 }, { "epoch": 122.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 71.22487176013084, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.6451612903225806, \"f1-score\": 0.689655172413793, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.7027027027027027, \"f1-score\": 0.7323943661971832, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7575435729847495, \"recall\": 0.7813302213503825, \"f1-score\": 0.7620833896299759, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7343325069491398, \"recall\": 0.7241379310344828, \"f1-score\": 0.723311550887401, \"support\": 116.0}}", "eval_f1_macro": 67.74074574488674, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.33115508874009, "eval_loss": 1.178951382637024, "eval_runtime": 2.1307, "eval_samples_per_second": 54.442, "step": 1952 }, { "epoch": 123.0, "learning_rate": 0.00012833333333333333, "loss": 0.0006, "step": 1968 }, { "epoch": 123.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 71.50574768110899, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6764705882352942, \"recall\": 0.7419354838709677, \"f1-score\": 0.7076923076923077, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.8148148148148148, \"recall\": 0.5945945945945946, \"f1-score\": 0.6875, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7682220931301813, \"recall\": 0.7990107042526398, \"f1-score\": 0.7757295706744236, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7399958241743231, \"recall\": 0.7241379310344828, \"f1-score\": 0.7224166490203509, \"support\": 116.0}}", "eval_f1_macro": 68.95373961550433, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.24166490203508, "eval_loss": 1.2310043573379517, "eval_runtime": 2.426, "eval_samples_per_second": 47.816, "step": 1968 }, { "epoch": 124.0, "learning_rate": 0.00012666666666666666, "loss": 0.0023, "step": 1984 }, { "epoch": 124.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 71.59508898183678, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7931034482758621, \"recall\": 0.6216216216216216, \"f1-score\": 0.696969696969697, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7748571770662288, \"recall\": 0.7948846023442797, \"f1-score\": 0.7773406332229862, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7407856647415209, \"recall\": 0.7241379310344828, \"f1-score\": 0.7245582454507403, \"support\": 116.0}}", "eval_f1_macro": 69.09694517537655, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.45582454507404, "eval_loss": 1.1732732057571411, "eval_runtime": 2.2592, "eval_samples_per_second": 51.345, "step": 1984 }, { "epoch": 125.0, "learning_rate": 0.000125, "loss": 0.0014, "step": 2000 }, { "epoch": 125.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 71.59508898183678, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7931034482758621, \"recall\": 0.6216216216216216, \"f1-score\": 0.696969696969697, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7748571770662288, \"recall\": 0.7948846023442797, \"f1-score\": 0.7773406332229862, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7407856647415209, \"recall\": 0.7241379310344828, \"f1-score\": 0.7245582454507403, \"support\": 116.0}}", "eval_f1_macro": 69.09694517537655, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.45582454507404, "eval_loss": 1.2506325244903564, "eval_runtime": 2.2452, "eval_samples_per_second": 51.666, "step": 2000 }, { "epoch": 126.0, "learning_rate": 0.0001233333333333333, "loss": 0.0002, "step": 2016 }, { "epoch": 126.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 71.59508898183678, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7931034482758621, \"recall\": 0.6216216216216216, \"f1-score\": 0.696969696969697, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7748571770662288, \"recall\": 0.7948846023442797, \"f1-score\": 0.7773406332229862, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7407856647415209, \"recall\": 0.7241379310344828, \"f1-score\": 0.7245582454507403, \"support\": 116.0}}", "eval_f1_macro": 69.09694517537655, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.45582454507404, "eval_loss": 1.2512825727462769, "eval_runtime": 2.344, "eval_samples_per_second": 49.489, "step": 2016 }, { "epoch": 127.0, "learning_rate": 0.00012166666666666665, "loss": 0.001, "step": 2032 }, { "epoch": 127.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 70.6663288540288, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7666666666666667, \"recall\": 0.6216216216216216, \"f1-score\": 0.6865671641791045, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7685763888888889, \"recall\": 0.7792596023442797, \"f1-score\": 0.7653260309098764, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7307112068965518, \"recall\": 0.7155172413793104, \"f1-score\": 0.7153288661493075, \"support\": 116.0}}", "eval_f1_macro": 68.02898052532235, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.53288661493076, "eval_loss": 1.2251347303390503, "eval_runtime": 2.2231, "eval_samples_per_second": 52.179, "step": 2032 }, { "epoch": 128.0, "learning_rate": 0.00011999999999999999, "loss": 0.0001, "step": 2048 }, { "epoch": 128.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 72.28350002650896, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.6774193548387096, \"f1-score\": 0.7, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.7027027027027027, \"f1-score\": 0.7323943661971832, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7937443655623169, \"recall\": 0.7853624794148988, \"f1-score\": 0.7781209757286809, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7513712084120211, \"recall\": 0.7327586206896551, \"f1-score\": 0.7341596701444429, \"support\": 116.0}}", "eval_f1_macro": 69.16630895366053, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.41596701444429, "eval_loss": 1.1956950426101685, "eval_runtime": 2.2387, "eval_samples_per_second": 51.816, "step": 2048 }, { "epoch": 129.0, "learning_rate": 0.00011833333333333331, "loss": 0.0002, "step": 2064 }, { "epoch": 129.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 72.19392646661663, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.7297297297297297, \"f1-score\": 0.7397260273972601, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7871260683760684, \"recall\": 0.7847085997287611, \"f1-score\": 0.7743547708399907, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7522656940760389, \"recall\": 0.7327586206896551, \"f1-score\": 0.7339244654275855, \"support\": 116.0}}", "eval_f1_macro": 68.83153518577694, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.39244654275855, "eval_loss": 1.2062249183654785, "eval_runtime": 2.0692, "eval_samples_per_second": 56.061, "step": 2064 }, { "epoch": 130.0, "learning_rate": 0.00011666666666666665, "loss": 0.0016, "step": 2080 }, { "epoch": 130.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 69.63518777151155, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7142857142857143, \"recall\": 0.6451612903225806, \"f1-score\": 0.6779661016949152, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7931034482758621, \"recall\": 0.6216216216216216, \"f1-score\": 0.696969696969697, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6923076923076923, \"recall\": 1.0, \"f1-score\": 0.8181818181818181, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7402398846364364, \"recall\": 0.7850839751041363, \"f1-score\": 0.7503301782962799, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7245843732169535, \"recall\": 0.7068965517241379, \"f1-score\": 0.7046542489266042, \"support\": 116.0}}", "eval_f1_macro": 66.6960158485582, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.46542489266044, "eval_loss": 1.2418709993362427, "eval_runtime": 2.0305, "eval_samples_per_second": 57.129, "step": 2080 }, { "epoch": 131.0, "learning_rate": 0.000115, "loss": 0.0005, "step": 2096 }, { "epoch": 131.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 69.90597139032764, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7586206896551724, \"recall\": 0.5945945945945946, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7639156710022181, \"recall\": 0.7758812239659014, \"f1-score\": 0.7601502966778109, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7251200142895884, \"recall\": 0.7068965517241379, \"f1-score\": 0.7067565995623314, \"support\": 116.0}}", "eval_f1_macro": 67.56891526024985, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.67565995623313, "eval_loss": 1.2399200201034546, "eval_runtime": 2.1094, "eval_samples_per_second": 54.992, "step": 2096 }, { "epoch": 132.0, "learning_rate": 0.00011333333333333331, "loss": 0.0004, "step": 2112 }, { "epoch": 132.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 69.96898151137239, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5, \"recall\": 0.8333333333333334, \"f1-score\": 0.625, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.6451612903225806, \"f1-score\": 0.689655172413793, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6756756756756757, \"f1-score\": 0.6944444444444444, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7787848725348725, \"recall\": 0.7640629540831154, \"f1-score\": 0.759137703489884, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7316276475759235, \"recall\": 0.7068965517241379, \"f1-score\": 0.7101770872378342, \"support\": 116.0}}", "eval_f1_macro": 67.47890697687858, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 71.01770872378343, "eval_loss": 1.286320686340332, "eval_runtime": 2.359, "eval_samples_per_second": 49.174, "step": 2112 }, { "epoch": 133.0, "learning_rate": 0.00011166666666666667, "loss": 0.0001, "step": 2128 }, { "epoch": 133.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 69.96898151137239, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5, \"recall\": 0.8333333333333334, \"f1-score\": 0.625, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.6451612903225806, \"f1-score\": 0.689655172413793, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6756756756756757, \"f1-score\": 0.6944444444444444, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7787848725348725, \"recall\": 0.7640629540831154, \"f1-score\": 0.759137703489884, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7316276475759235, \"recall\": 0.7068965517241379, \"f1-score\": 0.7101770872378342, \"support\": 116.0}}", "eval_f1_macro": 67.47890697687858, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 71.01770872378343, "eval_loss": 1.2882046699523926, "eval_runtime": 2.2026, "eval_samples_per_second": 52.665, "step": 2128 }, { "epoch": 134.0, "learning_rate": 0.00010999999999999998, "loss": 0.0041, "step": 2144 }, { "epoch": 134.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 70.70045157638818, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.6451612903225806, \"f1-score\": 0.689655172413793, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.7027027027027027, \"f1-score\": 0.7123287671232876, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7830664097111466, \"recall\": 0.7674413324614937, \"f1-score\": 0.763893405115062, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7368814429839838, \"recall\": 0.7155172413793104, \"f1-score\": 0.7179672201946288, \"support\": 116.0}}", "eval_f1_macro": 67.90163601022773, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.79672201946288, "eval_loss": 1.2369537353515625, "eval_runtime": 2.2854, "eval_samples_per_second": 50.757, "step": 2144 }, { "epoch": 135.0, "learning_rate": 0.00010833333333333333, "loss": 0.0003, "step": 2160 }, { "epoch": 135.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 72.28635318822373, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.7027027027027027, \"f1-score\": 0.7222222222222223, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7942460317460318, \"recall\": 0.7853624794148988, \"f1-score\": 0.7783325085792683, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7513136288998358, \"recall\": 0.7327586206896551, \"f1-score\": 0.7340857674125114, \"support\": 116.0}}", "eval_f1_macro": 69.18511187371273, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.40857674125114, "eval_loss": 1.1938015222549438, "eval_runtime": 2.5226, "eval_samples_per_second": 45.985, "step": 2160 }, { "epoch": 136.0, "learning_rate": 0.00010666666666666667, "loss": 0.0002, "step": 2176 }, { "epoch": 136.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 71.45726817405861, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.7096774193548387, \"f1-score\": 0.7096774193548389, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6756756756756757, \"f1-score\": 0.7142857142857142, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7872576572173346, \"recall\": 0.7721274702121477, \"f1-score\": 0.7693289764210067, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7428814002089864, \"recall\": 0.7241379310344828, \"f1-score\": 0.7261668858524842, \"support\": 116.0}}", "eval_f1_macro": 68.3847979040895, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.61668858524843, "eval_loss": 1.1832021474838257, "eval_runtime": 2.498, "eval_samples_per_second": 46.438, "step": 2176 }, { "epoch": 137.0, "learning_rate": 0.00010499999999999999, "loss": 0.0023, "step": 2192 }, { "epoch": 137.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 72.24180724602901, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.7096774193548387, \"f1-score\": 0.7096774193548389, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7741935483870968, \"recall\": 0.6486486486486487, \"f1-score\": 0.7058823529411764, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7777632827324479, \"recall\": 0.7965268696115471, \"f1-score\": 0.7793320243617616, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7432814783310432, \"recall\": 0.7327586206896551, \"f1-score\": 0.7314154712513956, \"support\": 116.0}}", "eval_f1_macro": 69.27395772104548, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.14154712513957, "eval_loss": 1.1572390794754028, "eval_runtime": 2.3161, "eval_samples_per_second": 50.085, "step": 2192 }, { "epoch": 138.0, "learning_rate": 0.00010333333333333333, "loss": 0.0015, "step": 2208 }, { "epoch": 138.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 73.31427426365428, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.625, \"recall\": 0.8333333333333334, \"f1-score\": 0.7142857142857143, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.7096774193548387, \"f1-score\": 0.7096774193548389, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7741935483870968, \"recall\": 0.6486486486486487, \"f1-score\": 0.7058823529411764, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"request\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7896505376344086, \"recall\": 0.8138879807226581, \"f1-score\": 0.7968241732176742, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7512513904338153, \"recall\": 0.7413793103448276, \"f1-score\": 0.7415241958852505, \"support\": 116.0}}", "eval_f1_macro": 70.8288153971266, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.15241958852505, "eval_loss": 1.127516746520996, "eval_runtime": 2.26, "eval_samples_per_second": 51.328, "step": 2208 }, { "epoch": 139.0, "learning_rate": 0.00010166666666666667, "loss": 0.0002, "step": 2224 }, { "epoch": 139.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 71.14146434132823, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.8095238095238095, \"recall\": 0.5483870967741935, \"f1-score\": 0.6538461538461537, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7567567567567568, \"recall\": 0.7567567567567568, \"f1-score\": 0.7567567567567567, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 0.7142857142857143, \"recall\": 0.625, \"f1-score\": 0.6666666666666666, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7503485628485629, \"recall\": 0.7898790928024799, \"f1-score\": 0.7605985467827573, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7411740558292282, \"recall\": 0.7241379310344828, \"f1-score\": 0.7212951144439347, \"support\": 116.0}}", "eval_f1_macro": 67.60875971402288, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.12951144439347, "eval_loss": 1.218409538269043, "eval_runtime": 2.3805, "eval_samples_per_second": 48.73, "step": 2224 }, { "epoch": 140.0, "learning_rate": 9.999999999999999e-05, "loss": 0.0007, "step": 2240 }, { "epoch": 140.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 71.34715800657082, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7105263157894737, \"recall\": 0.7297297297297297, \"f1-score\": 0.7200000000000001, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7881529106858054, \"recall\": 0.7609630800154994, \"f1-score\": 0.7636808914547252, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7449958981832848, \"recall\": 0.7241379310344828, \"f1-score\": 0.7267829991230006, \"support\": 116.0}}", "eval_f1_macro": 67.88274590708669, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.67829991230006, "eval_loss": 1.2071099281311035, "eval_runtime": 2.2973, "eval_samples_per_second": 50.493, "step": 2240 }, { "epoch": 141.0, "learning_rate": 9.833333333333333e-05, "loss": 0.0001, "step": 2256 }, { "epoch": 141.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 69.61180365216576, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.7096774193548387, \"f1-score\": 0.7096774193548389, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6756756756756757, \"f1-score\": 0.6944444444444444, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7801102906949682, \"recall\": 0.7426135813232587, \"f1-score\": 0.7458583702298124, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7279957829526795, \"recall\": 0.7068965517241379, \"f1-score\": 0.7076938246562994, \"support\": 116.0}}", "eval_f1_macro": 66.29852179820554, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.76938246562993, "eval_loss": 1.2121734619140625, "eval_runtime": 2.2118, "eval_samples_per_second": 52.447, "step": 2256 }, { "epoch": 142.0, "learning_rate": 9.666666666666667e-05, "loss": 0.0008, "step": 2272 }, { "epoch": 142.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 73.08583159191204, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.7297297297297297, \"f1-score\": 0.7397260273972601, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.798931623931624, \"recall\": 0.7985974886176499, \"f1-score\": 0.7865666731447722, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7595932802829355, \"recall\": 0.7413793103448276, \"f1-score\": 0.7415042668581396, \"support\": 116.0}}", "eval_f1_macro": 69.91703761286865, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.15042668581397, "eval_loss": 1.2578798532485962, "eval_runtime": 2.5066, "eval_samples_per_second": 46.278, "step": 2272 }, { "epoch": 143.0, "learning_rate": 9.499999999999999e-05, "loss": 0.0003, "step": 2288 }, { "epoch": 143.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 73.08094197020597, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7714285714285715, \"recall\": 0.7297297297297297, \"f1-score\": 0.75, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7979552245999615, \"recall\": 0.7985974886176499, \"f1-score\": 0.785162747677104, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7634034522646138, \"recall\": 0.7413793103448276, \"f1-score\": 0.7425566157389356, \"support\": 116.0}}", "eval_f1_macro": 69.79224423796481, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.25566157389356, "eval_loss": 1.2859212160110474, "eval_runtime": 2.1106, "eval_samples_per_second": 54.96, "step": 2288 }, { "epoch": 144.0, "learning_rate": 9.333333333333333e-05, "loss": 0.0004, "step": 2304 }, { "epoch": 144.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 72.21325519082055, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.7027027027027027, \"f1-score\": 0.7323943661971832, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7762815551321743, \"recall\": 0.7952191102392715, \"f1-score\": 0.7760939115836201, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7497648933927892, \"recall\": 0.7327586206896551, \"f1-score\": 0.7331517115125162, \"support\": 116.0}}", "eval_f1_macro": 68.98612547409955, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.31517115125162, "eval_loss": 1.2795659303665161, "eval_runtime": 2.3108, "eval_samples_per_second": 50.198, "step": 2304 }, { "epoch": 145.0, "learning_rate": 9.166666666666667e-05, "loss": 0.0001, "step": 2320 }, { "epoch": 145.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 72.97136593324288, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.7027027027027027, \"f1-score\": 0.7323943661971832, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7810049019607843, \"recall\": 0.7992513683037876, \"f1-score\": 0.7815800267603271, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7550738111336488, \"recall\": 0.7413793103448276, \"f1-score\": 0.7413582150753241, \"support\": 116.0}}", "eval_f1_macro": 69.47378015647352, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.13582150753241, "eval_loss": 1.259643793106079, "eval_runtime": 2.2374, "eval_samples_per_second": 51.845, "step": 2320 }, { "epoch": 146.0, "learning_rate": 8.999999999999999e-05, "loss": 0.0143, "step": 2336 }, { "epoch": 146.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 71.74999381388027, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.7096774193548387, \"f1-score\": 0.6875, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7666666666666667, \"recall\": 0.6216216216216216, \"f1-score\": 0.6865671641791045, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.786922268907563, \"recall\": 0.7948846023442797, \"f1-score\": 0.7840857402533838, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7387870665507582, \"recall\": 0.7241379310344828, \"f1-score\": 0.7247587880387926, \"support\": 116.0}}", "eval_f1_macro": 69.69651024474524, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.47587880387925, "eval_loss": 1.2264900207519531, "eval_runtime": 2.4493, "eval_samples_per_second": 47.361, "step": 2336 }, { "epoch": 147.0, "learning_rate": 8.833333333333333e-05, "loss": 0.0053, "step": 2352 }, { "epoch": 147.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 71.42834861077034, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7741935483870968, \"recall\": 0.6486486486486487, \"f1-score\": 0.7058823529411764, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7736022164242041, \"recall\": 0.7826379807226581, \"f1-score\": 0.7706139927235263, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7364926852275949, \"recall\": 0.7241379310344828, \"f1-score\": 0.7238678666076022, \"support\": 116.0}}", "eval_f1_macro": 68.49902157542456, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.38678666076022, "eval_loss": 1.2261444330215454, "eval_runtime": 2.3209, "eval_samples_per_second": 49.981, "step": 2352 }, { "epoch": 148.0, "learning_rate": 8.666666666666665e-05, "loss": 0.0014, "step": 2368 }, { "epoch": 148.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 72.94342125396005, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7857142857142857, \"recall\": 0.7096774193548387, \"f1-score\": 0.7457627118644068, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6756756756756757, \"f1-score\": 0.7142857142857142, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7853800611153553, \"recall\": 0.7999052479899253, \"f1-score\": 0.780938224545764, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.756981946383569, \"recall\": 0.7413793103448276, \"f1-score\": 0.7408109187614014, \"support\": 116.0}}", "eval_f1_macro": 69.41673107073457, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.08109187614014, "eval_loss": 1.25702702999115, "eval_runtime": 2.2587, "eval_samples_per_second": 51.357, "step": 2368 }, { "epoch": 149.0, "learning_rate": 8.499999999999999e-05, "loss": 0.0008, "step": 2384 }, { "epoch": 149.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 70.75520071430205, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.625, \"recall\": 0.8333333333333334, \"f1-score\": 0.7142857142857143, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6470588235294118, \"recall\": 0.7096774193548387, \"f1-score\": 0.6769230769230768, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6486486486486487, \"f1-score\": 0.6956521739130435, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7892156862745098, \"recall\": 0.7687490918337692, \"f1-score\": 0.768426032620406, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7321881338742394, \"recall\": 0.7155172413793104, \"f1-score\": 0.7161281834842113, \"support\": 116.0}}", "eval_f1_macro": 68.30453623292497, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.61281834842114, "eval_loss": 1.2749736309051514, "eval_runtime": 2.441, "eval_samples_per_second": 47.522, "step": 2384 }, { "epoch": 150.0, "learning_rate": 8.333333333333333e-05, "loss": 0.0153, "step": 2400 }, { "epoch": 150.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 73.04663385761674, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6756756756756757, \"f1-score\": 0.7246376811594203, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7713840744631186, \"recall\": 0.8114979899254092, \"f1-score\": 0.7849103643666888, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7537355818764288, \"recall\": 0.7413793103448276, \"f1-score\": 0.7414086319557359, \"support\": 116.0}}", "eval_f1_macro": 69.7698101659279, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.14086319557359, "eval_loss": 1.1729803085327148, "eval_runtime": 2.3845, "eval_samples_per_second": 48.648, "step": 2400 }, { "epoch": 151.0, "learning_rate": 8.166666666666665e-05, "loss": 0.0002, "step": 2416 }, { "epoch": 151.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 72.2925741490733, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6756756756756757, \"f1-score\": 0.7142857142857142, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7673564180917123, \"recall\": 0.8074657318608931, \"f1-score\": 0.7808184253737793, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7439002005432026, \"recall\": 0.7327586206896551, \"f1-score\": 0.7321249020291514, \"support\": 116.0}}", "eval_f1_macro": 69.40608225544706, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.21249020291513, "eval_loss": 1.2129653692245483, "eval_runtime": 2.2652, "eval_samples_per_second": 51.211, "step": 2416 }, { "epoch": 152.0, "learning_rate": 7.999999999999999e-05, "loss": 0.0016, "step": 2432 }, { "epoch": 152.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 72.41728534320436, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.8214285714285714, \"recall\": 0.6216216216216216, \"f1-score\": 0.7076923076923075, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7797867063492063, \"recall\": 0.8087734912331686, \"f1-score\": 0.7859909010837803, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7506824712643678, \"recall\": 0.7327586206896551, \"f1-score\": 0.7325155936077256, \"support\": 116.0}}", "eval_f1_macro": 69.86585787411381, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.25155936077256, "eval_loss": 1.2629878520965576, "eval_runtime": 2.2761, "eval_samples_per_second": 50.964, "step": 2432 }, { "epoch": 153.0, "learning_rate": 7.833333333333333e-05, "loss": 0.0011, "step": 2448 }, { "epoch": 153.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 70.92028682644413, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7857142857142857, \"recall\": 0.5945945945945946, \"f1-score\": 0.676923076923077, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7759829260651628, \"recall\": 0.7932423350770125, \"f1-score\": 0.7748835617787231, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7387906728026963, \"recall\": 0.7155172413793104, \"f1-score\": 0.7169916020513906, \"support\": 116.0}}", "eval_f1_macro": 68.87853882477538, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.69916020513907, "eval_loss": 1.2745832204818726, "eval_runtime": 2.4207, "eval_samples_per_second": 47.919, "step": 2448 }, { "epoch": 154.0, "learning_rate": 7.666666666666666e-05, "loss": 0.012, "step": 2464 }, { "epoch": 154.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 72.41728534320436, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.8214285714285714, \"recall\": 0.6216216216216216, \"f1-score\": 0.7076923076923075, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7797867063492063, \"recall\": 0.8087734912331686, \"f1-score\": 0.7859909010837803, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7506824712643678, \"recall\": 0.7327586206896551, \"f1-score\": 0.7325155936077256, \"support\": 116.0}}", "eval_f1_macro": 69.86585787411381, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.25155936077256, "eval_loss": 1.2336031198501587, "eval_runtime": 2.4363, "eval_samples_per_second": 47.613, "step": 2464 }, { "epoch": 155.0, "learning_rate": 7.5e-05, "loss": 0.0001, "step": 2480 }, { "epoch": 155.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 73.16539772739387, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.7096774193548387, \"f1-score\": 0.7096774193548389, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.8275862068965517, \"recall\": 0.6486486486486487, \"f1-score\": 0.7272727272727273, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7833285882020588, \"recall\": 0.8121518696115471, \"f1-score\": 0.7898465436491005, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7585732687843271, \"recall\": 0.7413793103448276, \"f1-score\": 0.7417714718291212, \"support\": 116.0}}", "eval_f1_macro": 70.20858165769782, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.17714718291212, "eval_loss": 1.221439003944397, "eval_runtime": 2.3711, "eval_samples_per_second": 48.922, "step": 2480 }, { "epoch": 156.0, "learning_rate": 7.333333333333332e-05, "loss": 0.002, "step": 2496 }, { "epoch": 156.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 73.1455226720313, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.7027027027027027, \"f1-score\": 0.7323943661971832, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7829127163685987, \"recall\": 0.8108441102392715, \"f1-score\": 0.7894963693409165, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7544317329053637, \"recall\": 0.7413793103448276, \"f1-score\": 0.7412877356663379, \"support\": 116.0}}", "eval_f1_macro": 70.17745505252591, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.12877356663378, "eval_loss": 1.2159744501113892, "eval_runtime": 2.149, "eval_samples_per_second": 53.978, "step": 2496 }, { "epoch": 157.0, "learning_rate": 7.166666666666667e-05, "loss": 0.0002, "step": 2512 }, { "epoch": 157.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 72.30710211656402, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.7027027027027027, \"f1-score\": 0.7323943661971832, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.775469866820486, \"recall\": 0.7969552213503827, \"f1-score\": 0.7790701020598105, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7490558323389006, \"recall\": 0.7327586206896551, \"f1-score\": 0.7342600858967526, \"support\": 116.0}}", "eval_f1_macro": 69.25067573864982, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.42600858967526, "eval_loss": 1.2059881687164307, "eval_runtime": 2.2612, "eval_samples_per_second": 51.3, "step": 2512 }, { "epoch": 158.0, "learning_rate": 7e-05, "loss": 0.0013, "step": 2528 }, { "epoch": 158.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 73.05038784859155, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6756756756756757, \"f1-score\": 0.7246376811594203, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7758218344155844, \"recall\": 0.8114979899254092, \"f1-score\": 0.7854036742133135, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7556448879185949, \"recall\": 0.7413793103448276, \"f1-score\": 0.7411202939532842, \"support\": 116.0}}", "eval_f1_macro": 69.81365993007232, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.11202939532842, "eval_loss": 1.2273198366165161, "eval_runtime": 2.2819, "eval_samples_per_second": 50.836, "step": 2528 }, { "epoch": 159.0, "learning_rate": 6.833333333333333e-05, "loss": 0.0006, "step": 2544 }, { "epoch": 159.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 72.30265892598246, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.6774193548387096, \"f1-score\": 0.7, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.8, \"recall\": 0.6486486486486487, \"f1-score\": 0.7164179104477612, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7714606035726725, \"recall\": 0.8081196115470308, \"f1-score\": 0.7813589614950458, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.747290704737197, \"recall\": 0.7327586206896551, \"f1-score\": 0.7320478165532811, \"support\": 116.0}}", "eval_f1_macro": 69.45412991067073, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.2047816553281, "eval_loss": 1.2536399364471436, "eval_runtime": 2.1983, "eval_samples_per_second": 52.768, "step": 2544 }, { "epoch": 160.0, "learning_rate": 6.666666666666666e-05, "loss": 0.0001, "step": 2560 }, { "epoch": 160.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 71.21555209201495, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6756756756756757, \"f1-score\": 0.7142857142857142, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 1.0, \"f1-score\": 0.8571428571428571, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7539635609488551, \"recall\": 0.7918407318608931, \"f1-score\": 0.7632119090830525, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7357926471770285, \"recall\": 0.7241379310344828, \"f1-score\": 0.7219356357600302, \"support\": 116.0}}", "eval_f1_macro": 67.84105858516021, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.19356357600302, "eval_loss": 1.2445656061172485, "eval_runtime": 2.2477, "eval_samples_per_second": 51.608, "step": 2560 }, { "epoch": 161.0, "learning_rate": 6.5e-05, "loss": 0.0005, "step": 2576 }, { "epoch": 161.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 72.0886472797046, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.7027027027027027, \"f1-score\": 0.7323943661971832, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 1.0, \"f1-score\": 0.8571428571428571, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7756881598793364, \"recall\": 0.7952191102392715, \"f1-score\": 0.7723436224401179, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7495611639881417, \"recall\": 0.7327586206896551, \"f1-score\": 0.7315009854176577, \"support\": 116.0}}", "eval_f1_macro": 68.65276643912159, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.15009854176577, "eval_loss": 1.2382774353027344, "eval_runtime": 2.1847, "eval_samples_per_second": 53.096, "step": 2576 }, { "epoch": 162.0, "learning_rate": 6.333333333333333e-05, "loss": 0.0002, "step": 2592 }, { "epoch": 162.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 71.21555209201495, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6756756756756757, \"f1-score\": 0.7142857142857142, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 1.0, \"f1-score\": 0.8571428571428571, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7539635609488551, \"recall\": 0.7918407318608931, \"f1-score\": 0.7632119090830525, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7357926471770285, \"recall\": 0.7241379310344828, \"f1-score\": 0.7219356357600302, \"support\": 116.0}}", "eval_f1_macro": 67.84105858516021, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.19356357600302, "eval_loss": 1.2521488666534424, "eval_runtime": 2.2733, "eval_samples_per_second": 51.028, "step": 2592 }, { "epoch": 163.0, "learning_rate": 6.166666666666666e-05, "loss": 0.0001, "step": 2608 }, { "epoch": 163.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 71.31191410231975, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.7027027027027027, \"f1-score\": 0.7323943661971832, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 1.0, \"f1-score\": 0.8571428571428571, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7745770096776289, \"recall\": 0.7813302213503827, \"f1-score\": 0.7639457527249349, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7472701180531863, \"recall\": 0.7241379310344828, \"f1-score\": 0.7251378107127714, \"support\": 116.0}}", "eval_f1_macro": 67.90628913110532, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.51378107127714, "eval_loss": 1.268195629119873, "eval_runtime": 2.3211, "eval_samples_per_second": 49.975, "step": 2608 }, { "epoch": 164.0, "learning_rate": 5.9999999999999995e-05, "loss": 0.0002, "step": 2624 }, { "epoch": 164.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 70.9771975436676, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5, \"recall\": 0.8333333333333334, \"f1-score\": 0.625, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6896551724137931, \"recall\": 0.6451612903225806, \"f1-score\": 0.6666666666666667, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7741935483870968, \"recall\": 0.6486486486486487, \"f1-score\": 0.7058823529411764, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7812765446455657, \"recall\": 0.790198464593626, \"f1-score\": 0.7759928313430635, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7379460720898525, \"recall\": 0.7155172413793104, \"f1-score\": 0.7182820133498045, \"support\": 116.0}}", "eval_f1_macro": 68.97714056382786, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.82820133498046, "eval_loss": 1.2899616956710815, "eval_runtime": 2.1951, "eval_samples_per_second": 52.844, "step": 2624 }, { "epoch": 165.0, "learning_rate": 5.8333333333333326e-05, "loss": 0.0004, "step": 2640 }, { "epoch": 165.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 70.83086708646859, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5, \"recall\": 0.8333333333333334, \"f1-score\": 0.625, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7931034482758621, \"recall\": 0.6216216216216216, \"f1-score\": 0.696969696969697, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7670762427227944, \"recall\": 0.7908523442797637, \"f1-score\": 0.77046813264093, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7368900273329524, \"recall\": 0.7155172413793104, \"f1-score\": 0.7173396383526295, \"support\": 116.0}}", "eval_f1_macro": 68.48605623474933, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.73396383526295, "eval_loss": 1.30678391456604, "eval_runtime": 2.3185, "eval_samples_per_second": 50.032, "step": 2640 }, { "epoch": 166.0, "learning_rate": 5.666666666666666e-05, "loss": 0.0005, "step": 2656 }, { "epoch": 166.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 70.83086708646859, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5, \"recall\": 0.8333333333333334, \"f1-score\": 0.625, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7931034482758621, \"recall\": 0.6216216216216216, \"f1-score\": 0.696969696969697, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7670762427227944, \"recall\": 0.7908523442797637, \"f1-score\": 0.77046813264093, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7368900273329524, \"recall\": 0.7155172413793104, \"f1-score\": 0.7173396383526295, \"support\": 116.0}}", "eval_f1_macro": 68.48605623474933, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.73396383526295, "eval_loss": 1.3133952617645264, "eval_runtime": 2.377, "eval_samples_per_second": 48.802, "step": 2656 }, { "epoch": 167.0, "learning_rate": 5.499999999999999e-05, "loss": 0.0002, "step": 2672 }, { "epoch": 167.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 69.96647365085389, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.47619047619047616, \"recall\": 0.8333333333333334, \"f1-score\": 0.6060606060606061, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.76, \"recall\": 0.6129032258064516, \"f1-score\": 0.6785714285714285, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6486486486486487, \"f1-score\": 0.6956521739130435, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7559848484848485, \"recall\": 0.7861662065291097, \"f1-score\": 0.7607708201857818, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7303649798477384, \"recall\": 0.7068965517241379, \"f1-score\": 0.7086251135318513, \"support\": 116.0}}", "eval_f1_macro": 67.62407290540283, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.86251135318513, "eval_loss": 1.300965666770935, "eval_runtime": 2.2349, "eval_samples_per_second": 51.904, "step": 2672 }, { "epoch": 168.0, "learning_rate": 5.333333333333333e-05, "loss": 0.0001, "step": 2688 }, { "epoch": 168.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 72.3130865910927, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.7027027027027027, \"f1-score\": 0.7222222222222223, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7763937451437452, \"recall\": 0.7969552213503827, \"f1-score\": 0.7804867561059512, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.745111641232331, \"recall\": 0.7327586206896551, \"f1-score\": 0.7332402168368859, \"support\": 116.0}}", "eval_f1_macro": 69.37660054275122, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.32402168368858, "eval_loss": 1.237136721611023, "eval_runtime": 2.2752, "eval_samples_per_second": 50.985, "step": 2688 }, { "epoch": 169.0, "learning_rate": 5.1666666666666664e-05, "loss": 0.0, "step": 2704 }, { "epoch": 169.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 73.07915281690298, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.7027027027027027, \"f1-score\": 0.7222222222222223, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7815470885323826, \"recall\": 0.8009874794148988, \"f1-score\": 0.7861582624580383, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7507764181952823, \"recall\": 0.7413793103448276, \"f1-score\": 0.7416001475793187, \"support\": 116.0}}", "eval_f1_macro": 69.88073444071452, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.16001475793188, "eval_loss": 1.2186721563339233, "eval_runtime": 2.4159, "eval_samples_per_second": 48.016, "step": 2704 }, { "epoch": 170.0, "learning_rate": 4.9999999999999996e-05, "loss": 0.0001, "step": 2720 }, { "epoch": 170.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 73.07915281690298, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.7027027027027027, \"f1-score\": 0.7222222222222223, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7815470885323826, \"recall\": 0.8009874794148988, \"f1-score\": 0.7861582624580383, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7507764181952823, \"recall\": 0.7413793103448276, \"f1-score\": 0.7416001475793187, \"support\": 116.0}}", "eval_f1_macro": 69.88073444071452, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.16001475793188, "eval_loss": 1.214310646057129, "eval_runtime": 2.3883, "eval_samples_per_second": 48.57, "step": 2720 }, { "epoch": 171.0, "learning_rate": 4.8333333333333334e-05, "loss": 0.0001, "step": 2736 }, { "epoch": 171.0, "eval_accuracy": 75.0, "eval_average_metrics": 73.91158889784157, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.7027027027027027, \"f1-score\": 0.7323943661971832, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.75, \"macro avg\": {\"precision\": 0.7880660597572362, \"recall\": 0.8148763683037876, \"f1-score\": 0.7951678756930036, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7600965098683151, \"recall\": 0.75, \"f1-score\": 0.7496476664087707, \"support\": 116.0}}", "eval_f1_macro": 70.68158895048921, "eval_f1_micro": 75.0, "eval_f1_weighted": 74.96476664087707, "eval_loss": 1.2283321619033813, "eval_runtime": 2.3482, "eval_samples_per_second": 49.4, "step": 2736 }, { "epoch": 172.0, "learning_rate": 4.6666666666666665e-05, "loss": 0.0118, "step": 2752 }, { "epoch": 172.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 73.06239436585838, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6756756756756757, \"f1-score\": 0.7142857142857142, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7769475214328155, \"recall\": 0.8114979899254092, \"f1-score\": 0.7869832415724911, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.75147428354832, \"recall\": 0.7413793103448276, \"f1-score\": 0.7401964947691324, \"support\": 116.0}}", "eval_f1_macro": 69.95406591755476, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.01964947691323, "eval_loss": 1.2305833101272583, "eval_runtime": 2.3308, "eval_samples_per_second": 49.768, "step": 2752 }, { "epoch": 173.0, "learning_rate": 4.4999999999999996e-05, "loss": 0.0001, "step": 2768 }, { "epoch": 173.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 73.06239436585838, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6756756756756757, \"f1-score\": 0.7142857142857142, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7769475214328155, \"recall\": 0.8114979899254092, \"f1-score\": 0.7869832415724911, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.75147428354832, \"recall\": 0.7413793103448276, \"f1-score\": 0.7401964947691324, \"support\": 116.0}}", "eval_f1_macro": 69.95406591755476, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.01964947691323, "eval_loss": 1.232358694076538, "eval_runtime": 2.353, "eval_samples_per_second": 49.298, "step": 2768 }, { "epoch": 174.0, "learning_rate": 4.333333333333333e-05, "loss": 0.0003, "step": 2784 }, { "epoch": 174.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 73.06239436585838, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6756756756756757, \"f1-score\": 0.7142857142857142, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7769475214328155, \"recall\": 0.8114979899254092, \"f1-score\": 0.7869832415724911, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.75147428354832, \"recall\": 0.7413793103448276, \"f1-score\": 0.7401964947691324, \"support\": 116.0}}", "eval_f1_macro": 69.95406591755476, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.01964947691323, "eval_loss": 1.2332383394241333, "eval_runtime": 2.4493, "eval_samples_per_second": 47.361, "step": 2784 }, { "epoch": 175.0, "learning_rate": 4.1666666666666665e-05, "loss": 0.0001, "step": 2800 }, { "epoch": 175.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 73.06239436585838, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6756756756756757, \"f1-score\": 0.7142857142857142, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7769475214328155, \"recall\": 0.8114979899254092, \"f1-score\": 0.7869832415724911, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.75147428354832, \"recall\": 0.7413793103448276, \"f1-score\": 0.7401964947691324, \"support\": 116.0}}", "eval_f1_macro": 69.95406591755476, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.01964947691323, "eval_loss": 1.2342748641967773, "eval_runtime": 2.2579, "eval_samples_per_second": 51.375, "step": 2800 }, { "epoch": 176.0, "learning_rate": 3.9999999999999996e-05, "loss": 0.0001, "step": 2816 }, { "epoch": 176.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 72.31492066595455, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6486486486486487, \"f1-score\": 0.6956521739130435, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7725283295136236, \"recall\": 0.8081196115470308, \"f1-score\": 0.7831198584940545, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7416345064671637, \"recall\": 0.7327586206896551, \"f1-score\": 0.7309730443752678, \"support\": 116.0}}", "eval_f1_macro": 69.6106540883604, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.09730443752677, "eval_loss": 1.210294485092163, "eval_runtime": 2.3919, "eval_samples_per_second": 48.497, "step": 2816 }, { "epoch": 177.0, "learning_rate": 3.833333333333333e-05, "loss": 0.0002, "step": 2832 }, { "epoch": 177.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 72.31269149223965, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.6774193548387096, \"f1-score\": 0.7, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7741935483870968, \"recall\": 0.6486486486486487, \"f1-score\": 0.7058823529411764, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7723197644413211, \"recall\": 0.8081196115470308, \"f1-score\": 0.7829155800251135, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7424399991601941, \"recall\": 0.7327586206896551, \"f1-score\": 0.7310654582879524, \"support\": 116.0}}", "eval_f1_macro": 69.59249600223232, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.10654582879525, "eval_loss": 1.207425832748413, "eval_runtime": 2.1302, "eval_samples_per_second": 54.455, "step": 2832 }, { "epoch": 178.0, "learning_rate": 3.666666666666666e-05, "loss": 0.0, "step": 2848 }, { "epoch": 178.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 72.31269149223965, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.6774193548387096, \"f1-score\": 0.7, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7741935483870968, \"recall\": 0.6486486486486487, \"f1-score\": 0.7058823529411764, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7723197644413211, \"recall\": 0.8081196115470308, \"f1-score\": 0.7829155800251135, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7424399991601941, \"recall\": 0.7327586206896551, \"f1-score\": 0.7310654582879524, \"support\": 116.0}}", "eval_f1_macro": 69.59249600223232, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.10654582879525, "eval_loss": 1.2198625802993774, "eval_runtime": 2.2148, "eval_samples_per_second": 52.376, "step": 2848 }, { "epoch": 179.0, "learning_rate": 3.5e-05, "loss": 0.0002, "step": 2864 }, { "epoch": 179.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 73.05760487940339, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6756756756756757, \"f1-score\": 0.7246376811594203, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7764345795136236, \"recall\": 0.8114979899254092, \"f1-score\": 0.7867430468998516, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7516021788809568, \"recall\": 0.7413793103448276, \"f1-score\": 0.7402184216866121, \"support\": 116.0}}", "eval_f1_macro": 69.93271527998681, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.02184216866121, "eval_loss": 1.2318569421768188, "eval_runtime": 2.3945, "eval_samples_per_second": 48.445, "step": 2864 }, { "epoch": 180.0, "learning_rate": 3.333333333333333e-05, "loss": 0.0001, "step": 2880 }, { "epoch": 180.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 72.9164008835973, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.7027027027027027, \"f1-score\": 0.7323943661971832, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 1.0, \"f1-score\": 0.8571428571428571, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7871732026143791, \"recall\": 0.7992513683037876, \"f1-score\": 0.780043526358128, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7583107955826008, \"recall\": 0.7413793103448276, \"f1-score\": 0.7405253912247894, \"support\": 116.0}}", "eval_f1_macro": 69.33720234294472, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.05253912247895, "eval_loss": 1.238344669342041, "eval_runtime": 2.1836, "eval_samples_per_second": 53.123, "step": 2880 }, { "epoch": 181.0, "learning_rate": 3.1666666666666666e-05, "loss": 0.0, "step": 2896 }, { "epoch": 181.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 71.54716438209535, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.6774193548387096, \"f1-score\": 0.7, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7931034482758621, \"recall\": 0.6216216216216216, \"f1-score\": 0.696969696969697, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7669435638469213, \"recall\": 0.8047412331686525, \"f1-score\": 0.776239762767277, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7420661391615456, \"recall\": 0.7241379310344828, \"f1-score\": 0.723619812977269, \"support\": 116.0}}", "eval_f1_macro": 68.99909002375796, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.3619812977269, "eval_loss": 1.264186978340149, "eval_runtime": 2.4029, "eval_samples_per_second": 48.275, "step": 2896 }, { "epoch": 182.0, "learning_rate": 2.9999999999999997e-05, "loss": 0.0015, "step": 2912 }, { "epoch": 182.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 70.66688997087125, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7931034482758621, \"recall\": 0.6216216216216216, \"f1-score\": 0.696969696969697, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 1.0, \"f1-score\": 0.8571428571428571, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7732607380520267, \"recall\": 0.7891162331686525, \"f1-score\": 0.7656020398344721, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7401777332749233, \"recall\": 0.7155172413793104, \"f1-score\": 0.7151059695566986, \"support\": 116.0}}", "eval_f1_macro": 68.05351465195308, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.51059695566985, "eval_loss": 1.265453815460205, "eval_runtime": 2.3141, "eval_samples_per_second": 50.127, "step": 2912 }, { "epoch": 183.0, "learning_rate": 2.833333333333333e-05, "loss": 0.0057, "step": 2928 }, { "epoch": 183.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 70.757218513014, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7666666666666667, \"recall\": 0.6216216216216216, \"f1-score\": 0.6865671641791045, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7784788676236045, \"recall\": 0.7891162331686525, \"f1-score\": 0.7696588660927909, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7370352802067865, \"recall\": 0.7155172413793104, \"f1-score\": 0.7151130434572363, \"support\": 116.0}}", "eval_f1_macro": 68.4141214304703, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.51130434572363, "eval_loss": 1.2580338716506958, "eval_runtime": 2.4483, "eval_samples_per_second": 47.38, "step": 2928 }, { "epoch": 184.0, "learning_rate": 2.6666666666666667e-05, "loss": 0.0017, "step": 2944 }, { "epoch": 184.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 70.757218513014, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7666666666666667, \"recall\": 0.6216216216216216, \"f1-score\": 0.6865671641791045, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7784788676236045, \"recall\": 0.7891162331686525, \"f1-score\": 0.7696588660927909, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7370352802067865, \"recall\": 0.7155172413793104, \"f1-score\": 0.7151130434572363, \"support\": 116.0}}", "eval_f1_macro": 68.4141214304703, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.51130434572363, "eval_loss": 1.2513511180877686, "eval_runtime": 2.2687, "eval_samples_per_second": 51.13, "step": 2944 }, { "epoch": 185.0, "learning_rate": 2.4999999999999998e-05, "loss": 0.0001, "step": 2960 }, { "epoch": 185.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 71.50327381251103, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5, \"recall\": 0.8333333333333334, \"f1-score\": 0.625, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6756756756756757, \"f1-score\": 0.7142857142857142, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7827068764568765, \"recall\": 0.7918407318608931, \"f1-score\": 0.7722572480409168, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7499145968973555, \"recall\": 0.7241379310344828, \"f1-score\": 0.7254042032839941, \"support\": 116.0}}", "eval_f1_macro": 68.64508871474816, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.5404203283994, "eval_loss": 1.290483832359314, "eval_runtime": 2.2629, "eval_samples_per_second": 51.262, "step": 2960 }, { "epoch": 186.0, "learning_rate": 2.3333333333333332e-05, "loss": 0.0001, "step": 2976 }, { "epoch": 186.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 71.50327381251103, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5, \"recall\": 0.8333333333333334, \"f1-score\": 0.625, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6756756756756757, \"f1-score\": 0.7142857142857142, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7827068764568765, \"recall\": 0.7918407318608931, \"f1-score\": 0.7722572480409168, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7499145968973555, \"recall\": 0.7241379310344828, \"f1-score\": 0.7254042032839941, \"support\": 116.0}}", "eval_f1_macro": 68.64508871474816, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.5404203283994, "eval_loss": 1.298017978668213, "eval_runtime": 2.339, "eval_samples_per_second": 49.594, "step": 2976 }, { "epoch": 187.0, "learning_rate": 2.1666666666666664e-05, "loss": 0.0066, "step": 2992 }, { "epoch": 187.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 72.35246834449421, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5, \"recall\": 0.8333333333333334, \"f1-score\": 0.625, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.7027027027027027, \"f1-score\": 0.7323943661971832, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7938254147812971, \"recall\": 0.7952191102392715, \"f1-score\": 0.7804418821614294, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7585368232173506, \"recall\": 0.7327586206896551, \"f1-score\": 0.7348553749236323, \"support\": 116.0}}", "eval_f1_macro": 69.37261174768261, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.48553749236324, "eval_loss": 1.2962392568588257, "eval_runtime": 2.4107, "eval_samples_per_second": 48.118, "step": 2992 }, { "epoch": 188.0, "learning_rate": 1.9999999999999998e-05, "loss": 0.0005, "step": 3008 }, { "epoch": 188.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 72.35124316220256, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.7027027027027027, \"f1-score\": 0.7222222222222223, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7943837960285329, \"recall\": 0.7952191102392715, \"f1-score\": 0.7816905254548818, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.754290151772003, \"recall\": 0.7327586206896551, \"f1-score\": 0.7336964624822306, \"support\": 116.0}}", "eval_f1_macro": 69.48360226265616, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.36964624822306, "eval_loss": 1.28173828125, "eval_runtime": 2.3316, "eval_samples_per_second": 49.751, "step": 3008 }, { "epoch": 189.0, "learning_rate": 1.833333333333333e-05, "loss": 0.0001, "step": 3024 }, { "epoch": 189.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 73.08583159191204, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.7297297297297297, \"f1-score\": 0.7397260273972601, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.798931623931624, \"recall\": 0.7985974886176499, \"f1-score\": 0.7865666731447722, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7595932802829355, \"recall\": 0.7413793103448276, \"f1-score\": 0.7415042668581396, \"support\": 116.0}}", "eval_f1_macro": 69.91703761286865, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.15042668581397, "eval_loss": 1.2783206701278687, "eval_runtime": 2.2566, "eval_samples_per_second": 51.406, "step": 3024 }, { "epoch": 190.0, "learning_rate": 1.6666666666666664e-05, "loss": 0.001, "step": 3040 }, { "epoch": 190.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 72.25528619564864, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7297297297297297, \"recall\": 0.7297297297297297, \"f1-score\": 0.7297297297297297, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7926099613599613, \"recall\": 0.7847085997287611, \"f1-score\": 0.7775790406982358, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7507766658628727, \"recall\": 0.7327586206896551, \"f1-score\": 0.7335128369370923, \"support\": 116.0}}", "eval_f1_macro": 69.1181369509543, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.35128369370923, "eval_loss": 1.278748631477356, "eval_runtime": 2.3035, "eval_samples_per_second": 50.357, "step": 3040 }, { "epoch": 191.0, "learning_rate": 1.4999999999999999e-05, "loss": 0.0001, "step": 3056 }, { "epoch": 191.0, "eval_accuracy": 75.0, "eval_average_metrics": 73.71814517325353, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.8333333333333334, \"recall\": 0.6451612903225806, \"f1-score\": 0.7272727272727272, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7435897435897436, \"recall\": 0.7837837837837838, \"f1-score\": 0.7631578947368421, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.75, \"macro avg\": {\"precision\": 0.8023552836052836, \"recall\": 0.7914653564855179, \"f1-score\": 0.7849473539876017, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7723283900008039, \"recall\": 0.75, \"f1-score\": 0.7509948256078287, \"support\": 116.0}}", "eval_f1_macro": 69.77309813223127, "eval_f1_micro": 75.0, "eval_f1_weighted": 75.09948256078287, "eval_loss": 1.2824381589889526, "eval_runtime": 2.3428, "eval_samples_per_second": 49.513, "step": 3056 }, { "epoch": 192.0, "learning_rate": 1.3333333333333333e-05, "loss": 0.0003, "step": 3072 }, { "epoch": 192.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 72.98695667868029, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.8, \"recall\": 0.6451612903225806, \"f1-score\": 0.7142857142857142, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7368421052631579, \"recall\": 0.7567567567567568, \"f1-score\": 0.7466666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7973451621477937, \"recall\": 0.7880869781071395, \"f1-score\": 0.7812625738554531, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7612680800747951, \"recall\": 0.7413793103448276, \"f1-score\": 0.742264025252709, \"support\": 116.0}}", "eval_f1_macro": 69.44556212048472, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.22640252527088, "eval_loss": 1.2781001329421997, "eval_runtime": 2.2829, "eval_samples_per_second": 50.812, "step": 3072 }, { "epoch": 193.0, "learning_rate": 1.1666666666666666e-05, "loss": 0.0006, "step": 3088 }, { "epoch": 193.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 72.25528619564864, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7297297297297297, \"recall\": 0.7297297297297297, \"f1-score\": 0.7297297297297297, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7926099613599613, \"recall\": 0.7847085997287611, \"f1-score\": 0.7775790406982358, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7507766658628727, \"recall\": 0.7327586206896551, \"f1-score\": 0.7335128369370923, \"support\": 116.0}}", "eval_f1_macro": 69.1181369509543, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.35128369370923, "eval_loss": 1.2707769870758057, "eval_runtime": 2.3425, "eval_samples_per_second": 49.519, "step": 3088 }, { "epoch": 194.0, "learning_rate": 9.999999999999999e-06, "loss": 0.0001, "step": 3104 }, { "epoch": 194.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 72.25528619564864, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7297297297297297, \"recall\": 0.7297297297297297, \"f1-score\": 0.7297297297297297, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7926099613599613, \"recall\": 0.7847085997287611, \"f1-score\": 0.7775790406982358, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7507766658628727, \"recall\": 0.7327586206896551, \"f1-score\": 0.7335128369370923, \"support\": 116.0}}", "eval_f1_macro": 69.1181369509543, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.35128369370923, "eval_loss": 1.267704725265503, "eval_runtime": 2.4427, "eval_samples_per_second": 47.488, "step": 3104 }, { "epoch": 195.0, "learning_rate": 8.333333333333332e-06, "loss": 0.0001, "step": 3120 }, { "epoch": 195.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 72.25528619564864, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7297297297297297, \"recall\": 0.7297297297297297, \"f1-score\": 0.7297297297297297, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7926099613599613, \"recall\": 0.7847085997287611, \"f1-score\": 0.7775790406982358, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7507766658628727, \"recall\": 0.7327586206896551, \"f1-score\": 0.7335128369370923, \"support\": 116.0}}", "eval_f1_macro": 69.1181369509543, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.35128369370923, "eval_loss": 1.2668423652648926, "eval_runtime": 2.2674, "eval_samples_per_second": 51.16, "step": 3120 }, { "epoch": 196.0, "learning_rate": 6.666666666666667e-06, "loss": 0.0001, "step": 3136 }, { "epoch": 196.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 71.52003226355563, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5, \"recall\": 0.8333333333333334, \"f1-score\": 0.625, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.7027027027027027, \"f1-score\": 0.7222222222222223, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7873064435564436, \"recall\": 0.7813302213503827, \"f1-score\": 0.7714322689264641, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7492167315443178, \"recall\": 0.7241379310344828, \"f1-score\": 0.7268078560941804, \"support\": 116.0}}", "eval_f1_macro": 68.57175723790792, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.68078560941804, "eval_loss": 1.2636795043945312, "eval_runtime": 2.3362, "eval_samples_per_second": 49.653, "step": 3136 }, { "epoch": 197.0, "learning_rate": 4.9999999999999996e-06, "loss": 0.0001, "step": 3152 }, { "epoch": 197.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 71.52003226355563, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5, \"recall\": 0.8333333333333334, \"f1-score\": 0.625, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.7027027027027027, \"f1-score\": 0.7222222222222223, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7873064435564436, \"recall\": 0.7813302213503827, \"f1-score\": 0.7714322689264641, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7492167315443178, \"recall\": 0.7241379310344828, \"f1-score\": 0.7268078560941804, \"support\": 116.0}}", "eval_f1_macro": 68.57175723790792, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.68078560941804, "eval_loss": 1.2627345323562622, "eval_runtime": 2.4056, "eval_samples_per_second": 48.222, "step": 3152 }, { "epoch": 198.0, "learning_rate": 3.3333333333333333e-06, "loss": 0.0001, "step": 3168 }, { "epoch": 198.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 70.70675091356331, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5, \"recall\": 0.8333333333333334, \"f1-score\": 0.625, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.7027027027027027, \"f1-score\": 0.7123287671232876, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7833381895881897, \"recall\": 0.7674413324614937, \"f1-score\": 0.7628856455186293, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7417728344452482, \"recall\": 0.7155172413793104, \"f1-score\": 0.7191149799895746, \"support\": 116.0}}", "eval_f1_macro": 67.81205737943371, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.91149799895746, "eval_loss": 1.2624974250793457, "eval_runtime": 2.3963, "eval_samples_per_second": 48.408, "step": 3168 }, { "epoch": 199.0, "learning_rate": 1.6666666666666667e-06, "loss": 0.0, "step": 3184 }, { "epoch": 199.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 70.70675091356331, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5, \"recall\": 0.8333333333333334, \"f1-score\": 0.625, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.7027027027027027, \"f1-score\": 0.7123287671232876, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7833381895881897, \"recall\": 0.7674413324614937, \"f1-score\": 0.7628856455186293, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7417728344452482, \"recall\": 0.7155172413793104, \"f1-score\": 0.7191149799895746, \"support\": 116.0}}", "eval_f1_macro": 67.81205737943371, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.91149799895746, "eval_loss": 1.262513279914856, "eval_runtime": 2.3048, "eval_samples_per_second": 50.331, "step": 3184 }, { "epoch": 200.0, "learning_rate": 0.0, "loss": 0.0009, "step": 3200 }, { "epoch": 200.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 70.70675091356331, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5, \"recall\": 0.8333333333333334, \"f1-score\": 0.625, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.7027027027027027, \"f1-score\": 0.7123287671232876, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7833381895881897, \"recall\": 0.7674413324614937, \"f1-score\": 0.7628856455186293, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7417728344452482, \"recall\": 0.7155172413793104, \"f1-score\": 0.7191149799895746, \"support\": 116.0}}", "eval_f1_macro": 67.81205737943371, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.91149799895746, "eval_loss": 1.2624309062957764, "eval_runtime": 2.4394, "eval_samples_per_second": 47.554, "step": 3200 }, { "epoch": 200.0, "step": 3200, "total_flos": 1.565785344442368e+16, "train_runtime": 2488.3254, "train_samples_per_second": 1.286 } ], "max_steps": 3200, "num_train_epochs": 200, "total_flos": 1.565785344442368e+16, "trial_name": null, "trial_params": null }