{ "best_metric": 76.93520799124248, "best_model_checkpoint": "/mounts/work/faeze/attempt/new_setting_outputs/source_adapter/crisis_13/0/checkpoint-1152", "epoch": 200.0, "global_step": 3200, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 1.0, "learning_rate": 1.4999999999999999e-05, "loss": 9.5194, "step": 16 }, { "epoch": 1.0, "eval_accuracy": 0.0, "eval_average_metrics": 0.0, "eval_classification_report": "{\"\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"\\\"media\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"How Dar\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"-\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \".\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"0:\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \":\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": D\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": Lots\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": apparently\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": good\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": please\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": very\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"????\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Anders\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Emers\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Shi\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Can use passport\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"False\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"No Friday Im\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"No they have\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"PPC 2019\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"SCP-04\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Shawn\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"True\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a friend\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"canada is under\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 31.0}, \"classification:\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"evaluation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 37.0}, \"instagram\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"weather\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.0, \"macro avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}}", "eval_f1_macro": 0.0, "eval_f1_micro": 0.0, "eval_f1_weighted": 0.0, "eval_loss": 9.631518363952637, "eval_runtime": 2.1772, "eval_samples_per_second": 53.278, "step": 16 }, { "epoch": 2.0, "learning_rate": 2.9999999999999997e-05, "loss": 8.9553, "step": 32 }, { "epoch": 2.0, "eval_accuracy": 0.0, "eval_average_metrics": 0.0, "eval_classification_report": "{\"\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"\\\"media\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"How Dar\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"-\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \".\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"0\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": D\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": Lots\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": apparently\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": good\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": please\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": very\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Anders\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Emers\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Can use passport\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Climate change is\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"False\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"No Friday Im\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"No they have\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"PPC 2019\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"SCP-04\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Shawn\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"True\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a friend\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"canada is under\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 37.0}, \"instagram\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"negative\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"vac\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"weather\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.0, \"macro avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}}", "eval_f1_macro": 0.0, "eval_f1_micro": 0.0, "eval_f1_weighted": 0.0, "eval_loss": 9.292896270751953, "eval_runtime": 2.1118, "eval_samples_per_second": 54.93, "step": 32 }, { "epoch": 3.0, "learning_rate": 4.4999999999999996e-05, "loss": 8.2015, "step": 48 }, { "epoch": 3.0, "eval_accuracy": 0.0, "eval_average_metrics": 0.0, "eval_classification_report": "{\"\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"How Dar\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"-\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": Lots\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": good\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": please\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": very\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Anders\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Emers\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Australian\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Can use passport\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Climate change is\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Ddx\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"False\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"No Friday Im\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"PPC 2019\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Shawn\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"True\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a boat\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a friend\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a real\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"canada is under\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 37.0}, \"instagram\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"negative\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_e\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"vac\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"weather\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.0, \"macro avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}}", "eval_f1_macro": 0.0, "eval_f1_micro": 0.0, "eval_f1_weighted": 0.0, "eval_loss": 8.517956733703613, "eval_runtime": 2.0756, "eval_samples_per_second": 55.889, "step": 48 }, { "epoch": 4.0, "learning_rate": 5.9999999999999995e-05, "loss": 7.0499, "step": 64 }, { "epoch": 4.0, "eval_accuracy": 0.0, "eval_average_metrics": 0.0, "eval_classification_report": "{\"\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"How Dar\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"Reality\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"116.\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": lots\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \":D\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"?\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"????\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@hara go\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Australian\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"False\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"No Friday Im\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"PPC 2019\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Please show us\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"SCP-04\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Sarah\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Shawn\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"True\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a friend\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"canada\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 31.0}, \"classification: D\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"classification: good\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"classification: very\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"climate change is\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"eder\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"entail\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"evaluation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"i will\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 37.0}, \"instagram\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"negative\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not a\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_e\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_en\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"panne\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"vac\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"weather\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.0, \"macro avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}}", "eval_f1_macro": 0.0, "eval_f1_micro": 0.0, "eval_f1_weighted": 0.0, "eval_loss": 6.531414031982422, "eval_runtime": 2.4081, "eval_samples_per_second": 48.171, "step": 64 }, { "epoch": 5.0, "learning_rate": 7.5e-05, "loss": 5.3806, "step": 80 }, { "epoch": 5.0, "eval_accuracy": 0.0, "eval_average_metrics": 0.0, "eval_classification_report": "{\"\\\"How Dar\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"Reality\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \".\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"0\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"1\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"33000f\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \":D\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"?\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Si\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@hara go\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Australia\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Australian\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Bruh\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"CNN\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"COVID\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Ddx\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"FEAR\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"FGS\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"False\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Florida\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"France\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"India\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"KGB\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"O U C\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"PPC\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"SCP-04\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"True\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"US\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a grand\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"canada\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 31.0}, \"classification\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"classification: good\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"conservative\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"dal\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"eder\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"entail\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"evaluation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"i\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 37.0}, \"instagram\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"jai hin\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"microwave\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"name\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"negative\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_d\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_e\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_en\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"panne\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"rn\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"to irri\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"trud\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"vac\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"weather\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.0, \"macro avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}}", "eval_f1_macro": 0.0, "eval_f1_micro": 0.0, "eval_f1_weighted": 0.0, "eval_loss": 3.8640472888946533, "eval_runtime": 2.1363, "eval_samples_per_second": 54.3, "step": 80 }, { "epoch": 6.0, "learning_rate": 8.999999999999999e-05, "loss": 3.0014, "step": 96 }, { "epoch": 6.0, "eval_accuracy": 12.068965517241379, "eval_average_metrics": 10.466996817571529, "eval_classification_report": "{\"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"challenge\": {\"precision\": 0.30434782608695654, \"recall\": 0.45161290322580644, \"f1-score\": 0.3636363636363636, \"support\": 31.0}, \"entail\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"evaluation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 37.0}, \"negative\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_e\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.1206896551724138, \"macro avg\": {\"precision\": 0.023411371237458196, \"recall\": 0.034739454094292806, \"f1-score\": 0.02797202797202797, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.08133433283358321, \"recall\": 0.1206896551724138, \"f1-score\": 0.09717868338557993, \"support\": 116.0}}", "eval_f1_macro": 2.7972027972027966, "eval_f1_micro": 17.28395061728395, "eval_f1_weighted": 9.717868338557993, "eval_loss": 1.5544441938400269, "eval_runtime": 2.0085, "eval_samples_per_second": 57.753, "step": 96 }, { "epoch": 7.0, "learning_rate": 0.00010499999999999999, "loss": 1.4145, "step": 112 }, { "epoch": 7.0, "eval_accuracy": 31.896551724137932, "eval_average_metrics": 20.735164092162066, "eval_classification_report": "{\"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.31896551724137934, \"recall\": 1.0, \"f1-score\": 0.48366013071895425, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accuracy\": 0.31896551724137934, \"macro avg\": {\"precision\": 0.03987068965517242, \"recall\": 0.125, \"f1-score\": 0.06045751633986928, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.10173900118906065, \"recall\": 0.31896551724137934, \"f1-score\": 0.15427090376380437, \"support\": 116.0}}", "eval_f1_macro": 3.720462543991956, "eval_f1_micro": 31.896551724137932, "eval_f1_weighted": 15.427090376380438, "eval_loss": 0.8676283359527588, "eval_runtime": 2.1145, "eval_samples_per_second": 54.861, "step": 112 }, { "epoch": 8.0, "learning_rate": 0.00011999999999999999, "loss": 0.938, "step": 128 }, { "epoch": 8.0, "eval_accuracy": 27.586206896551722, "eval_average_metrics": 18.972405875133617, "eval_classification_report": "{\"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"challenge\": {\"precision\": 0.27184466019417475, \"recall\": 0.9032258064516129, \"f1-score\": 0.417910447761194, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.3076923076923077, \"recall\": 0.10810810810810811, \"f1-score\": 0.16, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accuracy\": 0.27586206896551724, \"macro avg\": {\"precision\": 0.07244212098581031, \"recall\": 0.1264167393199651, \"f1-score\": 0.07223880597014926, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.17079137802271382, \"recall\": 0.27586206896551724, \"f1-score\": 0.16271744724652598, \"support\": 116.0}}", "eval_f1_macro": 4.445464982778416, "eval_f1_micro": 27.586206896551722, "eval_f1_weighted": 16.2717447246526, "eval_loss": 0.7047381401062012, "eval_runtime": 2.1706, "eval_samples_per_second": 53.441, "step": 128 }, { "epoch": 9.0, "learning_rate": 0.000135, "loss": 0.8323, "step": 144 }, { "epoch": 9.0, "eval_accuracy": 37.93103448275862, "eval_average_metrics": 27.735665281304225, "eval_classification_report": "{\"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"challenge\": {\"precision\": 0.36764705882352944, \"recall\": 0.8064516129032258, \"f1-score\": 0.5050505050505051, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.3958333333333333, \"recall\": 0.5135135135135135, \"f1-score\": 0.44705882352941173, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accuracy\": 0.3793103448275862, \"macro avg\": {\"precision\": 0.09543504901960784, \"recall\": 0.1649956408020924, \"f1-score\": 0.1190136660724896, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.22450769100743745, \"recall\": 0.3793103448275862, \"f1-score\": 0.27756674247546453, \"support\": 116.0}}", "eval_f1_macro": 7.323917912153206, "eval_f1_micro": 37.93103448275862, "eval_f1_weighted": 27.756674247546453, "eval_loss": 0.6856001615524292, "eval_runtime": 2.0791, "eval_samples_per_second": 55.794, "step": 144 }, { "epoch": 10.0, "learning_rate": 0.00015, "loss": 0.736, "step": 160 }, { "epoch": 10.0, "eval_accuracy": 36.206896551724135, "eval_average_metrics": 26.25442086648983, "eval_classification_report": "{\"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"challenge\": {\"precision\": 0.345679012345679, \"recall\": 0.9032258064516129, \"f1-score\": 0.5, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.4, \"recall\": 0.3783783783783784, \"f1-score\": 0.3888888888888889, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accuracy\": 0.3620689655172414, \"macro avg\": {\"precision\": 0.09320987654320988, \"recall\": 0.1602005231037489, \"f1-score\": 0.1111111111111111, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.21996594295444868, \"recall\": 0.3620689655172414, \"f1-score\": 0.25766283524904215, \"support\": 116.0}}", "eval_f1_macro": 6.837606837606837, "eval_f1_micro": 36.20689655172414, "eval_f1_weighted": 25.766283524904214, "eval_loss": 0.6892978549003601, "eval_runtime": 2.0856, "eval_samples_per_second": 55.619, "step": 160 }, { "epoch": 11.0, "learning_rate": 0.000165, "loss": 0.7248, "step": 176 }, { "epoch": 11.0, "eval_accuracy": 39.6551724137931, "eval_average_metrics": 29.085576143629396, "eval_classification_report": "{\"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"challenge\": {\"precision\": 0.4166666666666667, \"recall\": 0.6451612903225806, \"f1-score\": 0.5063291139240507, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.38235294117647056, \"recall\": 0.7027027027027027, \"f1-score\": 0.4952380952380953, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accuracy\": 0.39655172413793105, \"macro avg\": {\"precision\": 0.09987745098039216, \"recall\": 0.16848299912816042, \"f1-score\": 0.12519590114526824, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.2333079783637593, \"recall\": 0.39655172413793105, \"f1-score\": 0.2932759659953026, \"support\": 116.0}}", "eval_f1_macro": 7.704363147401122, "eval_f1_micro": 39.6551724137931, "eval_f1_weighted": 29.32759659953026, "eval_loss": 0.6886057257652283, "eval_runtime": 2.1055, "eval_samples_per_second": 55.093, "step": 176 }, { "epoch": 12.0, "learning_rate": 0.00017999999999999998, "loss": 0.6738, "step": 192 }, { "epoch": 12.0, "eval_accuracy": 37.93103448275862, "eval_average_metrics": 27.760615065515246, "eval_classification_report": "{\"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"challenge\": {\"precision\": 0.41025641025641024, \"recall\": 0.5161290322580645, \"f1-score\": 0.4571428571428572, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.36363636363636365, \"recall\": 0.7567567567567568, \"f1-score\": 0.4912280701754386, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accuracy\": 0.3793103448275862, \"macro avg\": {\"precision\": 0.09673659673659674, \"recall\": 0.15911072362685266, \"f1-score\": 0.11854636591478698, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.2256249497628808, \"recall\": 0.3793103448275862, \"f1-score\": 0.27885230317172244, \"support\": 116.0}}", "eval_f1_macro": 7.295160979371507, "eval_f1_micro": 37.93103448275862, "eval_f1_weighted": 27.885230317172244, "eval_loss": 0.6640838980674744, "eval_runtime": 2.2026, "eval_samples_per_second": 52.665, "step": 192 }, { "epoch": 13.0, "learning_rate": 0.000195, "loss": 0.7337, "step": 208 }, { "epoch": 13.0, "eval_accuracy": 37.06896551724138, "eval_average_metrics": 27.156620348964545, "eval_classification_report": "{\"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"challenge\": {\"precision\": 0.38636363636363635, \"recall\": 0.5483870967741935, \"f1-score\": 0.4533333333333333, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.3611111111111111, \"recall\": 0.7027027027027027, \"f1-score\": 0.47706422018348627, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accuracy\": 0.3706896551724138, \"macro avg\": {\"precision\": 0.09343434343434343, \"recall\": 0.15638622493461202, \"f1-score\": 0.11629969418960245, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.21843434343434343, \"recall\": 0.3706896551724138, \"f1-score\": 0.2733164610355373, \"support\": 116.0}}", "eval_f1_macro": 7.156904257821689, "eval_f1_micro": 37.06896551724138, "eval_f1_weighted": 27.33164610355373, "eval_loss": 0.6635391712188721, "eval_runtime": 2.234, "eval_samples_per_second": 51.925, "step": 208 }, { "epoch": 14.0, "learning_rate": 0.00020999999999999998, "loss": 0.7431, "step": 224 }, { "epoch": 14.0, "eval_accuracy": 37.93103448275862, "eval_average_metrics": 27.575873831071966, "eval_classification_report": "{\"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"challenge\": {\"precision\": 0.46153846153846156, \"recall\": 0.3870967741935484, \"f1-score\": 0.42105263157894735, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.35555555555555557, \"recall\": 0.8648648648648649, \"f1-score\": 0.5039370078740157, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accuracy\": 0.3793103448275862, \"macro avg\": {\"precision\": 0.10213675213675213, \"recall\": 0.15649520488230168, \"f1-score\": 0.11562370493162039, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.23675213675213677, \"recall\": 0.3793103448275862, \"f1-score\": 0.2732612143990168, \"support\": 116.0}}", "eval_f1_macro": 7.115304918868946, "eval_f1_micro": 37.93103448275862, "eval_f1_weighted": 27.32612143990168, "eval_loss": 0.6511205434799194, "eval_runtime": 2.2897, "eval_samples_per_second": 50.662, "step": 224 }, { "epoch": 15.0, "learning_rate": 0.000225, "loss": 0.7529, "step": 240 }, { "epoch": 15.0, "eval_accuracy": 38.793103448275865, "eval_average_metrics": 28.441580913172167, "eval_classification_report": "{\"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"challenge\": {\"precision\": 0.39285714285714285, \"recall\": 0.7096774193548387, \"f1-score\": 0.5057471264367817, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.38333333333333336, \"recall\": 0.6216216216216216, \"f1-score\": 0.4742268041237113, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accuracy\": 0.3879310344827586, \"macro avg\": {\"precision\": 0.09702380952380953, \"recall\": 0.16641238012205756, \"f1-score\": 0.12249674132006162, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.22725779967159276, \"recall\": 0.3879310344827586, \"f1-score\": 0.28641855751825473, \"support\": 116.0}}", "eval_f1_macro": 7.538261004311485, "eval_f1_micro": 38.793103448275865, "eval_f1_weighted": 28.641855751825474, "eval_loss": 0.6641126871109009, "eval_runtime": 2.221, "eval_samples_per_second": 52.229, "step": 240 }, { "epoch": 16.0, "learning_rate": 0.00023999999999999998, "loss": 0.7239, "step": 256 }, { "epoch": 16.0, "eval_accuracy": 36.206896551724135, "eval_average_metrics": 26.256044568490083, "eval_classification_report": "{\"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"challenge\": {\"precision\": 0.4230769230769231, \"recall\": 0.3548387096774194, \"f1-score\": 0.3859649122807018, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.34444444444444444, \"recall\": 0.8378378378378378, \"f1-score\": 0.4881889763779528, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accuracy\": 0.3620689655172414, \"macro avg\": {\"precision\": 0.09594017094017093, \"recall\": 0.14908456843940715, \"f1-score\": 0.10926923608233183, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.22292956086059534, \"recall\": 0.3620689655172414, \"f1-score\": 0.25886124488522416, \"support\": 116.0}}", "eval_f1_macro": 6.724260681989651, "eval_f1_micro": 36.20689655172414, "eval_f1_weighted": 25.886124488522416, "eval_loss": 0.6616472601890564, "eval_runtime": 2.2101, "eval_samples_per_second": 52.487, "step": 256 }, { "epoch": 17.0, "learning_rate": 0.00025499999999999996, "loss": 0.7121, "step": 272 }, { "epoch": 17.0, "eval_accuracy": 38.793103448275865, "eval_average_metrics": 28.220869647593787, "eval_classification_report": "{\"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"challenge\": {\"precision\": 0.4057971014492754, \"recall\": 0.9032258064516129, \"f1-score\": 0.5599999999999999, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.3617021276595745, \"recall\": 0.4594594594594595, \"f1-score\": 0.40476190476190477, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accuracy\": 0.3879310344827586, \"macro avg\": {\"precision\": 0.09593740363860623, \"recall\": 0.17033565823888405, \"f1-score\": 0.12059523809523809, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.22381628334768788, \"recall\": 0.3879310344827586, \"f1-score\": 0.27876026272578, \"support\": 116.0}}", "eval_f1_macro": 7.421245421245421, "eval_f1_micro": 38.793103448275865, "eval_f1_weighted": 27.876026272578, "eval_loss": 0.6310274600982666, "eval_runtime": 2.2051, "eval_samples_per_second": 52.605, "step": 272 }, { "epoch": 18.0, "learning_rate": 0.00027, "loss": 0.6385, "step": 288 }, { "epoch": 18.0, "eval_accuracy": 42.241379310344826, "eval_average_metrics": 31.05062473823817, "eval_classification_report": "{\"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"challenge\": {\"precision\": 0.4888888888888889, \"recall\": 0.7096774193548387, \"f1-score\": 0.5789473684210527, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.38028169014084506, \"recall\": 0.7297297297297297, \"f1-score\": 0.5, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accuracy\": 0.4224137931034483, \"macro avg\": {\"precision\": 0.10864632237871674, \"recall\": 0.17992589363557104, \"f1-score\": 0.13486842105263158, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.25194808698936916, \"recall\": 0.4224137931034483, \"f1-score\": 0.3142014519056261, \"support\": 116.0}}", "eval_f1_macro": 8.299595141700406, "eval_f1_micro": 42.24137931034483, "eval_f1_weighted": 31.42014519056261, "eval_loss": 0.6071025133132935, "eval_runtime": 2.27, "eval_samples_per_second": 51.101, "step": 288 }, { "epoch": 19.0, "learning_rate": 0.000285, "loss": 0.5922, "step": 304 }, { "epoch": 19.0, "eval_accuracy": 45.689655172413794, "eval_average_metrics": 36.16077557122249, "eval_classification_report": "{\"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.25, \"recall\": 0.125, \"f1-score\": 0.16666666666666666, \"support\": 8.0}, \"challenge\": {\"precision\": 0.5510204081632653, \"recall\": 0.8709677419354839, \"f1-score\": 0.6749999999999999, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.1, \"recall\": 0.1111111111111111, \"f1-score\": 0.10526315789473685, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.44680851063829785, \"recall\": 0.5675675675675675, \"f1-score\": 0.5, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5, \"recall\": 0.3333333333333333, \"f1-score\": 0.4, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accuracy\": 0.45689655172413796, \"macro avg\": {\"precision\": 0.23097861485019539, \"recall\": 0.25099746924343697, \"f1-score\": 0.23086622807017543, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.353565065057571, \"recall\": 0.45689655172413796, \"f1-score\": 0.39056639443436175, \"support\": 116.0}}", "eval_f1_macro": 14.207152496626179, "eval_f1_micro": 45.689655172413794, "eval_f1_weighted": 39.056639443436175, "eval_loss": 0.5678648352622986, "eval_runtime": 2.2264, "eval_samples_per_second": 52.103, "step": 304 }, { "epoch": 20.0, "learning_rate": 0.0003, "loss": 0.6507, "step": 320 }, { "epoch": 20.0, "eval_accuracy": 50.0, "eval_average_metrics": 40.86473713418625, "eval_classification_report": "{\"accusation\": {\"precision\": 0.375, \"recall\": 0.25, \"f1-score\": 0.3, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.5, \"recall\": 0.375, \"f1-score\": 0.42857142857142855, \"support\": 8.0}, \"challenge\": {\"precision\": 0.5094339622641509, \"recall\": 0.8709677419354839, \"f1-score\": 0.6428571428571428, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.5, \"recall\": 0.1111111111111111, \"f1-score\": 0.1818181818181818, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.4888888888888889, \"recall\": 0.5945945945945946, \"f1-score\": 0.5365853658536586, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.2222222222222222, \"f1-score\": 0.3636363636363636, \"support\": 9.0}, \"accuracy\": 0.5, \"macro avg\": {\"precision\": 0.42166535639413, \"recall\": 0.30298695873292647, \"f1-score\": 0.3066835603420969, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.4817357044748066, \"recall\": 0.5, \"f1-score\": 0.44586114054154424, \"support\": 116.0}}", "eval_f1_macro": 18.87283448259058, "eval_f1_micro": 50.0, "eval_f1_weighted": 44.58611405415442, "eval_loss": 0.5527870655059814, "eval_runtime": 2.2006, "eval_samples_per_second": 52.713, "step": 320 }, { "epoch": 21.0, "learning_rate": 0.00029833333333333334, "loss": 0.6159, "step": 336 }, { "epoch": 21.0, "eval_accuracy": 50.0, "eval_average_metrics": 39.89382649262413, "eval_classification_report": "{\"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.375, \"f1-score\": 0.5, \"support\": 8.0}, \"challenge\": {\"precision\": 0.5306122448979592, \"recall\": 0.8387096774193549, \"f1-score\": 0.65, \"support\": 31.0}, \"evaluation\": {\"precision\": 1.0, \"recall\": 0.1111111111111111, \"f1-score\": 0.19999999999999998, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.4482758620689655, \"recall\": 0.7027027027027027, \"f1-score\": 0.5473684210526315, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.2222222222222222, \"f1-score\": 0.3333333333333333, \"support\": 9.0}, \"accuracy\": 0.5, \"macro avg\": {\"precision\": 0.42444434670419895, \"recall\": 0.28121821418192383, \"f1-score\": 0.27883771929824563, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.4658205731757626, \"recall\": 0.5, \"f1-score\": 0.4241606170598911, \"support\": 116.0}}", "eval_f1_macro": 17.15924426450742, "eval_f1_micro": 50.0, "eval_f1_weighted": 42.41606170598911, "eval_loss": 0.49902039766311646, "eval_runtime": 2.1471, "eval_samples_per_second": 54.026, "step": 336 }, { "epoch": 22.0, "learning_rate": 0.00029666666666666665, "loss": 0.4905, "step": 352 }, { "epoch": 22.0, "eval_accuracy": 55.172413793103445, "eval_average_metrics": 47.567859823700545, "eval_classification_report": "{\"accusation\": {\"precision\": 0.3, \"recall\": 0.5, \"f1-score\": 0.37499999999999994, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.5, \"f1-score\": 0.5714285714285715, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.6451612903225806, \"f1-score\": 0.689655172413793, \"support\": 31.0}, \"evaluation\": {\"precision\": 1.0, \"recall\": 0.4444444444444444, \"f1-score\": 0.6153846153846153, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.49019607843137253, \"recall\": 0.6756756756756757, \"f1-score\": 0.5681818181818181, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.625, \"recall\": 0.5555555555555556, \"f1-score\": 0.5882352941176471, \"support\": 9.0}, \"accuracy\": 0.5517241379310345, \"macro avg\": {\"precision\": 0.4778254357298475, \"recall\": 0.41510462074978205, \"f1-score\": 0.4259856839408056, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.5574013034332508, \"recall\": 0.5517241379310345, \"f1-score\": 0.5371210808146878, \"support\": 116.0}}", "eval_f1_macro": 26.2145036271265, "eval_f1_micro": 55.172413793103445, "eval_f1_weighted": 53.71210808146878, "eval_loss": 0.4539774954319, "eval_runtime": 2.1666, "eval_samples_per_second": 53.539, "step": 352 }, { "epoch": 23.0, "learning_rate": 0.00029499999999999996, "loss": 0.4571, "step": 368 }, { "epoch": 23.0, "eval_accuracy": 57.758620689655174, "eval_average_metrics": 51.260575594623866, "eval_classification_report": "{\"accusation\": {\"precision\": 0.4166666666666667, \"recall\": 0.8333333333333334, \"f1-score\": 0.5555555555555556, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.8333333333333334, \"recall\": 0.3225806451612903, \"f1-score\": 0.4651162790697674, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8571428571428571, \"recall\": 0.6666666666666666, \"f1-score\": 0.75, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6, \"recall\": 0.6486486486486487, \"f1-score\": 0.6233766233766234, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.35, \"recall\": 0.7777777777777778, \"f1-score\": 0.48275862068965514, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.4444444444444444, \"f1-score\": 0.5714285714285714, \"support\": 9.0}, \"accuracy\": 0.5775862068965517, \"macro avg\": {\"precision\": 0.5758928571428572, \"recall\": 0.5554314395040202, \"f1-score\": 0.5247794562650215, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.6646346469622332, \"recall\": 0.5775862068965517, \"f1-score\": 0.572309406136453, \"support\": 116.0}}", "eval_f1_macro": 32.29412038553979, "eval_f1_micro": 57.758620689655174, "eval_f1_weighted": 57.23094061364532, "eval_loss": 0.4593088924884796, "eval_runtime": 2.1213, "eval_samples_per_second": 54.683, "step": 368 }, { "epoch": 24.0, "learning_rate": 0.00029333333333333327, "loss": 0.3941, "step": 384 }, { "epoch": 24.0, "eval_accuracy": 66.37931034482759, "eval_average_metrics": 58.64231561905229, "eval_classification_report": "{\"accusation\": {\"precision\": 0.45454545454545453, \"recall\": 0.8333333333333334, \"f1-score\": 0.5882352941176471, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.7419354838709677, \"f1-score\": 0.7419354838709677, \"support\": 31.0}, \"evaluation\": {\"precision\": 1.0, \"recall\": 0.6666666666666666, \"f1-score\": 0.8, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6666666666666666, \"recall\": 0.6486486486486487, \"f1-score\": 0.6575342465753425, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.2857142857142857, \"recall\": 0.2222222222222222, \"f1-score\": 0.25, \"support\": 9.0}, \"request\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accuracy\": 0.6637931034482759, \"macro avg\": {\"precision\": 0.6123577363496717, \"recall\": 0.5661841276760632, \"f1-score\": 0.5750912793309989, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.6848503507986266, \"recall\": 0.6637931034482759, \"f1-score\": 0.6642040921233869, \"support\": 116.0}}", "eval_f1_macro": 35.39023257421532, "eval_f1_micro": 66.37931034482759, "eval_f1_weighted": 66.4204092123387, "eval_loss": 0.388798326253891, "eval_runtime": 2.0824, "eval_samples_per_second": 55.705, "step": 384 }, { "epoch": 25.0, "learning_rate": 0.00029166666666666664, "loss": 0.3707, "step": 400 }, { "epoch": 25.0, "eval_accuracy": 60.3448275862069, "eval_average_metrics": 53.652705233850234, "eval_classification_report": "{\"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.3333333333333333, \"recall\": 0.875, \"f1-score\": 0.48275862068965514, \"support\": 8.0}, \"challenge\": {\"precision\": 0.5806451612903226, \"recall\": 0.5806451612903226, \"f1-score\": 0.5806451612903226, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.782608695652174, \"recall\": 0.4864864864864865, \"f1-score\": 0.6000000000000001, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.42857142857142855, \"recall\": 0.3333333333333333, \"f1-score\": 0.375, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.6666666666666666, \"f1-score\": 0.75, \"support\": 9.0}, \"accuracy\": 0.603448275862069, \"macro avg\": {\"precision\": 0.5533432400543199, \"recall\": 0.5830442337498789, \"f1-score\": 0.5429949171919417, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.6539765831370029, \"recall\": 0.603448275862069, \"f1-score\": 0.6050609393579073, \"support\": 116.0}}", "eval_f1_macro": 33.415071827196414, "eval_f1_micro": 60.3448275862069, "eval_f1_weighted": 60.50609393579073, "eval_loss": 0.4242806136608124, "eval_runtime": 2.2084, "eval_samples_per_second": 52.527, "step": 400 }, { "epoch": 26.0, "learning_rate": 0.00029, "loss": 0.339, "step": 416 }, { "epoch": 26.0, "eval_accuracy": 69.82758620689656, "eval_average_metrics": 61.77302749117365, "eval_classification_report": "{\"accusation\": {\"precision\": 0.625, \"recall\": 0.8333333333333334, \"f1-score\": 0.7142857142857143, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6571428571428571, \"recall\": 0.7419354838709677, \"f1-score\": 0.6969696969696969, \"support\": 31.0}, \"evaluation\": {\"precision\": 1.0, \"recall\": 0.6666666666666666, \"f1-score\": 0.8, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6923076923076923, \"recall\": 0.7297297297297297, \"f1-score\": 0.7105263157894737, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5, \"recall\": 0.4444444444444444, \"f1-score\": 0.47058823529411764, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.5555555555555556, \"f1-score\": 0.6666666666666667, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.6634729853479853, \"recall\": 0.5902081517000872, \"f1-score\": 0.6145224357685658, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7110932171276998, \"recall\": 0.6982758620689655, \"f1-score\": 0.6962017227283588, \"support\": 116.0}}", "eval_f1_macro": 37.81676527806559, "eval_f1_micro": 69.82758620689656, "eval_f1_weighted": 69.62017227283587, "eval_loss": 0.3623848557472229, "eval_runtime": 2.1161, "eval_samples_per_second": 54.817, "step": 416 }, { "epoch": 27.0, "learning_rate": 0.0002883333333333333, "loss": 0.3184, "step": 432 }, { "epoch": 27.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 62.719846550445325, "eval_classification_report": "{\"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.875, \"f1-score\": 0.9333333333333333, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.7096774193548387, \"f1-score\": 0.7096774193548389, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7419354838709677, \"recall\": 0.6216216216216216, \"f1-score\": 0.676470588235294, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.4444444444444444, \"recall\": 0.4444444444444444, \"f1-score\": 0.4444444444444444, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.6264456711325782, \"recall\": 0.657731824566502, \"f1-score\": 0.6369607776174497, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7091025084318285, \"recall\": 0.7068965517241379, \"f1-score\": 0.7030248954203372, \"support\": 116.0}}", "eval_f1_macro": 39.19758631491999, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.30248954203373, "eval_loss": 0.3541596531867981, "eval_runtime": 2.1012, "eval_samples_per_second": 55.205, "step": 432 }, { "epoch": 28.0, "learning_rate": 0.0002866666666666667, "loss": 0.2422, "step": 448 }, { "epoch": 28.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 63.936806022349785, "eval_classification_report": "{\"accusation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6857142857142857, \"recall\": 0.7741935483870968, \"f1-score\": 0.7272727272727272, \"support\": 31.0}, \"evaluation\": {\"precision\": 1.0, \"recall\": 0.6666666666666666, \"f1-score\": 0.8, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6590909090909091, \"recall\": 0.7837837837837838, \"f1-score\": 0.7160493827160493, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 0.3333333333333333, \"f1-score\": 0.46153846153846156, \"support\": 9.0}, \"request\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7028228715728716, \"recall\": 0.6044693887435824, \"f1-score\": 0.6362226508059842, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7361509180474698, \"recall\": 0.7241379310344828, \"f1-score\": 0.717674747559805, \"support\": 116.0}}", "eval_f1_macro": 39.152163126522105, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 71.7674747559805, "eval_loss": 0.3609805405139923, "eval_runtime": 2.2843, "eval_samples_per_second": 50.782, "step": 448 }, { "epoch": 29.0, "learning_rate": 0.000285, "loss": 0.2425, "step": 464 }, { "epoch": 29.0, "eval_accuracy": 69.82758620689656, "eval_average_metrics": 62.08778783554499, "eval_classification_report": "{\"accusation\": {\"precision\": 0.6923076923076923, \"recall\": 0.75, \"f1-score\": 0.7199999999999999, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.6451612903225806, \"f1-score\": 0.689655172413793, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7096774193548387, \"recall\": 0.5945945945945946, \"f1-score\": 0.6470588235294118, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5384615384615384, \"recall\": 0.7777777777777778, \"f1-score\": 0.6363636363636364, \"support\": 9.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.8888888888888888, \"f1-score\": 0.6956521739130435, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.6159519952866727, \"recall\": 0.6775389300590913, \"f1-score\": 0.6382293836722224, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7044631963430629, \"recall\": 0.6982758620689655, \"f1-score\": 0.6942032454855775, \"support\": 116.0}}", "eval_f1_macro": 39.275654379829064, "eval_f1_micro": 69.82758620689656, "eval_f1_weighted": 69.42032454855776, "eval_loss": 0.36870959401130676, "eval_runtime": 2.0508, "eval_samples_per_second": 56.564, "step": 464 }, { "epoch": 30.0, "learning_rate": 0.0002833333333333333, "loss": 0.1556, "step": 480 }, { "epoch": 30.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 64.72567209776729, "eval_classification_report": "{\"accusation\": {\"precision\": 0.6111111111111112, \"recall\": 0.9166666666666666, \"f1-score\": 0.7333333333333334, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"challenge\": {\"precision\": 0.76, \"recall\": 0.6129032258064516, \"f1-score\": 0.6785714285714285, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7692307692307693, \"recall\": 0.5405405405405406, \"f1-score\": 0.6349206349206349, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5384615384615384, \"recall\": 0.7777777777777778, \"f1-score\": 0.6363636363636364, \"support\": 9.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.8888888888888888, \"f1-score\": 0.6956521739130435, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.742926726051726, \"recall\": 0.8264721374600407, \"f1-score\": 0.7692301508877595, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7302372340303375, \"recall\": 0.7068965517241379, \"f1-score\": 0.7018613799161024, \"support\": 116.0}}", "eval_f1_macro": 47.33724005463135, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.18613799161024, "eval_loss": 0.409758597612381, "eval_runtime": 2.1135, "eval_samples_per_second": 54.884, "step": 480 }, { "epoch": 31.0, "learning_rate": 0.00028166666666666666, "loss": 0.1426, "step": 496 }, { "epoch": 31.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 66.17670734030372, "eval_classification_report": "{\"accusation\": {\"precision\": 0.625, \"recall\": 0.8333333333333334, \"f1-score\": 0.7142857142857143, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6756756756756757, \"f1-score\": 0.6944444444444444, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5454545454545454, \"recall\": 0.6666666666666666, \"f1-score\": 0.6, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7652417027417027, \"recall\": 0.7837202121476314, \"f1-score\": 0.7695496876911684, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7347100935463005, \"recall\": 0.7241379310344828, \"f1-score\": 0.7252233929640028, \"support\": 116.0}}", "eval_f1_macro": 47.356903857918056, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.52233929640029, "eval_loss": 0.38606610894203186, "eval_runtime": 2.0992, "eval_samples_per_second": 55.258, "step": 496 }, { "epoch": 32.0, "learning_rate": 0.00028, "loss": 0.1403, "step": 512 }, { "epoch": 32.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 64.83586644295046, "eval_classification_report": "{\"accusation\": {\"precision\": 0.5, \"recall\": 0.8333333333333334, \"f1-score\": 0.625, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7037037037037037, \"recall\": 0.6129032258064516, \"f1-score\": 0.6551724137931035, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7741935483870968, \"recall\": 0.6486486486486487, \"f1-score\": 0.7058823529411764, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5, \"recall\": 0.5555555555555556, \"f1-score\": 0.5263157894736842, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7552232676224612, \"recall\": 0.7879023176402209, \"f1-score\": 0.7647955885318435, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7255170353891154, \"recall\": 0.7068965517241379, \"f1-score\": 0.7089981151732236, \"support\": 116.0}}", "eval_f1_macro": 47.0643439096519, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.89981151732236, "eval_loss": 0.4265669584274292, "eval_runtime": 2.1308, "eval_samples_per_second": 54.439, "step": 512 }, { "epoch": 33.0, "learning_rate": 0.00027833333333333334, "loss": 0.1089, "step": 528 }, { "epoch": 33.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 66.9963380759733, "eval_classification_report": "{\"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"challenge\": {\"precision\": 0.8571428571428571, \"recall\": 0.5806451612903226, \"f1-score\": 0.6923076923076923, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.7297297297297297, \"f1-score\": 0.7397260273972601, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5454545454545454, \"recall\": 0.6666666666666666, \"f1-score\": 0.6, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7712593510290878, \"recall\": 0.8078940835997288, \"f1-score\": 0.7801354873645527, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7594117543073986, \"recall\": 0.7327586206896551, \"f1-score\": 0.7342529048198969, \"support\": 116.0}}", "eval_f1_macro": 48.00833768397247, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.4252904819897, "eval_loss": 0.5401732325553894, "eval_runtime": 2.2438, "eval_samples_per_second": 51.697, "step": 528 }, { "epoch": 34.0, "learning_rate": 0.00027666666666666665, "loss": 0.1282, "step": 544 }, { "epoch": 34.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 64.73343394035153, "eval_classification_report": "{\"accusation\": {\"precision\": 0.55, \"recall\": 0.9166666666666666, \"f1-score\": 0.6874999999999999, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7058823529411765, \"recall\": 0.6486486486486487, \"f1-score\": 0.676056338028169, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5454545454545454, \"recall\": 0.6666666666666666, \"f1-score\": 0.6, \"support\": 9.0}, \"request\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7600610378183907, \"recall\": 0.7728373534825147, \"f1-score\": 0.7583272391919456, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7259054713035443, \"recall\": 0.7068965517241379, \"f1-score\": 0.7088813377399731, \"support\": 116.0}}", "eval_f1_macro": 46.666291642581264, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.88813377399731, "eval_loss": 0.5074975490570068, "eval_runtime": 2.3982, "eval_samples_per_second": 48.37, "step": 544 }, { "epoch": 35.0, "learning_rate": 0.00027499999999999996, "loss": 0.113, "step": 560 }, { "epoch": 35.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 67.0757956863876, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7272727272727273, \"recall\": 0.6666666666666666, \"f1-score\": 0.6956521739130435, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6857142857142857, \"recall\": 0.7741935483870968, \"f1-score\": 0.7272727272727272, \"support\": 31.0}, \"evaluation\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.7027027027027027, \"f1-score\": 0.7027027027027027, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"request\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.8116834366834367, \"recall\": 0.7644731424973361, \"f1-score\": 0.7833908982721912, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7446932377966862, \"recall\": 0.7327586206896551, \"f1-score\": 0.7354278794471526, \"support\": 116.0}}", "eval_f1_macro": 48.208670662904076, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.54278794471526, "eval_loss": 0.47861307859420776, "eval_runtime": 2.0863, "eval_samples_per_second": 55.601, "step": 560 }, { "epoch": 36.0, "learning_rate": 0.00027333333333333333, "loss": 0.0821, "step": 576 }, { "epoch": 36.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 65.40700236389213, "eval_classification_report": "{\"accusation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8333333333333334, \"f1-score\": 0.7407407407407408, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7666666666666667, \"recall\": 0.6216216216216216, \"f1-score\": 0.6865671641791045, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.5333333333333333, \"recall\": 0.8888888888888888, \"f1-score\": 0.6666666666666667, \"support\": 9.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7447649572649573, \"recall\": 0.8007089751041363, \"f1-score\": 0.7626605641383726, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7350243147656942, \"recall\": 0.7155172413793104, \"f1-score\": 0.715916033865759, \"support\": 116.0}}", "eval_f1_macro": 46.93295779313061, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.59160338657588, "eval_loss": 0.542853057384491, "eval_runtime": 2.0928, "eval_samples_per_second": 55.428, "step": 576 }, { "epoch": 37.0, "learning_rate": 0.00027166666666666664, "loss": 0.0757, "step": 592 }, { "epoch": 37.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 65.66949561846265, "eval_classification_report": "{\"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.696969696969697, \"recall\": 0.7419354838709677, \"f1-score\": 0.71875, \"support\": 31.0}, \"evaluation\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.5675675675675675, \"f1-score\": 0.6461538461538462, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5454545454545454, \"recall\": 0.6666666666666666, \"f1-score\": 0.6, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7844065656565656, \"recall\": 0.7921601036520391, \"f1-score\": 0.7787362300689507, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7368730407523512, \"recall\": 0.7155172413793104, \"f1-score\": 0.7165230465528387, \"support\": 116.0}}", "eval_f1_macro": 47.922229542704656, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.65230465528387, "eval_loss": 0.6507958769798279, "eval_runtime": 1.9704, "eval_samples_per_second": 58.871, "step": 592 }, { "epoch": 38.0, "learning_rate": 0.00027, "loss": 0.0588, "step": 608 }, { "epoch": 38.0, "eval_accuracy": 69.82758620689656, "eval_average_metrics": 63.73703365588325, "eval_classification_report": "{\"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.631578947368421, \"recall\": 0.7741935483870968, \"f1-score\": 0.6956521739130435, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7419354838709677, \"recall\": 0.6216216216216216, \"f1-score\": 0.676470588235294, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5, \"recall\": 0.3333333333333333, \"f1-score\": 0.4, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7638768039049236, \"recall\": 0.7369768962510898, \"f1-score\": 0.7422196796338673, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7099617264797143, \"recall\": 0.6982758620689655, \"f1-score\": 0.696179050015019, \"support\": 116.0}}", "eval_f1_macro": 45.67505720823799, "eval_f1_micro": 69.82758620689656, "eval_f1_weighted": 69.6179050015019, "eval_loss": 0.6356076002120972, "eval_runtime": 2.1143, "eval_samples_per_second": 54.865, "step": 608 }, { "epoch": 39.0, "learning_rate": 0.0002683333333333333, "loss": 0.109, "step": 624 }, { "epoch": 39.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 67.851106421197, "eval_classification_report": "{\"accusation\": {\"precision\": 0.5238095238095238, \"recall\": 0.9166666666666666, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7857142857142857, \"recall\": 0.7096774193548387, \"f1-score\": 0.7457627118644068, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.8, \"recall\": 0.6486486486486487, \"f1-score\": 0.7164179104477612, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5454545454545454, \"recall\": 0.6666666666666666, \"f1-score\": 0.6, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7929834054834055, \"recall\": 0.8086796473893247, \"f1-score\": 0.7896230372710594, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7702754142409315, \"recall\": 0.7413793103448276, \"f1-score\": 0.7453637670683421, \"support\": 116.0}}", "eval_f1_macro": 48.592186908988275, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.5363767068342, "eval_loss": 0.6028116345405579, "eval_runtime": 2.1145, "eval_samples_per_second": 54.859, "step": 624 }, { "epoch": 40.0, "learning_rate": 0.0002666666666666666, "loss": 0.0434, "step": 640 }, { "epoch": 40.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 67.73082981916082, "eval_classification_report": "{\"accusation\": {\"precision\": 0.6923076923076923, \"recall\": 0.75, \"f1-score\": 0.7199999999999999, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.7096774193548387, \"f1-score\": 0.7096774193548389, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.7027027027027027, \"f1-score\": 0.7323943661971832, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5454545454545454, \"recall\": 0.6666666666666666, \"f1-score\": 0.6, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7822721606877204, \"recall\": 0.7946030708127482, \"f1-score\": 0.7866332421998508, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7462762389740077, \"recall\": 0.7413793103448276, \"f1-score\": 0.7423925768768693, \"support\": 116.0}}", "eval_f1_macro": 48.40819951999081, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.23925768768693, "eval_loss": 0.5737775564193726, "eval_runtime": 1.9992, "eval_samples_per_second": 58.024, "step": 640 }, { "epoch": 41.0, "learning_rate": 0.000265, "loss": 0.0482, "step": 656 }, { "epoch": 41.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 66.19735127899169, "eval_classification_report": "{\"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7058823529411765, \"recall\": 0.7741935483870968, \"f1-score\": 0.7384615384615385, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7666666666666667, \"recall\": 0.6216216216216216, \"f1-score\": 0.6865671641791045, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5, \"recall\": 0.5555555555555556, \"f1-score\": 0.5263157894736842, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7794730392156863, \"recall\": 0.7751713406955343, \"f1-score\": 0.7704721495741973, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7403693373901284, \"recall\": 0.7241379310344828, \"f1-score\": 0.7254814816604266, \"support\": 116.0}}", "eval_f1_macro": 47.413670743027524, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.54814816604267, "eval_loss": 0.6634430885314941, "eval_runtime": 2.0377, "eval_samples_per_second": 56.927, "step": 656 }, { "epoch": 42.0, "learning_rate": 0.0002633333333333333, "loss": 0.0329, "step": 672 }, { "epoch": 42.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 66.16273743305499, "eval_classification_report": "{\"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.75, \"f1-score\": 0.7058823529411765, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.6451612903225806, \"f1-score\": 0.689655172413793, \"support\": 31.0}, \"evaluation\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.7027027027027027, \"f1-score\": 0.7323943661971832, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5454545454545454, \"recall\": 0.6666666666666666, \"f1-score\": 0.6, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7632253911665676, \"recall\": 0.7830663324614937, \"f1-score\": 0.7668365408904801, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7392959343263603, \"recall\": 0.7241379310344828, \"f1-score\": 0.7263342254744771, \"support\": 116.0}}", "eval_f1_macro": 47.1899409778757, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.63342254744771, "eval_loss": 0.6547279357910156, "eval_runtime": 2.1903, "eval_samples_per_second": 52.96, "step": 672 }, { "epoch": 43.0, "learning_rate": 0.00026166666666666667, "loss": 0.0316, "step": 688 }, { "epoch": 43.0, "eval_accuracy": 69.82758620689656, "eval_average_metrics": 64.21844203108832, "eval_classification_report": "{\"accusation\": {\"precision\": 0.4583333333333333, \"recall\": 0.9166666666666666, \"f1-score\": 0.611111111111111, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.8571428571428571, \"recall\": 0.5806451612903226, \"f1-score\": 0.6923076923076923, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.71875, \"recall\": 0.6216216216216216, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5384615384615384, \"recall\": 0.7777777777777778, \"f1-score\": 0.6363636363636364, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7718700570263071, \"recall\": 0.7891722367528818, \"f1-score\": 0.7608901719195837, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7494119565779048, \"recall\": 0.6982758620689655, \"f1-score\": 0.7039458513089345, \"support\": 116.0}}", "eval_f1_macro": 46.824010579666684, "eval_f1_micro": 69.82758620689656, "eval_f1_weighted": 70.39458513089345, "eval_loss": 0.7738772034645081, "eval_runtime": 2.0549, "eval_samples_per_second": 56.451, "step": 688 }, { "epoch": 44.0, "learning_rate": 0.00026, "loss": 0.0505, "step": 704 }, { "epoch": 44.0, "eval_accuracy": 69.82758620689656, "eval_average_metrics": 63.97464242891904, "eval_classification_report": "{\"accusation\": {\"precision\": 0.5333333333333333, \"recall\": 0.6666666666666666, \"f1-score\": 0.5925925925925926, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.75, \"f1-score\": 0.7058823529411765, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7096774193548387, \"recall\": 0.5945945945945946, \"f1-score\": 0.6470588235294118, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.739340990550668, \"recall\": 0.7805295214569408, \"f1-score\": 0.7567426750203485, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7020418309906631, \"recall\": 0.6982758620689655, \"f1-score\": 0.6967461730063084, \"support\": 116.0}}", "eval_f1_macro": 46.56878000125221, "eval_f1_micro": 69.82758620689656, "eval_f1_weighted": 69.67461730063084, "eval_loss": 0.6857355833053589, "eval_runtime": 2.1516, "eval_samples_per_second": 53.912, "step": 704 }, { "epoch": 45.0, "learning_rate": 0.00025833333333333334, "loss": 0.0194, "step": 720 }, { "epoch": 45.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 67.93408505993239, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7272727272727273, \"recall\": 0.6666666666666666, \"f1-score\": 0.6956521739130435, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.7096774193548387, \"f1-score\": 0.7213114754098361, \"support\": 31.0}, \"evaluation\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.675, \"recall\": 0.7297297297297297, \"f1-score\": 0.7012987012987014, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.8203598484848486, \"recall\": 0.7875647825244599, \"f1-score\": 0.7984059141284494, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7524229362591431, \"recall\": 0.7413793103448276, \"f1-score\": 0.7432780653209022, \"support\": 116.0}}", "eval_f1_macro": 49.132671638673806, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.32780653209022, "eval_loss": 0.6950328946113586, "eval_runtime": 2.1933, "eval_samples_per_second": 52.888, "step": 720 }, { "epoch": 46.0, "learning_rate": 0.00025666666666666665, "loss": 0.0199, "step": 736 }, { "epoch": 46.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 64.64614471265455, "eval_classification_report": "{\"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.7096774193548387, \"f1-score\": 0.7096774193548389, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.5675675675675675, \"f1-score\": 0.6461538461538462, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7441835336285052, \"recall\": 0.788127845587523, \"f1-score\": 0.7597323795105688, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7173174442190671, \"recall\": 0.7068965517241379, \"f1-score\": 0.7045250668975559, \"support\": 116.0}}", "eval_f1_macro": 46.752761816035, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.4525066897556, "eval_loss": 0.8003849983215332, "eval_runtime": 2.0532, "eval_samples_per_second": 56.498, "step": 736 }, { "epoch": 47.0, "learning_rate": 0.00025499999999999996, "loss": 0.0231, "step": 752 }, { "epoch": 47.0, "eval_accuracy": 65.51724137931035, "eval_average_metrics": 60.04096842765952, "eval_classification_report": "{\"accusation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.5641025641025641, \"recall\": 0.7096774193548387, \"f1-score\": 0.6285714285714286, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.72, \"recall\": 0.4864864864864865, \"f1-score\": 0.5806451612903226, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5714285714285714, \"recall\": 0.4444444444444444, \"f1-score\": 0.5, \"support\": 9.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.8888888888888888, \"f1-score\": 0.6956521739130435, \"support\": 9.0}, \"accuracy\": 0.6551724137931034, \"macro avg\": {\"precision\": 0.7158699633699634, \"recall\": 0.7397982660079434, \"f1-score\": 0.7183717533665861, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.6690766704559807, \"recall\": 0.6551724137931034, \"f1-score\": 0.6492189843715057, \"support\": 116.0}}", "eval_f1_macro": 44.207492514866836, "eval_f1_micro": 65.51724137931035, "eval_f1_weighted": 64.92189843715057, "eval_loss": 0.827684760093689, "eval_runtime": 2.1386, "eval_samples_per_second": 54.241, "step": 752 }, { "epoch": 48.0, "learning_rate": 0.00025333333333333333, "loss": 0.0259, "step": 768 }, { "epoch": 48.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 67.95719742042525, "eval_classification_report": "{\"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"evaluation\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.7027027027027027, \"f1-score\": 0.7123287671232876, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.811489898989899, \"recall\": 0.8009874794148988, \"f1-score\": 0.7971726856406487, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7623040752351098, \"recall\": 0.7413793103448276, \"f1-score\": 0.744961469579263, \"support\": 116.0}}", "eval_f1_macro": 49.05678065480915, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.4961469579263, "eval_loss": 0.7849622964859009, "eval_runtime": 2.115, "eval_samples_per_second": 54.847, "step": 768 }, { "epoch": 49.0, "learning_rate": 0.00025166666666666664, "loss": 0.0132, "step": 784 }, { "epoch": 49.0, "eval_accuracy": 69.82758620689656, "eval_average_metrics": 63.94372313914248, "eval_classification_report": "{\"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7272727272727273, \"recall\": 0.7741935483870968, \"f1-score\": 0.7500000000000001, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6896551724137931, \"recall\": 0.5405405405405406, \"f1-score\": 0.6060606060606061, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7142857142857143, \"recall\": 0.5555555555555556, \"f1-score\": 0.6250000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7684892523538168, \"recall\": 0.7650362055603992, \"f1-score\": 0.7542708615704347, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7174182617713971, \"recall\": 0.6982758620689655, \"f1-score\": 0.6970305173844238, \"support\": 116.0}}", "eval_f1_macro": 46.41666840433444, "eval_f1_micro": 69.82758620689656, "eval_f1_weighted": 69.70305173844238, "eval_loss": 0.8554142117500305, "eval_runtime": 2.2913, "eval_samples_per_second": 50.627, "step": 784 }, { "epoch": 50.0, "learning_rate": 0.00025, "loss": 0.0138, "step": 800 }, { "epoch": 50.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 66.15480831429107, "eval_classification_report": "{\"accusation\": {\"precision\": 0.6923076923076923, \"recall\": 0.75, \"f1-score\": 0.7199999999999999, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.71875, \"recall\": 0.7419354838709677, \"f1-score\": 0.7301587301587302, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7586206896551724, \"recall\": 0.5945945945945946, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7575394403250006, \"recall\": 0.7990107042526398, \"f1-score\": 0.7728733766233766, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7321003628505858, \"recall\": 0.7241379310344828, \"f1-score\": 0.7223020848882918, \"support\": 116.0}}", "eval_f1_macro": 47.561438561438564, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.23020848882918, "eval_loss": 0.7713599801063538, "eval_runtime": 2.2362, "eval_samples_per_second": 51.874, "step": 800 }, { "epoch": 51.0, "learning_rate": 0.0002483333333333333, "loss": 0.0306, "step": 816 }, { "epoch": 51.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 66.99442064338926, "eval_classification_report": "{\"accusation\": {\"precision\": 0.5789473684210527, \"recall\": 0.9166666666666666, \"f1-score\": 0.7096774193548387, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.84, \"recall\": 0.6774193548387096, \"f1-score\": 0.75, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7419354838709677, \"recall\": 0.6216216216216216, \"f1-score\": 0.676470588235294, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7696143247904708, \"recall\": 0.815157899835319, \"f1-score\": 0.7817010406313063, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.754707967081218, \"recall\": 0.7327586206896551, \"f1-score\": 0.7332127901216103, \"support\": 116.0}}", "eval_f1_macro": 48.104679423465, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.32127901216103, "eval_loss": 0.8219650387763977, "eval_runtime": 2.1993, "eval_samples_per_second": 52.743, "step": 816 }, { "epoch": 52.0, "learning_rate": 0.0002466666666666666, "loss": 0.0168, "step": 832 }, { "epoch": 52.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 65.40322943340185, "eval_classification_report": "{\"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7586206896551724, \"recall\": 0.7096774193548387, \"f1-score\": 0.7333333333333333, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7586206896551724, \"recall\": 0.5945945945945946, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7462973149934357, \"recall\": 0.7949784461881235, \"f1-score\": 0.763627344877345, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7293270789852241, \"recall\": 0.7155172413793104, \"f1-score\": 0.7151701746529333, \"support\": 116.0}}", "eval_f1_macro": 46.992451992452, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.51701746529334, "eval_loss": 0.865384042263031, "eval_runtime": 2.3172, "eval_samples_per_second": 50.06, "step": 832 }, { "epoch": 53.0, "learning_rate": 0.000245, "loss": 0.0092, "step": 848 }, { "epoch": 53.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 64.62124905081899, "eval_classification_report": "{\"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7586206896551724, \"recall\": 0.7096774193548387, \"f1-score\": 0.7333333333333333, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7333333333333333, \"recall\": 0.5945945945945946, \"f1-score\": 0.6567164179104478, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7414002843420947, \"recall\": 0.7810895572992347, \"f1-score\": 0.7542136291422947, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.720183698089465, \"recall\": 0.7068965517241379, \"f1-score\": 0.7069253944969183, \"support\": 116.0}}", "eval_f1_macro": 46.41314640875659, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.69253944969182, "eval_loss": 0.9035179018974304, "eval_runtime": 2.2008, "eval_samples_per_second": 52.708, "step": 848 }, { "epoch": 54.0, "learning_rate": 0.0002433333333333333, "loss": 0.0081, "step": 864 }, { "epoch": 54.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 64.59618900858153, "eval_classification_report": "{\"accusation\": {\"precision\": 0.6923076923076923, \"recall\": 0.75, \"f1-score\": 0.7199999999999999, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7142857142857143, \"recall\": 0.6451612903225806, \"f1-score\": 0.6779661016949152, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6756756756756757, \"recall\": 0.6756756756756757, \"f1-score\": 0.6756756756756757, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7142857142857143, \"recall\": 0.5555555555555556, \"f1-score\": 0.6250000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.751652682902683, \"recall\": 0.7692712874164487, \"f1-score\": 0.756581475304156, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7075786282682834, \"recall\": 0.7068965517241379, \"f1-score\": 0.7044658567078128, \"support\": 116.0}}", "eval_f1_macro": 46.55886001871728, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.44658567078127, "eval_loss": 0.8619075417518616, "eval_runtime": 2.0413, "eval_samples_per_second": 56.826, "step": 864 }, { "epoch": 55.0, "learning_rate": 0.00024166666666666664, "loss": 0.0174, "step": 880 }, { "epoch": 55.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 66.96407889073964, "eval_classification_report": "{\"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6756756756756757, \"recall\": 0.8064516129032258, \"f1-score\": 0.7352941176470588, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7931034482758621, \"recall\": 0.6216216216216216, \"f1-score\": 0.696969696969697, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7142857142857143, \"recall\": 0.5555555555555556, \"f1-score\": 0.6250000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7952158067998585, \"recall\": 0.7826758209822725, \"f1-score\": 0.7807690879382057, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7501275819498173, \"recall\": 0.7327586206896551, \"f1-score\": 0.7325726293652257, \"support\": 116.0}}", "eval_f1_macro": 48.04732848850496, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.25726293652257, "eval_loss": 0.8389707803726196, "eval_runtime": 2.2977, "eval_samples_per_second": 50.486, "step": 880 }, { "epoch": 56.0, "learning_rate": 0.00023999999999999998, "loss": 0.0107, "step": 896 }, { "epoch": 56.0, "eval_accuracy": 69.82758620689656, "eval_average_metrics": 63.90321776361796, "eval_classification_report": "{\"accusation\": {\"precision\": 0.8181818181818182, \"recall\": 0.75, \"f1-score\": 0.7826086956521738, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6363636363636364, \"recall\": 0.6774193548387096, \"f1-score\": 0.65625, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.71875, \"recall\": 0.6216216216216216, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5555555555555556, \"recall\": 0.5555555555555556, \"f1-score\": 0.5555555555555556, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7431897095959596, \"recall\": 0.766546788724208, \"f1-score\": 0.7518863678671316, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.701200039184953, \"recall\": 0.6982758620689655, \"f1-score\": 0.6968776831039368, \"support\": 116.0}}", "eval_f1_macro": 46.26993033028503, "eval_f1_micro": 69.82758620689656, "eval_f1_weighted": 69.68776831039368, "eval_loss": 0.8277653455734253, "eval_runtime": 2.0327, "eval_samples_per_second": 57.068, "step": 896 }, { "epoch": 57.0, "learning_rate": 0.0002383333333333333, "loss": 0.0078, "step": 912 }, { "epoch": 57.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 66.14248466318766, "eval_classification_report": "{\"accusation\": {\"precision\": 0.5, \"recall\": 0.9166666666666666, \"f1-score\": 0.6470588235294118, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.8181818181818182, \"recall\": 0.5806451612903226, \"f1-score\": 0.679245283018868, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.7297297297297297, \"f1-score\": 0.7397260273972601, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.625, \"recall\": 0.5555555555555556, \"f1-score\": 0.5882352941176471, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7708107864357865, \"recall\": 0.7887968613775065, \"f1-score\": 0.7678942896190095, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7512175324675325, \"recall\": 0.7241379310344828, \"f1-score\": 0.7248731923853042, \"support\": 116.0}}", "eval_f1_macro": 47.25503320732366, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.48731923853042, "eval_loss": 0.9054758548736572, "eval_runtime": 2.224, "eval_samples_per_second": 52.158, "step": 912 }, { "epoch": 58.0, "learning_rate": 0.00023666666666666663, "loss": 0.0117, "step": 928 }, { "epoch": 58.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 64.79079973704312, "eval_classification_report": "{\"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.76, \"recall\": 0.6129032258064516, \"f1-score\": 0.6785714285714285, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7419354838709677, \"recall\": 0.6216216216216216, \"f1-score\": 0.676470588235294, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7532539088905812, \"recall\": 0.7966767170396203, \"f1-score\": 0.7645456154113649, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.728019650860234, \"recall\": 0.7068965517241379, \"f1-score\": 0.7073492765495317, \"support\": 116.0}}", "eval_f1_macro": 47.048960948391695, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.73492765495317, "eval_loss": 0.8908520340919495, "eval_runtime": 2.1575, "eval_samples_per_second": 53.766, "step": 928 }, { "epoch": 59.0, "learning_rate": 0.00023499999999999997, "loss": 0.0229, "step": 944 }, { "epoch": 59.0, "eval_accuracy": 68.10344827586206, "eval_average_metrics": 62.18145432466472, "eval_classification_report": "{\"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6111111111111112, \"recall\": 0.7096774193548387, \"f1-score\": 0.6567164179104478, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7777777777777778, \"recall\": 0.5675675675675675, \"f1-score\": 0.65625, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5, \"recall\": 0.3333333333333333, \"f1-score\": 0.4, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.6810344827586207, \"macro avg\": {\"precision\": 0.7321428571428571, \"recall\": 0.7464611789208564, \"f1-score\": 0.7288033419213458, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.6960864805692392, \"recall\": 0.6810344827586207, \"f1-score\": 0.676694843210058, \"support\": 116.0}}", "eval_f1_macro": 44.849436425928964, "eval_f1_micro": 68.10344827586206, "eval_f1_weighted": 67.6694843210058, "eval_loss": 0.9175754189491272, "eval_runtime": 2.3118, "eval_samples_per_second": 50.177, "step": 944 }, { "epoch": 60.0, "learning_rate": 0.0002333333333333333, "loss": 0.0061, "step": 960 }, { "epoch": 60.0, "eval_accuracy": 68.96551724137932, "eval_average_metrics": 63.110820405911554, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7272727272727273, \"recall\": 0.6666666666666666, \"f1-score\": 0.6956521739130435, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6470588235294118, \"recall\": 0.7096774193548387, \"f1-score\": 0.6769230769230768, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7096774193548387, \"recall\": 0.5945945945945946, \"f1-score\": 0.6470588235294118, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5555555555555556, \"recall\": 0.5555555555555556, \"f1-score\": 0.5555555555555556, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.6896551724137931, \"macro avg\": {\"precision\": 0.7367297537482547, \"recall\": 0.7567840017436792, \"f1-score\": 0.742668905760338, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.6946777785454746, \"recall\": 0.6896551724137931, \"f1-score\": 0.6880954524794368, \"support\": 116.0}}", "eval_f1_macro": 45.70270189294388, "eval_f1_micro": 68.96551724137932, "eval_f1_weighted": 68.80954524794369, "eval_loss": 0.8836420178413391, "eval_runtime": 2.1103, "eval_samples_per_second": 54.968, "step": 960 }, { "epoch": 61.0, "learning_rate": 0.00023166666666666667, "loss": 0.0147, "step": 976 }, { "epoch": 61.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 66.25025679739119, "eval_classification_report": "{\"accusation\": {\"precision\": 0.6153846153846154, \"recall\": 0.6666666666666666, \"f1-score\": 0.64, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6756756756756757, \"recall\": 0.6756756756756757, \"f1-score\": 0.6756756756756757, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"request\": {\"precision\": 0.6428571428571429, \"recall\": 1.0, \"f1-score\": 0.782608695652174, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7759658697158698, \"recall\": 0.7906646565920759, \"f1-score\": 0.7771491930240497, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7317981348153763, \"recall\": 0.7241379310344828, \"f1-score\": 0.7234887525811131, \"support\": 116.0}}", "eval_f1_macro": 47.824565724556905, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.3488752581113, "eval_loss": 0.8437548875808716, "eval_runtime": 2.2297, "eval_samples_per_second": 52.024, "step": 976 }, { "epoch": 62.0, "learning_rate": 0.00023, "loss": 0.0049, "step": 992 }, { "epoch": 62.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 66.96754083591335, "eval_classification_report": "{\"accusation\": {\"precision\": 0.6470588235294118, \"recall\": 0.9166666666666666, \"f1-score\": 0.7586206896551724, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7857142857142857, \"recall\": 0.7096774193548387, \"f1-score\": 0.7457627118644068, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7096774193548387, \"recall\": 0.5945945945945946, \"f1-score\": 0.6470588235294118, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6428571428571429, \"recall\": 1.0, \"f1-score\": 0.782608695652174, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7747507605192614, \"recall\": 0.815811779521457, \"f1-score\": 0.78620080953209, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7415756068834654, \"recall\": 0.7327586206896551, \"f1-score\": 0.7293685092682456, \"support\": 116.0}}", "eval_f1_macro": 48.381588278897844, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 72.93685092682455, "eval_loss": 0.8428878784179688, "eval_runtime": 2.0652, "eval_samples_per_second": 56.169, "step": 992 }, { "epoch": 63.0, "learning_rate": 0.0002283333333333333, "loss": 0.0022, "step": 1008 }, { "epoch": 63.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 66.1127357566638, "eval_classification_report": "{\"accusation\": {\"precision\": 0.6111111111111112, \"recall\": 0.9166666666666666, \"f1-score\": 0.7333333333333334, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.8, \"recall\": 0.6451612903225806, \"f1-score\": 0.7142857142857142, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6486486486486487, \"f1-score\": 0.6956521739130435, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5454545454545454, \"recall\": 0.6666666666666666, \"f1-score\": 0.6, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7515151515151515, \"recall\": 0.8006151312602925, \"f1-score\": 0.7680081090407177, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7395898641588295, \"recall\": 0.7241379310344828, \"f1-score\": 0.7236131934032982, \"support\": 116.0}}", "eval_f1_macro": 47.262037479428784, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.36131934032983, "eval_loss": 0.8622753024101257, "eval_runtime": 2.2456, "eval_samples_per_second": 51.658, "step": 1008 }, { "epoch": 64.0, "learning_rate": 0.00022666666666666663, "loss": 0.002, "step": 1024 }, { "epoch": 64.0, "eval_accuracy": 75.86206896551724, "eval_average_metrics": 69.18501875699872, "eval_classification_report": "{\"accusation\": {\"precision\": 0.6875, \"recall\": 0.9166666666666666, \"f1-score\": 0.7857142857142857, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.8, \"recall\": 0.6451612903225806, \"f1-score\": 0.7142857142857142, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.7297297297297297, \"f1-score\": 0.7397260273972601, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7586206896551724, \"macro avg\": {\"precision\": 0.787957702020202, \"recall\": 0.8246391552843165, \"f1-score\": 0.8019321276936633, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7641849529780564, \"recall\": 0.7586206896551724, \"f1-score\": 0.7566626770042725, \"support\": 116.0}}", "eval_f1_macro": 49.34966939653312, "eval_f1_micro": 75.86206896551724, "eval_f1_weighted": 75.66626770042724, "eval_loss": 0.8323994874954224, "eval_runtime": 2.1112, "eval_samples_per_second": 54.945, "step": 1024 }, { "epoch": 65.0, "learning_rate": 0.000225, "loss": 0.0043, "step": 1040 }, { "epoch": 65.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 65.35337967908377, "eval_classification_report": "{\"accusation\": {\"precision\": 0.6111111111111112, \"recall\": 0.9166666666666666, \"f1-score\": 0.7333333333333334, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.75, \"f1-score\": 0.7058823529411765, \"support\": 8.0}, \"challenge\": {\"precision\": 0.8571428571428571, \"recall\": 0.5806451612903226, \"f1-score\": 0.6923076923076923, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7419354838709677, \"recall\": 0.6216216216216216, \"f1-score\": 0.676470588235294, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6, \"recall\": 1.0, \"f1-score\": 0.7499999999999999, \"support\": 9.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7436347926267282, \"recall\": 0.8169500145306596, \"f1-score\": 0.7641936902966314, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7443078288044918, \"recall\": 0.7155172413793104, \"f1-score\": 0.7128276642221876, \"support\": 116.0}}", "eval_f1_macro": 47.02730401825425, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.28276642221876, "eval_loss": 1.015073299407959, "eval_runtime": 2.0888, "eval_samples_per_second": 55.535, "step": 1040 }, { "epoch": 66.0, "learning_rate": 0.00022333333333333333, "loss": 0.0134, "step": 1056 }, { "epoch": 66.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 66.25018013630203, "eval_classification_report": "{\"accusation\": {\"precision\": 0.55, \"recall\": 0.9166666666666666, \"f1-score\": 0.6874999999999999, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.7096774193548387, \"f1-score\": 0.7213114754098361, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7272727272727273, \"recall\": 0.6486486486486487, \"f1-score\": 0.6857142857142857, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7142857142857143, \"recall\": 0.5555555555555556, \"f1-score\": 0.6250000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7908955627705627, \"recall\": 0.7809018696115471, \"f1-score\": 0.7750247537539606, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7421676742797432, \"recall\": 0.7241379310344828, \"f1-score\": 0.7247930333806785, \"support\": 116.0}}", "eval_f1_macro": 47.693831000243726, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.47930333806785, "eval_loss": 0.8981208205223083, "eval_runtime": 2.2335, "eval_samples_per_second": 51.937, "step": 1056 }, { "epoch": 67.0, "learning_rate": 0.00022166666666666667, "loss": 0.0019, "step": 1072 }, { "epoch": 67.0, "eval_accuracy": 69.82758620689656, "eval_average_metrics": 64.00331673049328, "eval_classification_report": "{\"accusation\": {\"precision\": 0.6875, \"recall\": 0.9166666666666666, \"f1-score\": 0.7857142857142857, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.7096774193548387, \"f1-score\": 0.6875, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.5405405405405406, \"f1-score\": 0.6153846153846154, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5, \"recall\": 0.5555555555555556, \"f1-score\": 0.5263157894736842, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7590907356532357, \"recall\": 0.7812772449869224, \"f1-score\": 0.7610273954846323, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7102043071870658, \"recall\": 0.6982758620689655, \"f1-score\": 0.6952563940143341, \"support\": 116.0}}", "eval_f1_macro": 46.8324551067466, "eval_f1_micro": 69.82758620689656, "eval_f1_weighted": 69.52563940143341, "eval_loss": 0.871522068977356, "eval_runtime": 2.0887, "eval_samples_per_second": 55.536, "step": 1072 }, { "epoch": 68.0, "learning_rate": 0.00021999999999999995, "loss": 0.0042, "step": 1088 }, { "epoch": 68.0, "eval_accuracy": 68.10344827586206, "eval_average_metrics": 62.42749733192013, "eval_classification_report": "{\"accusation\": {\"precision\": 0.55, \"recall\": 0.9166666666666666, \"f1-score\": 0.6874999999999999, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7307692307692307, \"recall\": 0.6129032258064516, \"f1-score\": 0.6666666666666667, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7307692307692307, \"recall\": 0.5135135135135135, \"f1-score\": 0.6031746031746031, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.8888888888888888, \"f1-score\": 0.6956521739130435, \"support\": 9.0}, \"accuracy\": 0.6810344827586207, \"macro avg\": {\"precision\": 0.7256486568986569, \"recall\": 0.7935798701927734, \"f1-score\": 0.7448186249137336, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.704182455475559, \"recall\": 0.6810344827586207, \"f1-score\": 0.6766810047357272, \"support\": 116.0}}", "eval_f1_macro": 45.83499230238361, "eval_f1_micro": 68.10344827586206, "eval_f1_weighted": 67.66810047357275, "eval_loss": 0.9840335249900818, "eval_runtime": 2.1773, "eval_samples_per_second": 53.276, "step": 1088 }, { "epoch": 69.0, "learning_rate": 0.0002183333333333333, "loss": 0.0116, "step": 1104 }, { "epoch": 69.0, "eval_accuracy": 75.0, "eval_average_metrics": 68.42151422003346, "eval_classification_report": "{\"accusation\": {\"precision\": 0.8181818181818182, \"recall\": 0.75, \"f1-score\": 0.7826086956521738, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7857142857142857, \"recall\": 0.7096774193548387, \"f1-score\": 0.7457627118644068, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6923076923076923, \"recall\": 0.7297297297297297, \"f1-score\": 0.7105263157894737, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.75, \"macro avg\": {\"precision\": 0.7904345654345655, \"recall\": 0.7979814491911266, \"f1-score\": 0.7921106674256406, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.752466929622102, \"recall\": 0.75, \"f1-score\": 0.7494078503855592, \"support\": 116.0}}", "eval_f1_macro": 48.74527184157789, "eval_f1_micro": 75.0, "eval_f1_weighted": 74.94078503855592, "eval_loss": 0.8145465850830078, "eval_runtime": 2.2025, "eval_samples_per_second": 52.668, "step": 1104 }, { "epoch": 70.0, "learning_rate": 0.00021666666666666666, "loss": 0.0124, "step": 1120 }, { "epoch": 70.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 64.76443464447134, "eval_classification_report": "{\"accusation\": {\"precision\": 0.5238095238095238, \"recall\": 0.9166666666666666, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.7096774193548387, \"f1-score\": 0.7213114754098361, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7096774193548387, \"recall\": 0.5945945945945946, \"f1-score\": 0.6470588235294118, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7142857142857143, \"recall\": 0.5555555555555556, \"f1-score\": 0.6250000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7778465821812596, \"recall\": 0.7741451128547903, \"f1-score\": 0.76282674955228, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7291438238254145, \"recall\": 0.7068965517241379, \"f1-score\": 0.7073524364522519, \"support\": 116.0}}", "eval_f1_macro": 46.94318458783261, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.73524364522518, "eval_loss": 0.97468101978302, "eval_runtime": 2.1281, "eval_samples_per_second": 54.508, "step": 1120 }, { "epoch": 71.0, "learning_rate": 0.000215, "loss": 0.0042, "step": 1136 }, { "epoch": 71.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 64.6756069152224, "eval_classification_report": "{\"accusation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8333333333333334, \"f1-score\": 0.7407407407407408, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7586206896551724, \"recall\": 0.7096774193548387, \"f1-score\": 0.7333333333333333, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7241379310344828, \"recall\": 0.5675675675675675, \"f1-score\": 0.6363636363636365, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5384615384615384, \"recall\": 0.7777777777777778, \"f1-score\": 0.6363636363636364, \"support\": 9.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.8888888888888888, \"f1-score\": 0.6956521739130435, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7489322817986611, \"recall\": 0.788127845587523, \"f1-score\": 0.7582478665598871, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7244094713253454, \"recall\": 0.7068965517241379, \"f1-score\": 0.7066171014314591, \"support\": 116.0}}", "eval_f1_macro": 46.66140717291613, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.66171014314591, "eval_loss": 0.9142933487892151, "eval_runtime": 2.1078, "eval_samples_per_second": 55.033, "step": 1136 }, { "epoch": 72.0, "learning_rate": 0.00021333333333333333, "loss": 0.0025, "step": 1152 }, { "epoch": 72.0, "eval_accuracy": 76.72413793103449, "eval_average_metrics": 70.1131066055635, "eval_classification_report": "{\"accusation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8333333333333334, \"f1-score\": 0.7407407407407408, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7931034482758621, \"recall\": 0.7419354838709677, \"f1-score\": 0.7666666666666667, \"support\": 31.0}, \"evaluation\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7297297297297297, \"recall\": 0.7297297297297297, \"f1-score\": 0.7297297297297297, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7672413793103449, \"macro avg\": {\"precision\": 0.8292430361395879, \"recall\": 0.8124303739223094, \"f1-score\": 0.8136203167453168, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7809155766944115, \"recall\": 0.7672413793103449, \"f1-score\": 0.7693520799124248, \"support\": 116.0}}", "eval_f1_macro": 50.06894256894257, "eval_f1_micro": 76.72413793103448, "eval_f1_weighted": 76.93520799124248, "eval_loss": 0.9065346121788025, "eval_runtime": 2.2637, "eval_samples_per_second": 51.244, "step": 1152 }, { "epoch": 73.0, "learning_rate": 0.00021166666666666667, "loss": 0.0118, "step": 1168 }, { "epoch": 73.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 67.77121142508423, "eval_classification_report": "{\"accusation\": {\"precision\": 0.5789473684210527, \"recall\": 0.9166666666666666, \"f1-score\": 0.7096774193548387, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.8076923076923077, \"recall\": 0.6774193548387096, \"f1-score\": 0.7368421052631579, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6756756756756757, \"f1-score\": 0.7142857142857142, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7902807170899276, \"recall\": 0.808025767703187, \"f1-score\": 0.7879227837145045, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.763951773423189, \"recall\": 0.7413793103448276, \"f1-score\": 0.7432142771047883, \"support\": 116.0}}", "eval_f1_macro": 48.487555920892575, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.32142771047884, "eval_loss": 1.009997844696045, "eval_runtime": 2.2483, "eval_samples_per_second": 51.595, "step": 1168 }, { "epoch": 74.0, "learning_rate": 0.00020999999999999998, "loss": 0.0026, "step": 1184 }, { "epoch": 74.0, "eval_accuracy": 76.72413793103449, "eval_average_metrics": 70.049578589584, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.7741935483870968, \"f1-score\": 0.7619047619047619, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6756756756756757, \"f1-score\": 0.7142857142857142, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7672413793103449, \"macro avg\": {\"precision\": 0.8173160173160172, \"recall\": 0.8235947641189576, \"f1-score\": 0.8147149548078341, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7732907896701001, \"recall\": 0.7672413793103449, \"f1-score\": 0.7661373358501574, \"support\": 116.0}}", "eval_f1_macro": 50.13630491125133, "eval_f1_micro": 76.72413793103448, "eval_f1_weighted": 76.61373358501574, "eval_loss": 0.8809800148010254, "eval_runtime": 2.1396, "eval_samples_per_second": 54.215, "step": 1184 }, { "epoch": 75.0, "learning_rate": 0.00020833333333333332, "loss": 0.008, "step": 1200 }, { "epoch": 75.0, "eval_accuracy": 76.72413793103449, "eval_average_metrics": 70.02564938184499, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.8518518518518519, \"recall\": 0.7419354838709677, \"f1-score\": 0.7931034482758621, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6756756756756757, \"f1-score\": 0.6944444444444444, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"request\": {\"precision\": 0.6428571428571429, \"recall\": 1.0, \"f1-score\": 0.782608695652174, \"support\": 9.0}, \"accuracy\": 0.7672413793103449, \"macro avg\": {\"precision\": 0.8030321530321531, \"recall\": 0.8334513949433304, \"f1-score\": 0.8119859478155659, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7759448724965966, \"recall\": 0.7672413793103449, \"f1-score\": 0.7668595564589151, \"support\": 116.0}}", "eval_f1_macro": 49.96836601941944, "eval_f1_micro": 76.72413793103448, "eval_f1_weighted": 76.68595564589151, "eval_loss": 0.9097422957420349, "eval_runtime": 2.1959, "eval_samples_per_second": 52.826, "step": 1200 }, { "epoch": 76.0, "learning_rate": 0.00020666666666666666, "loss": 0.0013, "step": 1216 }, { "epoch": 76.0, "eval_accuracy": 75.0, "eval_average_metrics": 68.43649890784998, "eval_classification_report": "{\"accusation\": {\"precision\": 0.6111111111111112, \"recall\": 0.9166666666666666, \"f1-score\": 0.7333333333333334, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.8148148148148148, \"recall\": 0.7096774193548387, \"f1-score\": 0.7586206896551724, \"support\": 31.0}, \"evaluation\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.7027027027027027, \"f1-score\": 0.7222222222222223, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.75, \"macro avg\": {\"precision\": 0.7949074074074074, \"recall\": 0.8015475152571927, \"f1-score\": 0.7897184592227696, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7666849115124977, \"recall\": 0.75, \"f1-score\": 0.751479366023064, \"support\": 116.0}}", "eval_f1_macro": 48.59805902909351, "eval_f1_micro": 75.0, "eval_f1_weighted": 75.1479366023064, "eval_loss": 0.9198105931282043, "eval_runtime": 2.12, "eval_samples_per_second": 54.717, "step": 1216 }, { "epoch": 77.0, "learning_rate": 0.000205, "loss": 0.0034, "step": 1232 }, { "epoch": 77.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 66.9730843369445, "eval_classification_report": "{\"accusation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.7419354838709677, \"f1-score\": 0.7419354838709677, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7419354838709677, \"recall\": 0.6216216216216216, \"f1-score\": 0.676470588235294, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7723275828114539, \"recall\": 0.802389082631018, \"f1-score\": 0.781404294366818, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.742217390507157, \"recall\": 0.7327586206896551, \"f1-score\": 0.7325419509496588, \"support\": 116.0}}", "eval_f1_macro": 48.086418114881106, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.25419509496588, "eval_loss": 0.9556060433387756, "eval_runtime": 2.2754, "eval_samples_per_second": 50.979, "step": 1232 }, { "epoch": 78.0, "learning_rate": 0.00020333333333333333, "loss": 0.0014, "step": 1248 }, { "epoch": 78.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 66.08748867920372, "eval_classification_report": "{\"accusation\": {\"precision\": 0.625, \"recall\": 0.8333333333333334, \"f1-score\": 0.7142857142857143, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.75, \"f1-score\": 0.7058823529411765, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7931034482758621, \"recall\": 0.6216216216216216, \"f1-score\": 0.696969696969697, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.743769341290893, \"recall\": 0.8047412331686525, \"f1-score\": 0.7655129972574192, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7420494273199386, \"recall\": 0.7241379310344828, \"f1-score\": 0.7241387637100023, \"support\": 116.0}}", "eval_f1_macro": 47.1084921389181, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.41387637100023, "eval_loss": 1.0091474056243896, "eval_runtime": 2.0996, "eval_samples_per_second": 55.249, "step": 1248 }, { "epoch": 79.0, "learning_rate": 0.00020166666666666667, "loss": 0.003, "step": 1264 }, { "epoch": 79.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 66.80793026911971, "eval_classification_report": "{\"accusation\": {\"precision\": 0.6428571428571429, \"recall\": 0.75, \"f1-score\": 0.6923076923076924, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7058823529411765, \"recall\": 0.7741935483870968, \"f1-score\": 0.7384615384615385, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.8214285714285714, \"recall\": 0.6216216216216216, \"f1-score\": 0.7076923076923075, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7594654528478058, \"recall\": 0.7925324518066452, \"f1-score\": 0.7713542670779513, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7447370327441322, \"recall\": 0.7327586206896551, \"f1-score\": 0.7321204204144313, \"support\": 116.0}}", "eval_f1_macro": 47.46795489710468, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.21204204144313, "eval_loss": 1.009720802307129, "eval_runtime": 2.21, "eval_samples_per_second": 52.488, "step": 1264 }, { "epoch": 80.0, "learning_rate": 0.00019999999999999998, "loss": 0.001, "step": 1280 }, { "epoch": 80.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 66.2278718761285, "eval_classification_report": "{\"accusation\": {\"precision\": 0.625, \"recall\": 0.8333333333333334, \"f1-score\": 0.7142857142857143, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7419354838709677, \"recall\": 0.6216216216216216, \"f1-score\": 0.676470588235294, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7611828695498051, \"recall\": 0.8047412331686525, \"f1-score\": 0.7747153145479717, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.738864900085701, \"recall\": 0.7241379310344828, \"f1-score\": 0.7240911271004996, \"support\": 116.0}}", "eval_f1_macro": 47.67478858756749, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.40911271004997, "eval_loss": 1.0751197338104248, "eval_runtime": 2.1451, "eval_samples_per_second": 54.077, "step": 1280 }, { "epoch": 81.0, "learning_rate": 0.00019833333333333332, "loss": 0.0034, "step": 1296 }, { "epoch": 81.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 64.67391316856995, "eval_classification_report": "{\"accusation\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6764705882352942, \"recall\": 0.7419354838709677, \"f1-score\": 0.7076923076923077, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7777777777777778, \"recall\": 0.5675675675675675, \"f1-score\": 0.65625, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7519580909286792, \"recall\": 0.7817434369853724, \"f1-score\": 0.7597936922361264, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7219307793851404, \"recall\": 0.7068965517241379, \"f1-score\": 0.7055980742261369, \"support\": 116.0}}", "eval_f1_macro": 46.756534906838546, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.55980742261369, "eval_loss": 1.0636241436004639, "eval_runtime": 2.1338, "eval_samples_per_second": 54.363, "step": 1296 }, { "epoch": 82.0, "learning_rate": 0.00019666666666666666, "loss": 0.0042, "step": 1312 }, { "epoch": 82.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 66.11615353179951, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.75, \"f1-score\": 0.7058823529411765, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7916666666666666, \"recall\": 0.6129032258064516, \"f1-score\": 0.6909090909090909, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.7027027027027027, \"f1-score\": 0.7027027027027027, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7566690066690066, \"recall\": 0.7929229632858665, \"f1-score\": 0.7692698861816509, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7331784594715629, \"recall\": 0.7241379310344828, \"f1-score\": 0.7229734261681524, \"support\": 116.0}}", "eval_f1_macro": 47.33968530348621, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.29734261681524, "eval_loss": 1.0073676109313965, "eval_runtime": 2.1573, "eval_samples_per_second": 53.77, "step": 1312 }, { "epoch": 83.0, "learning_rate": 0.000195, "loss": 0.0031, "step": 1328 }, { "epoch": 83.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 64.7525479876361, "eval_classification_report": "{\"accusation\": {\"precision\": 0.6923076923076923, \"recall\": 0.75, \"f1-score\": 0.7199999999999999, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.696969696969697, \"recall\": 0.7419354838709677, \"f1-score\": 0.71875, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7096774193548387, \"recall\": 0.5945945945945946, \"f1-score\": 0.6470588235294118, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5454545454545454, \"recall\": 0.6666666666666666, \"f1-score\": 0.6, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7668492461839236, \"recall\": 0.771232926474862, \"f1-score\": 0.7617192274637128, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7197790506675379, \"recall\": 0.7068965517241379, \"f1-score\": 0.7075585222333448, \"support\": 116.0}}", "eval_f1_macro": 46.87502938238233, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.75585222333449, "eval_loss": 1.0074830055236816, "eval_runtime": 2.3548, "eval_samples_per_second": 49.262, "step": 1328 }, { "epoch": 84.0, "learning_rate": 0.00019333333333333333, "loss": 0.0023, "step": 1344 }, { "epoch": 84.0, "eval_accuracy": 69.82758620689656, "eval_average_metrics": 63.96007923927792, "eval_classification_report": "{\"accusation\": {\"precision\": 0.625, \"recall\": 0.8333333333333334, \"f1-score\": 0.7142857142857143, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6764705882352942, \"recall\": 0.6216216216216216, \"f1-score\": 0.6478873239436619, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5, \"recall\": 0.5555555555555556, \"f1-score\": 0.5263157894736842, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7552319004524887, \"recall\": 0.7630745665019858, \"f1-score\": 0.7510372788328765, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7128695974410986, \"recall\": 0.6982758620689655, \"f1-score\": 0.6996746584591074, \"support\": 116.0}}", "eval_f1_macro": 46.21767869740778, "eval_f1_micro": 69.82758620689656, "eval_f1_weighted": 69.96746584591075, "eval_loss": 0.9842262864112854, "eval_runtime": 2.2804, "eval_samples_per_second": 50.867, "step": 1344 }, { "epoch": 85.0, "learning_rate": 0.00019166666666666665, "loss": 0.0013, "step": 1360 }, { "epoch": 85.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 65.34671058116388, "eval_classification_report": "{\"accusation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8333333333333334, \"f1-score\": 0.7407407407407408, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.875, \"f1-score\": 0.7368421052631579, \"support\": 8.0}, \"challenge\": {\"precision\": 0.76, \"recall\": 0.6129032258064516, \"f1-score\": 0.6785714285714285, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7666666666666667, \"recall\": 0.6216216216216216, \"f1-score\": 0.6865671641791045, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7349621212121212, \"recall\": 0.8123017170396203, \"f1-score\": 0.7641747663104694, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7281687565308255, \"recall\": 0.7155172413793104, \"f1-score\": 0.712572545835338, \"support\": 116.0}}", "eval_f1_macro": 47.02613946525966, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.25725458353381, "eval_loss": 0.9387104511260986, "eval_runtime": 2.2626, "eval_samples_per_second": 51.269, "step": 1360 }, { "epoch": 86.0, "learning_rate": 0.00018999999999999998, "loss": 0.0014, "step": 1376 }, { "epoch": 86.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 67.0203074746172, "eval_classification_report": "{\"accusation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8333333333333334, \"f1-score\": 0.7407407407407408, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7857142857142857, \"recall\": 0.7096774193548387, \"f1-score\": 0.7457627118644068, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7058823529411765, \"recall\": 0.6486486486486487, \"f1-score\": 0.676056338028169, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7808308929632459, \"recall\": 0.7982629807226582, \"f1-score\": 0.7830183865275773, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7437090039777667, \"recall\": 0.7327586206896551, \"f1-score\": 0.7334375889730225, \"support\": 116.0}}", "eval_f1_macro": 48.18574686323552, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.34375889730225, "eval_loss": 0.9032027125358582, "eval_runtime": 2.3121, "eval_samples_per_second": 50.17, "step": 1376 }, { "epoch": 87.0, "learning_rate": 0.00018833333333333332, "loss": 0.0005, "step": 1392 }, { "epoch": 87.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 67.12735809570236, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.8076923076923077, \"recall\": 0.6774193548387096, \"f1-score\": 0.7368421052631579, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6857142857142857, \"recall\": 0.6486486486486487, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7869123931623933, \"recall\": 0.8081196115470308, \"f1-score\": 0.7890888351414667, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7480153467222432, \"recall\": 0.7327586206896551, \"f1-score\": 0.7339839531309585, \"support\": 116.0}}", "eval_f1_macro": 48.559312931782564, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.39839531309586, "eval_loss": 0.9623767733573914, "eval_runtime": 2.1627, "eval_samples_per_second": 53.637, "step": 1392 }, { "epoch": 88.0, "learning_rate": 0.00018666666666666666, "loss": 0.0008, "step": 1408 }, { "epoch": 88.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 64.55394117974937, "eval_classification_report": "{\"accusation\": {\"precision\": 0.625, \"recall\": 0.8333333333333334, \"f1-score\": 0.7142857142857143, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.7, \"recall\": 0.875, \"f1-score\": 0.7777777777777777, \"support\": 8.0}, \"challenge\": {\"precision\": 0.696969696969697, \"recall\": 0.7419354838709677, \"f1-score\": 0.71875, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.8260869565217391, \"recall\": 0.5135135135135135, \"f1-score\": 0.6333333333333333, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5454545454545454, \"recall\": 0.6666666666666666, \"f1-score\": 0.6, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7372230879024357, \"recall\": 0.8010283468952824, \"f1-score\": 0.7575385551948053, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7303347452148053, \"recall\": 0.7068965517241379, \"f1-score\": 0.7021869713141263, \"support\": 116.0}}", "eval_f1_macro": 46.61775724275724, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.21869713141264, "eval_loss": 1.013329029083252, "eval_runtime": 2.0721, "eval_samples_per_second": 55.981, "step": 1408 }, { "epoch": 89.0, "learning_rate": 0.000185, "loss": 0.0006, "step": 1424 }, { "epoch": 89.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 67.65086863670423, "eval_classification_report": "{\"accusation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8333333333333334, \"f1-score\": 0.7407407407407408, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7857142857142857, \"recall\": 0.7096774193548387, \"f1-score\": 0.7457627118644068, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6756756756756757, \"f1-score\": 0.7142857142857142, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7740124458874459, \"recall\": 0.8016413591010365, \"f1-score\": 0.7816112509033746, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7531861098671444, \"recall\": 0.7413793103448276, \"f1-score\": 0.7422845857610528, \"support\": 116.0}}", "eval_f1_macro": 48.09915390174613, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.22845857610528, "eval_loss": 0.9453557729721069, "eval_runtime": 2.079, "eval_samples_per_second": 55.796, "step": 1424 }, { "epoch": 90.0, "learning_rate": 0.00018333333333333334, "loss": 0.0012, "step": 1440 }, { "epoch": 90.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 66.21906865002549, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7058823529411765, \"recall\": 0.6486486486486487, \"f1-score\": 0.676056338028169, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7661008313581843, \"recall\": 0.794230722658142, \"f1-score\": 0.77336173049969, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7355240615255829, \"recall\": 0.7241379310344828, \"f1-score\": 0.7245719728553219, \"support\": 116.0}}", "eval_f1_macro": 47.591491107673235, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.45719728553219, "eval_loss": 0.988357424736023, "eval_runtime": 2.409, "eval_samples_per_second": 48.153, "step": 1440 }, { "epoch": 91.0, "learning_rate": 0.00018166666666666665, "loss": 0.0058, "step": 1456 }, { "epoch": 91.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 64.72942944952882, "eval_classification_report": "{\"accusation\": {\"precision\": 0.625, \"recall\": 0.8333333333333334, \"f1-score\": 0.7142857142857143, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6875, \"recall\": 0.5945945945945946, \"f1-score\": 0.6376811594202898, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7526423229548229, \"recall\": 0.7874739659013852, \"f1-score\": 0.7616967013055733, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7204242716096164, \"recall\": 0.7068965517241379, \"f1-score\": 0.7066476429602169, \"support\": 116.0}}", "eval_f1_macro": 46.873643157266045, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.6647642960217, "eval_loss": 1.003797173500061, "eval_runtime": 2.3032, "eval_samples_per_second": 50.364, "step": 1456 }, { "epoch": 92.0, "learning_rate": 0.00017999999999999998, "loss": 0.0037, "step": 1472 }, { "epoch": 92.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 65.49007788853585, "eval_classification_report": "{\"accusation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8333333333333334, \"f1-score\": 0.7407407407407408, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.8076923076923077, \"recall\": 0.6774193548387096, \"f1-score\": 0.7368421052631579, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6764705882352942, \"recall\": 0.6216216216216216, \"f1-score\": 0.6478873239436619, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7602112960568843, \"recall\": 0.7908523442797636, \"f1-score\": 0.7678673719564575, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7292110146445846, \"recall\": 0.7155172413793104, \"f1-score\": 0.7160348654249928, \"support\": 116.0}}", "eval_f1_macro": 47.253376735782005, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.6034865424993, "eval_loss": 0.9861951470375061, "eval_runtime": 2.189, "eval_samples_per_second": 52.993, "step": 1472 }, { "epoch": 93.0, "learning_rate": 0.00017833333333333332, "loss": 0.0004, "step": 1488 }, { "epoch": 93.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 67.0835534197603, "eval_classification_report": "{\"accusation\": {\"precision\": 0.8333333333333334, \"recall\": 0.8333333333333334, \"f1-score\": 0.8333333333333334, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7586206896551724, \"recall\": 0.7096774193548387, \"f1-score\": 0.7333333333333333, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.696969696969697, \"recall\": 0.6216216216216216, \"f1-score\": 0.6571428571428571, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7792091768384872, \"recall\": 0.8087734912331687, \"f1-score\": 0.7883297258297258, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7409545220930478, \"recall\": 0.7327586206896551, \"f1-score\": 0.7326989102851171, \"support\": 116.0}}", "eval_f1_macro": 48.512598512598515, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.26989102851171, "eval_loss": 0.9672034978866577, "eval_runtime": 2.3052, "eval_samples_per_second": 50.322, "step": 1488 }, { "epoch": 94.0, "learning_rate": 0.00017666666666666666, "loss": 0.0047, "step": 1504 }, { "epoch": 94.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 64.61081018532731, "eval_classification_report": "{\"accusation\": {\"precision\": 0.8333333333333334, \"recall\": 0.8333333333333334, \"f1-score\": 0.8333333333333334, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6486486486486487, \"recall\": 0.7741935483870968, \"f1-score\": 0.7058823529411764, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7, \"recall\": 0.5675675675675675, \"f1-score\": 0.626865671641791, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5714285714285714, \"recall\": 0.4444444444444444, \"f1-score\": 0.5, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7643533643533644, \"recall\": 0.7684145839387775, \"f1-score\": 0.7602803717597397, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7116083590221521, \"recall\": 0.7068965517241379, \"f1-score\": 0.7027744598049765, \"support\": 116.0}}", "eval_f1_macro": 46.78648441598398, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.27744598049766, "eval_loss": 0.986122190952301, "eval_runtime": 2.1905, "eval_samples_per_second": 52.957, "step": 1504 }, { "epoch": 95.0, "learning_rate": 0.000175, "loss": 0.0018, "step": 1520 }, { "epoch": 95.0, "eval_accuracy": 69.82758620689656, "eval_average_metrics": 63.83607096081741, "eval_classification_report": "{\"accusation\": {\"precision\": 0.8333333333333334, \"recall\": 0.8333333333333334, \"f1-score\": 0.8333333333333334, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6486486486486487, \"recall\": 0.7741935483870968, \"f1-score\": 0.7058823529411764, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6774193548387096, \"recall\": 0.5675675675675675, \"f1-score\": 0.6176470588235294, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5714285714285714, \"recall\": 0.4444444444444444, \"f1-score\": 0.5, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7597946725970919, \"recall\": 0.7545256950498886, \"f1-score\": 0.7509581105169341, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7033283256517405, \"recall\": 0.6982758620689655, \"f1-score\": 0.6947630462843445, \"support\": 116.0}}", "eval_f1_macro": 46.21280680104209, "eval_f1_micro": 69.82758620689656, "eval_f1_weighted": 69.47630462843445, "eval_loss": 0.9718347787857056, "eval_runtime": 2.3653, "eval_samples_per_second": 49.042, "step": 1520 }, { "epoch": 96.0, "learning_rate": 0.0001733333333333333, "loss": 0.0007, "step": 1536 }, { "epoch": 96.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 67.76175141658041, "eval_classification_report": "{\"accusation\": {\"precision\": 0.8333333333333334, \"recall\": 0.8333333333333334, \"f1-score\": 0.8333333333333334, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7666666666666667, \"recall\": 0.7419354838709677, \"f1-score\": 0.7540983606557377, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7058823529411765, \"recall\": 0.6486486486486487, \"f1-score\": 0.676056338028169, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7860031512605042, \"recall\": 0.8022952387871742, \"f1-score\": 0.7894486090441719, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7488487636433885, \"recall\": 0.7413793103448276, \"f1-score\": 0.7418969073309942, \"support\": 116.0}}", "eval_f1_macro": 48.581452864256725, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.18969073309943, "eval_loss": 0.9660546779632568, "eval_runtime": 2.1618, "eval_samples_per_second": 53.659, "step": 1536 }, { "epoch": 97.0, "learning_rate": 0.00017166666666666665, "loss": 0.0005, "step": 1552 }, { "epoch": 97.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 66.37793808045248, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6666666666666666, \"recall\": 0.6486486486486487, \"f1-score\": 0.6575342465753425, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7854662698412698, \"recall\": 0.794230722658142, \"f1-score\": 0.7825183305399483, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7368466064586755, \"recall\": 0.7241379310344828, \"f1-score\": 0.7252919192783961, \"support\": 116.0}}", "eval_f1_macro": 48.15497418707374, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.5291919278396, "eval_loss": 0.9771972894668579, "eval_runtime": 2.1794, "eval_samples_per_second": 53.226, "step": 1552 }, { "epoch": 98.0, "learning_rate": 0.00016999999999999999, "loss": 0.0034, "step": 1568 }, { "epoch": 98.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 64.7984190383411, "eval_classification_report": "{\"accusation\": {\"precision\": 0.5789473684210527, \"recall\": 0.9166666666666666, \"f1-score\": 0.7096774193548387, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7241379310344828, \"recall\": 0.5675675675675675, \"f1-score\": 0.6363636363636365, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7658504060216855, \"recall\": 0.7945122541896734, \"f1-score\": 0.7665646406681366, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7297752879981898, \"recall\": 0.7068965517241379, \"f1-score\": 0.7064115715203608, \"support\": 116.0}}", "eval_f1_macro": 47.17320865650072, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.6411571520361, "eval_loss": 1.0846428871154785, "eval_runtime": 2.1972, "eval_samples_per_second": 52.795, "step": 1568 }, { "epoch": 99.0, "learning_rate": 0.00016833333333333332, "loss": 0.0011, "step": 1584 }, { "epoch": 99.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 67.07366957347286, "eval_classification_report": "{\"accusation\": {\"precision\": 0.625, \"recall\": 0.8333333333333334, \"f1-score\": 0.7142857142857143, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6756756756756757, \"f1-score\": 0.6944444444444444, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7878336940836941, \"recall\": 0.7976091010365204, \"f1-score\": 0.7850258781673589, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7475011817684232, \"recall\": 0.7327586206896551, \"f1-score\": 0.7343366934566137, \"support\": 116.0}}", "eval_f1_macro": 48.309284810299005, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.43366934566137, "eval_loss": 0.9998502135276794, "eval_runtime": 2.0893, "eval_samples_per_second": 55.521, "step": 1584 }, { "epoch": 100.0, "learning_rate": 0.00016666666666666666, "loss": 0.0058, "step": 1600 }, { "epoch": 100.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 67.84299596892572, "eval_classification_report": "{\"accusation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8333333333333334, \"f1-score\": 0.7407407407407408, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"evaluation\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7, \"recall\": 0.7567567567567568, \"f1-score\": 0.7272727272727273, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7142857142857143, \"recall\": 0.5555555555555556, \"f1-score\": 0.6250000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.8226190476190476, \"recall\": 0.7799664583938777, \"f1-score\": 0.7944732723097693, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7519704433497537, \"recall\": 0.7413793103448276, \"f1-score\": 0.7420545889536696, \"support\": 116.0}}", "eval_f1_macro": 48.89066291137042, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.20545889536697, "eval_loss": 0.9746664762496948, "eval_runtime": 2.2781, "eval_samples_per_second": 50.92, "step": 1600 }, { "epoch": 101.0, "learning_rate": 0.000165, "loss": 0.0019, "step": 1616 }, { "epoch": 101.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 66.95791619492573, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7857142857142857, \"recall\": 0.9166666666666666, \"f1-score\": 0.8461538461538461, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6470588235294118, \"recall\": 0.7096774193548387, \"f1-score\": 0.6769230769230768, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.7027027027027027, \"f1-score\": 0.7123287671232876, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5714285714285714, \"recall\": 0.4444444444444444, \"f1-score\": 0.5, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.8001779878618114, \"recall\": 0.7737697374794148, \"f1-score\": 0.7822729027832086, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7364437441643324, \"recall\": 0.7327586206896551, \"f1-score\": 0.7314006970126677, \"support\": 116.0}}", "eval_f1_macro": 48.13987094050513, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.14006970126677, "eval_loss": 1.0031696557998657, "eval_runtime": 2.174, "eval_samples_per_second": 53.358, "step": 1616 }, { "epoch": 102.0, "learning_rate": 0.0001633333333333333, "loss": 0.0003, "step": 1632 }, { "epoch": 102.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 64.6906117356637, "eval_classification_report": "{\"accusation\": {\"precision\": 0.625, \"recall\": 0.8333333333333334, \"f1-score\": 0.7142857142857143, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7916666666666666, \"recall\": 0.6129032258064516, \"f1-score\": 0.6909090909090909, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6666666666666666, \"recall\": 0.6486486486486487, \"f1-score\": 0.6575342465753425, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7468434343434344, \"recall\": 0.7861662065291097, \"f1-score\": 0.7613133536934906, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7153539707419017, \"recall\": 0.7068965517241379, \"f1-score\": 0.7053308406284317, \"support\": 116.0}}", "eval_f1_macro": 46.85005253498404, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.53308406284316, "eval_loss": 1.0247207880020142, "eval_runtime": 2.125, "eval_samples_per_second": 54.588, "step": 1632 }, { "epoch": 103.0, "learning_rate": 0.00016166666666666665, "loss": 0.0076, "step": 1648 }, { "epoch": 103.0, "eval_accuracy": 67.24137931034483, "eval_average_metrics": 61.543610356616945, "eval_classification_report": "{\"accusation\": {\"precision\": 0.5789473684210527, \"recall\": 0.9166666666666666, \"f1-score\": 0.7096774193548387, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6363636363636364, \"recall\": 0.6774193548387096, \"f1-score\": 0.65625, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.76, \"recall\": 0.5135135135135135, \"f1-score\": 0.6129032258064517, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5555555555555556, \"recall\": 0.5555555555555556, \"f1-score\": 0.5555555555555556, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.6724137931034483, \"macro avg\": {\"precision\": 0.7214063969656075, \"recall\": 0.7599777196551389, \"f1-score\": 0.7293985424692849, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.6914491610929905, \"recall\": 0.6724137931034483, \"f1-score\": 0.6680561865382214, \"support\": 116.0}}", "eval_f1_macro": 44.88606415195599, "eval_f1_micro": 67.24137931034483, "eval_f1_weighted": 66.80561865382214, "eval_loss": 1.1534621715545654, "eval_runtime": 2.1588, "eval_samples_per_second": 53.734, "step": 1648 }, { "epoch": 104.0, "learning_rate": 0.00015999999999999999, "loss": 0.001, "step": 1664 }, { "epoch": 104.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 64.83670185178806, "eval_classification_report": "{\"accusation\": {\"precision\": 0.625, \"recall\": 0.8333333333333334, \"f1-score\": 0.7142857142857143, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7916666666666666, \"recall\": 0.6129032258064516, \"f1-score\": 0.6909090909090909, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6571428571428571, \"recall\": 0.6216216216216216, \"f1-score\": 0.6388888888888888, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7589486901986902, \"recall\": 0.7966767170396203, \"f1-score\": 0.7700306637806638, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7196450819295647, \"recall\": 0.7068965517241379, \"f1-score\": 0.7058099467582226, \"support\": 116.0}}", "eval_f1_macro": 47.386502386502386, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.58099467582227, "eval_loss": 1.0140974521636963, "eval_runtime": 2.0908, "eval_samples_per_second": 55.482, "step": 1664 }, { "epoch": 105.0, "learning_rate": 0.00015833333333333332, "loss": 0.0019, "step": 1680 }, { "epoch": 105.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 65.47830212383965, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.6774193548387096, \"f1-score\": 0.7, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.696969696969697, \"recall\": 0.6216216216216216, \"f1-score\": 0.6571428571428571, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7583858311229001, \"recall\": 0.7908523442797636, \"f1-score\": 0.7697431245225362, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7207556736763039, \"recall\": 0.7155172413793104, \"f1-score\": 0.7144095255657122, \"support\": 116.0}}", "eval_f1_macro": 47.36880766292531, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.44095255657122, "eval_loss": 0.9925776124000549, "eval_runtime": 2.1371, "eval_samples_per_second": 54.278, "step": 1680 }, { "epoch": 106.0, "learning_rate": 0.00015666666666666666, "loss": 0.0006, "step": 1696 }, { "epoch": 106.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 66.27571069810807, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7142857142857143, \"recall\": 0.6451612903225806, \"f1-score\": 0.6779661016949152, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6756756756756757, \"f1-score\": 0.684931506849315, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7708246614496614, \"recall\": 0.7935768429720043, \"f1-score\": 0.7795533775386171, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7259982594034318, \"recall\": 0.7241379310344828, \"f1-score\": 0.7230274104469775, \"support\": 116.0}}", "eval_f1_macro": 47.97251554083797, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.30274104469775, "eval_loss": 0.9963064789772034, "eval_runtime": 2.1179, "eval_samples_per_second": 54.77, "step": 1696 }, { "epoch": 107.0, "learning_rate": 0.000155, "loss": 0.0024, "step": 1712 }, { "epoch": 107.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 64.7377253700138, "eval_classification_report": "{\"accusation\": {\"precision\": 0.6923076923076923, \"recall\": 0.75, \"f1-score\": 0.7199999999999999, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.631578947368421, \"recall\": 0.7741935483870968, \"f1-score\": 0.6956521739130435, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7333333333333333, \"recall\": 0.5945945945945946, \"f1-score\": 0.6567164179104478, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.625, \"recall\": 0.5555555555555556, \"f1-score\": 0.5882352941176471, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7779858299595142, \"recall\": 0.7613762956504893, \"f1-score\": 0.7628976145941829, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7199997673228162, \"recall\": 0.7068965517241379, \"f1-score\": 0.7062404562173948, \"support\": 116.0}}", "eval_f1_macro": 46.94754551348818, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.62404562173948, "eval_loss": 1.0944117307662964, "eval_runtime": 2.2195, "eval_samples_per_second": 52.264, "step": 1712 }, { "epoch": 108.0, "learning_rate": 0.0001533333333333333, "loss": 0.0002, "step": 1728 }, { "epoch": 108.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 66.35003338809494, "eval_classification_report": "{\"accusation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8333333333333334, \"f1-score\": 0.7407407407407408, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.7096774193548387, \"f1-score\": 0.7213114754098361, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.696969696969697, \"recall\": 0.6216216216216216, \"f1-score\": 0.6571428571428571, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.784375, \"recall\": 0.7948846023442797, \"f1-score\": 0.7827215130132199, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.733823145245559, \"recall\": 0.7241379310344828, \"f1-score\": 0.7240506962159275, \"support\": 116.0}}", "eval_f1_macro": 48.16747772389046, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.40506962159276, "eval_loss": 1.0319750308990479, "eval_runtime": 2.2528, "eval_samples_per_second": 51.491, "step": 1728 }, { "epoch": 109.0, "learning_rate": 0.00015166666666666665, "loss": 0.0013, "step": 1744 }, { "epoch": 109.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 65.66781775563415, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6875, \"recall\": 0.5945945945945946, \"f1-score\": 0.6376811594202898, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7836538461538461, \"recall\": 0.7915062239659014, \"f1-score\": 0.7806671944139975, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7250497347480106, \"recall\": 0.7155172413793104, \"f1-score\": 0.7152676462889004, \"support\": 116.0}}", "eval_f1_macro": 48.04105811778446, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.52676462889004, "eval_loss": 1.0217504501342773, "eval_runtime": 2.24, "eval_samples_per_second": 51.786, "step": 1744 }, { "epoch": 110.0, "learning_rate": 0.00015, "loss": 0.0011, "step": 1760 }, { "epoch": 110.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 67.81005930236041, "eval_classification_report": "{\"accusation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8333333333333334, \"f1-score\": 0.7407407407407408, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7857142857142857, \"recall\": 0.7096774193548387, \"f1-score\": 0.7457627118644068, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6756756756756757, \"f1-score\": 0.6944444444444444, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7940340909090908, \"recall\": 0.8016413591010365, \"f1-score\": 0.7910358540779776, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7539324899238693, \"recall\": 0.7413793103448276, \"f1-score\": 0.7428524565875445, \"support\": 116.0}}", "eval_f1_macro": 48.6791294817217, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.28524565875443, "eval_loss": 1.0107359886169434, "eval_runtime": 2.059, "eval_samples_per_second": 56.337, "step": 1760 }, { "epoch": 111.0, "learning_rate": 0.00014833333333333332, "loss": 0.0006, "step": 1776 }, { "epoch": 111.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 64.72662641030664, "eval_classification_report": "{\"accusation\": {\"precision\": 0.8181818181818182, \"recall\": 0.75, \"f1-score\": 0.7826086956521738, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7096774193548387, \"recall\": 0.5945945945945946, \"f1-score\": 0.6470588235294118, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7548846482818258, \"recall\": 0.7810895572992347, \"f1-score\": 0.761943627912298, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7159673092480378, \"recall\": 0.7068965517241379, \"f1-score\": 0.7063835665564212, \"support\": 116.0}}", "eval_f1_macro": 46.888838640756795, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.63835665564213, "eval_loss": 1.0691028833389282, "eval_runtime": 2.2902, "eval_samples_per_second": 50.652, "step": 1776 }, { "epoch": 112.0, "learning_rate": 0.00014666666666666664, "loss": 0.0003, "step": 1792 }, { "epoch": 112.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 66.27387783514109, "eval_classification_report": "{\"accusation\": {\"precision\": 0.8181818181818182, \"recall\": 0.75, \"f1-score\": 0.7826086956521738, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.696969696969697, \"recall\": 0.7419354838709677, \"f1-score\": 0.71875, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.71875, \"recall\": 0.6216216216216216, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7770055465367965, \"recall\": 0.7885001937421292, \"f1-score\": 0.7775157753318719, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7327597869084939, \"recall\": 0.7241379310344828, \"f1-score\": 0.7242080049786032, \"support\": 116.0}}", "eval_f1_macro": 47.8471246358075, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.42080049786031, "eval_loss": 1.070449709892273, "eval_runtime": 2.1308, "eval_samples_per_second": 54.439, "step": 1792 }, { "epoch": 113.0, "learning_rate": 0.000145, "loss": 0.0013, "step": 1808 }, { "epoch": 113.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 66.94971808239175, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.6451612903225806, \"f1-score\": 0.689655172413793, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.7027027027027027, \"f1-score\": 0.7123287671232876, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7709749740999741, \"recall\": 0.7969552213503825, \"f1-score\": 0.7802725022460566, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7378138703569739, \"recall\": 0.7327586206896551, \"f1-score\": 0.7323037882264791, \"support\": 116.0}}", "eval_f1_macro": 48.0167693689881, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.2303788226479, "eval_loss": 1.0506659746170044, "eval_runtime": 2.1797, "eval_samples_per_second": 53.219, "step": 1808 }, { "epoch": 114.0, "learning_rate": 0.00014333333333333334, "loss": 0.0014, "step": 1824 }, { "epoch": 114.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 64.76925999714823, "eval_classification_report": "{\"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.7096774193548387, \"f1-score\": 0.7213114754098361, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.5675675675675675, \"f1-score\": 0.6461538461538462, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5384615384615384, \"recall\": 0.7777777777777778, \"f1-score\": 0.6363636363636364, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7625518476621418, \"recall\": 0.788127845587523, \"f1-score\": 0.7626786408151753, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7310497607531076, \"recall\": 0.7068965517241379, \"f1-score\": 0.7076365943975451, \"support\": 116.0}}", "eval_f1_macro": 46.934070204010794, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.7636594397545, "eval_loss": 1.1001496315002441, "eval_runtime": 2.131, "eval_samples_per_second": 54.435, "step": 1824 }, { "epoch": 115.0, "learning_rate": 0.00014166666666666665, "loss": 0.0007, "step": 1840 }, { "epoch": 115.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 64.70115661864702, "eval_classification_report": "{\"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7333333333333333, \"recall\": 0.5945945945945946, \"f1-score\": 0.6567164179104478, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7503036791639733, \"recall\": 0.7874739659013852, \"f1-score\": 0.7594631003510002, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7238169073995647, \"recall\": 0.7068965517241379, \"f1-score\": 0.7068912533892974, \"support\": 116.0}}", "eval_f1_macro": 46.736190790830776, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.68912533892974, "eval_loss": 1.0808982849121094, "eval_runtime": 2.193, "eval_samples_per_second": 52.895, "step": 1840 }, { "epoch": 116.0, "learning_rate": 0.00014, "loss": 0.0002, "step": 1856 }, { "epoch": 116.0, "eval_accuracy": 69.82758620689656, "eval_average_metrics": 63.88409380563921, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7307692307692307, \"recall\": 0.6129032258064516, \"f1-score\": 0.6666666666666667, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7333333333333333, \"recall\": 0.5945945945945946, \"f1-score\": 0.6567164179104478, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.8888888888888888, \"f1-score\": 0.6956521739130435, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.727754884004884, \"recall\": 0.7932983386612419, \"f1-score\": 0.7513110312428937, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7116505620815966, \"recall\": 0.6982758620689655, \"f1-score\": 0.6964667780920103, \"support\": 116.0}}", "eval_f1_macro": 46.23452499956269, "eval_f1_micro": 69.82758620689656, "eval_f1_weighted": 69.64667780920102, "eval_loss": 1.1061365604400635, "eval_runtime": 2.1218, "eval_samples_per_second": 54.67, "step": 1856 }, { "epoch": 117.0, "learning_rate": 0.00013833333333333333, "loss": 0.0058, "step": 1872 }, { "epoch": 117.0, "eval_accuracy": 68.96551724137932, "eval_average_metrics": 63.282710359282376, "eval_classification_report": "{\"accusation\": {\"precision\": 0.625, \"recall\": 0.8333333333333334, \"f1-score\": 0.7142857142857143, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7307692307692307, \"recall\": 0.6129032258064516, \"f1-score\": 0.6666666666666667, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6666666666666666, \"recall\": 0.5945945945945946, \"f1-score\": 0.6285714285714286, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.6896551724137931, \"macro avg\": {\"precision\": 0.7441620879120879, \"recall\": 0.7794094497723529, \"f1-score\": 0.7533740769034887, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7012165277251484, \"recall\": 0.6896551724137931, \"f1-score\": 0.6883832529877154, \"support\": 116.0}}", "eval_f1_macro": 46.3614816555993, "eval_f1_micro": 68.96551724137932, "eval_f1_weighted": 68.83832529877154, "eval_loss": 1.0625555515289307, "eval_runtime": 2.1789, "eval_samples_per_second": 53.238, "step": 1872 }, { "epoch": 118.0, "learning_rate": 0.00013666666666666666, "loss": 0.0003, "step": 1888 }, { "epoch": 118.0, "eval_accuracy": 68.96551724137932, "eval_average_metrics": 63.282710359282376, "eval_classification_report": "{\"accusation\": {\"precision\": 0.625, \"recall\": 0.8333333333333334, \"f1-score\": 0.7142857142857143, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7307692307692307, \"recall\": 0.6129032258064516, \"f1-score\": 0.6666666666666667, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6666666666666666, \"recall\": 0.5945945945945946, \"f1-score\": 0.6285714285714286, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.6896551724137931, \"macro avg\": {\"precision\": 0.7441620879120879, \"recall\": 0.7794094497723529, \"f1-score\": 0.7533740769034887, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7012165277251484, \"recall\": 0.6896551724137931, \"f1-score\": 0.6883832529877154, \"support\": 116.0}}", "eval_f1_macro": 46.3614816555993, "eval_f1_micro": 68.96551724137932, "eval_f1_weighted": 68.83832529877154, "eval_loss": 1.0653810501098633, "eval_runtime": 2.1294, "eval_samples_per_second": 54.477, "step": 1888 }, { "epoch": 119.0, "learning_rate": 0.000135, "loss": 0.0026, "step": 1904 }, { "epoch": 119.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 65.66956317091271, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.7096774193548387, \"f1-score\": 0.7096774193548389, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6666666666666666, \"recall\": 0.5945945945945946, \"f1-score\": 0.6285714285714286, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7840403883750657, \"recall\": 0.7915062239659014, \"f1-score\": 0.7807742305133197, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.72446698416526, \"recall\": 0.7155172413793104, \"f1-score\": 0.7152715945312295, \"support\": 116.0}}", "eval_f1_macro": 48.04764495466583, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.52715945312295, "eval_loss": 1.0476528406143188, "eval_runtime": 2.0205, "eval_samples_per_second": 57.412, "step": 1904 }, { "epoch": 120.0, "learning_rate": 0.0001333333333333333, "loss": 0.003, "step": 1920 }, { "epoch": 120.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 64.70006170794646, "eval_classification_report": "{\"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.6774193548387096, \"f1-score\": 0.7, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.5675675675675675, \"f1-score\": 0.6461538461538462, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7461895102868734, \"recall\": 0.7979844764118957, \"f1-score\": 0.7633107137871202, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7203813986950309, \"recall\": 0.7068965517241379, \"f1-score\": 0.7044796948467391, \"support\": 116.0}}", "eval_f1_macro": 46.97296700228433, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.4479694846739, "eval_loss": 1.0797429084777832, "eval_runtime": 2.0702, "eval_samples_per_second": 56.032, "step": 1920 }, { "epoch": 121.0, "learning_rate": 0.00013166666666666665, "loss": 0.0049, "step": 1936 }, { "epoch": 121.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 64.70006170794646, "eval_classification_report": "{\"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.6774193548387096, \"f1-score\": 0.7, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.5675675675675675, \"f1-score\": 0.6461538461538462, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7461895102868734, \"recall\": 0.7979844764118957, \"f1-score\": 0.7633107137871202, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7203813986950309, \"recall\": 0.7068965517241379, \"f1-score\": 0.7044796948467391, \"support\": 116.0}}", "eval_f1_macro": 46.97296700228433, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.4479694846739, "eval_loss": 1.094364881515503, "eval_runtime": 2.0626, "eval_samples_per_second": 56.239, "step": 1936 }, { "epoch": 122.0, "learning_rate": 0.00013, "loss": 0.0013, "step": 1952 }, { "epoch": 122.0, "eval_accuracy": 69.82758620689656, "eval_average_metrics": 63.88829920079921, "eval_classification_report": "{\"accusation\": {\"precision\": 0.6111111111111112, \"recall\": 0.9166666666666666, \"f1-score\": 0.7333333333333334, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.76, \"recall\": 0.6129032258064516, \"f1-score\": 0.6785714285714285, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7241379310344828, \"recall\": 0.5675675675675675, \"f1-score\": 0.6363636363636365, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.734797821909891, \"recall\": 0.8003366269495301, \"f1-score\": 0.755275974025974, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7144611700230011, \"recall\": 0.6982758620689655, \"f1-score\": 0.6941950291088222, \"support\": 116.0}}", "eval_f1_macro": 46.47852147852148, "eval_f1_micro": 69.82758620689656, "eval_f1_weighted": 69.41950291088223, "eval_loss": 1.1192301511764526, "eval_runtime": 2.2276, "eval_samples_per_second": 52.075, "step": 1952 }, { "epoch": 123.0, "learning_rate": 0.00012833333333333333, "loss": 0.0002, "step": 1968 }, { "epoch": 123.0, "eval_accuracy": 69.82758620689656, "eval_average_metrics": 64.06779904745557, "eval_classification_report": "{\"accusation\": {\"precision\": 0.5, \"recall\": 0.9166666666666666, \"f1-score\": 0.6470588235294118, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.8571428571428571, \"recall\": 0.5806451612903226, \"f1-score\": 0.6923076923076923, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7333333333333333, \"recall\": 0.5945945945945946, \"f1-score\": 0.6567164179104478, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.8888888888888888, \"f1-score\": 0.6956521739130435, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7556547619047619, \"recall\": 0.7996827472633924, \"f1-score\": 0.7572062368285017, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7439449917898194, \"recall\": 0.6982758620689655, \"f1-score\": 0.7001871689427523, \"support\": 116.0}}", "eval_f1_macro": 46.59730688175395, "eval_f1_micro": 69.82758620689656, "eval_f1_weighted": 70.01871689427523, "eval_loss": 1.1883636713027954, "eval_runtime": 2.2785, "eval_samples_per_second": 50.911, "step": 1968 }, { "epoch": 124.0, "learning_rate": 0.00012666666666666666, "loss": 0.001, "step": 1984 }, { "epoch": 124.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 67.65132265328175, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.8636363636363636, \"recall\": 0.6129032258064516, \"f1-score\": 0.7169811320754716, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6829268292682927, \"recall\": 0.7567567567567568, \"f1-score\": 0.7179487179487181, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7806580614507443, \"recall\": 0.7996797200426232, \"f1-score\": 0.7847306037226593, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7553021200834489, \"recall\": 0.7413793103448276, \"f1-score\": 0.7403831446892091, \"support\": 116.0}}", "eval_f1_macro": 48.29111407524058, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.0383144689209, "eval_loss": 1.0762184858322144, "eval_runtime": 2.2546, "eval_samples_per_second": 51.451, "step": 1984 }, { "epoch": 125.0, "learning_rate": 0.000125, "loss": 0.0002, "step": 2000 }, { "epoch": 125.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 67.0707623347613, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.7096774193548387, \"f1-score\": 0.7213114754098361, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6857142857142857, \"recall\": 0.6486486486486487, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.783531746031746, \"recall\": 0.7982629807226582, \"f1-score\": 0.7892417571882627, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7330459770114942, \"recall\": 0.7327586206896551, \"f1-score\": 0.7316260168183646, \"support\": 116.0}}", "eval_f1_macro": 48.56872351927771, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.16260168183646, "eval_loss": 1.0365558862686157, "eval_runtime": 2.1837, "eval_samples_per_second": 53.12, "step": 2000 }, { "epoch": 126.0, "learning_rate": 0.0001233333333333333, "loss": 0.0008, "step": 2016 }, { "epoch": 126.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 66.18828070182492, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.6451612903225806, \"f1-score\": 0.689655172413793, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7272727272727273, \"recall\": 0.6486486486486487, \"f1-score\": 0.6857142857142857, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7553631553631555, \"recall\": 0.8040873534825148, \"f1-score\": 0.7754215196479262, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7279218705942844, \"recall\": 0.7241379310344828, \"f1-score\": 0.7220728923745383, \"support\": 116.0}}", "eval_f1_macro": 47.7182473629493, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.20728923745384, "eval_loss": 1.0847417116165161, "eval_runtime": 2.1565, "eval_samples_per_second": 53.791, "step": 2016 }, { "epoch": 127.0, "learning_rate": 0.00012166666666666665, "loss": 0.0015, "step": 2032 }, { "epoch": 127.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 67.8450022809264, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.696969696969697, \"recall\": 0.7419354838709677, \"f1-score\": 0.71875, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7272727272727273, \"recall\": 0.6486486486486487, \"f1-score\": 0.6857142857142857, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7963338744588745, \"recall\": 0.8022952387871742, \"f1-score\": 0.7970214793913517, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.744126921928646, \"recall\": 0.7413793103448276, \"f1-score\": 0.7405667139988761, \"support\": 116.0}}", "eval_f1_macro": 49.04747565485241, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.0566713998876, "eval_loss": 1.0856082439422607, "eval_runtime": 2.2011, "eval_samples_per_second": 52.701, "step": 2032 }, { "epoch": 128.0, "learning_rate": 0.00011999999999999999, "loss": 0.0002, "step": 2048 }, { "epoch": 128.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 67.14339660631728, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7058823529411765, \"recall\": 0.6486486486486487, \"f1-score\": 0.676056338028169, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7924763655462185, \"recall\": 0.7982629807226582, \"f1-score\": 0.7932720732321743, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7347734352361635, \"recall\": 0.7327586206896551, \"f1-score\": 0.7320511931920426, \"support\": 116.0}}", "eval_f1_macro": 48.81674296813381, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.20511931920429, "eval_loss": 1.0866280794143677, "eval_runtime": 2.1867, "eval_samples_per_second": 53.047, "step": 2048 }, { "epoch": 129.0, "learning_rate": 0.00011833333333333331, "loss": 0.0002, "step": 2064 }, { "epoch": 129.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 67.85334504297987, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.7096774193548387, \"f1-score\": 0.7096774193548389, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6756756756756757, \"f1-score\": 0.6944444444444444, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7962989631336406, \"recall\": 0.8016413591010365, \"f1-score\": 0.7969786766519763, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7433805418719213, \"recall\": 0.7413793103448276, \"f1-score\": 0.7409267646283232, \"support\": 116.0}}", "eval_f1_macro": 49.044841640121625, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.09267646283233, "eval_loss": 1.0848969221115112, "eval_runtime": 2.0553, "eval_samples_per_second": 56.44, "step": 2064 }, { "epoch": 130.0, "learning_rate": 0.00011666666666666665, "loss": 0.0002, "step": 2080 }, { "epoch": 130.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 67.18824972788626, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.7096774193548387, \"f1-score\": 0.6875, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7058823529411765, \"recall\": 0.6486486486486487, \"f1-score\": 0.676056338028169, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7967403307476837, \"recall\": 0.7982629807226582, \"f1-score\": 0.7957541397767409, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7348898777017439, \"recall\": 0.7327586206896551, \"f1-score\": 0.7323178924889144, \"support\": 116.0}}", "eval_f1_macro": 48.969485524722515, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.23178924889146, "eval_loss": 1.095829963684082, "eval_runtime": 2.0853, "eval_samples_per_second": 55.627, "step": 2080 }, { "epoch": 131.0, "learning_rate": 0.000115, "loss": 0.0005, "step": 2096 }, { "epoch": 131.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 66.29230917287957, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.6774193548387096, \"f1-score\": 0.7, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7058823529411765, \"recall\": 0.6486486486486487, \"f1-score\": 0.676056338028169, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7733818645012864, \"recall\": 0.794230722658142, \"f1-score\": 0.7794353982112888, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7298702490941247, \"recall\": 0.7241379310344828, \"f1-score\": 0.7237639521008089, \"support\": 116.0}}", "eval_f1_macro": 47.96525527454085, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.37639521008089, "eval_loss": 1.105514407157898, "eval_runtime": 2.1232, "eval_samples_per_second": 54.634, "step": 2096 }, { "epoch": 132.0, "learning_rate": 0.00011333333333333331, "loss": 0.0002, "step": 2112 }, { "epoch": 132.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 67.80477325033718, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.6774193548387096, \"f1-score\": 0.7, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.7027027027027027, \"f1-score\": 0.7123287671232876, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7911440451311141, \"recall\": 0.8009874794148988, \"f1-score\": 0.7933992764095822, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.744839157242546, \"recall\": 0.7413793103448276, \"f1-score\": 0.7411866007640886, \"support\": 116.0}}", "eval_f1_macro": 48.82457085597429, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.11866007640886, "eval_loss": 1.083550214767456, "eval_runtime": 2.2453, "eval_samples_per_second": 51.664, "step": 2112 }, { "epoch": 133.0, "learning_rate": 0.00011166666666666667, "loss": 0.0007, "step": 2128 }, { "epoch": 133.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 66.12366410068819, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.75, \"f1-score\": 0.7058823529411765, \"support\": 8.0}, \"challenge\": {\"precision\": 0.76, \"recall\": 0.6129032258064516, \"f1-score\": 0.6785714285714285, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.7027027027027027, \"f1-score\": 0.7123287671232876, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.756097582972583, \"recall\": 0.7929229632858665, \"f1-score\": 0.7695261745372544, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7315296437279195, \"recall\": 0.7241379310344828, \"f1-score\": 0.7231161330125597, \"support\": 116.0}}", "eval_f1_macro": 47.355456894600266, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.31161330125597, "eval_loss": 1.164588451385498, "eval_runtime": 2.2443, "eval_samples_per_second": 51.686, "step": 2128 }, { "epoch": 134.0, "learning_rate": 0.00010999999999999998, "loss": 0.0012, "step": 2144 }, { "epoch": 134.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 65.51241324001258, "eval_classification_report": "{\"accusation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6470588235294118, \"recall\": 0.7096774193548387, \"f1-score\": 0.6769230769230768, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7419354838709677, \"recall\": 0.6216216216216216, \"f1-score\": 0.676470588235294, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7628919651927243, \"recall\": 0.7844679356776131, \"f1-score\": 0.7716595882617554, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7192905962061232, \"recall\": 0.7155172413793104, \"f1-score\": 0.7145946079115717, \"support\": 116.0}}", "eval_f1_macro": 47.486743893031104, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.45946079115717, "eval_loss": 1.159348487854004, "eval_runtime": 2.0817, "eval_samples_per_second": 55.725, "step": 2144 }, { "epoch": 135.0, "learning_rate": 0.00010833333333333333, "loss": 0.0001, "step": 2160 }, { "epoch": 135.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 67.09424591991048, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6571428571428571, \"recall\": 0.7419354838709677, \"f1-score\": 0.6969696969696969, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7272727272727273, \"recall\": 0.6486486486486487, \"f1-score\": 0.6857142857142857, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7940569846819847, \"recall\": 0.7884063498982854, \"f1-score\": 0.7893731655341563, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.73658130662441, \"recall\": 0.7327586206896551, \"f1-score\": 0.732484493549936, \"support\": 116.0}}", "eval_f1_macro": 48.5768101867173, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.24844935499361, "eval_loss": 1.1159605979919434, "eval_runtime": 2.0586, "eval_samples_per_second": 56.35, "step": 2160 }, { "epoch": 136.0, "learning_rate": 0.00010666666666666667, "loss": 0.0008, "step": 2176 }, { "epoch": 136.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 67.66205711093811, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.75, \"f1-score\": 0.7058823529411765, \"support\": 8.0}, \"challenge\": {\"precision\": 0.76, \"recall\": 0.6129032258064516, \"f1-score\": 0.6785714285714285, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.717948717948718, \"recall\": 0.7567567567567568, \"f1-score\": 0.736842105263158, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7792376373626374, \"recall\": 0.7996797200426232, \"f1-score\": 0.7866254295240364, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7448609005936592, \"recall\": 0.7413793103448276, \"f1-score\": 0.739646476348462, \"support\": 116.0}}", "eval_f1_macro": 48.4077187399407, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 73.9646476348462, "eval_loss": 1.1159119606018066, "eval_runtime": 2.2156, "eval_samples_per_second": 52.355, "step": 2176 }, { "epoch": 137.0, "learning_rate": 0.00010499999999999999, "loss": 0.0002, "step": 2192 }, { "epoch": 137.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 64.57872110723775, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.75, \"f1-score\": 0.7058823529411765, \"support\": 8.0}, \"challenge\": {\"precision\": 0.76, \"recall\": 0.6129032258064516, \"f1-score\": 0.6785714285714285, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7058823529411765, \"recall\": 0.6486486486486487, \"f1-score\": 0.676056338028169, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7343265639589169, \"recall\": 0.7861662065291097, \"f1-score\": 0.7545112618900938, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7140724910653917, \"recall\": 0.7068965517241379, \"f1-score\": 0.7050411181396382, \"support\": 116.0}}", "eval_f1_macro": 46.43146227015962, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.50411181396383, "eval_loss": 1.1595603227615356, "eval_runtime": 2.1661, "eval_samples_per_second": 53.552, "step": 2192 }, { "epoch": 138.0, "learning_rate": 0.00010333333333333333, "loss": 0.0016, "step": 2208 }, { "epoch": 138.0, "eval_accuracy": 68.96551724137932, "eval_average_metrics": 63.281002597401425, "eval_classification_report": "{\"accusation\": {\"precision\": 0.8181818181818182, \"recall\": 0.75, \"f1-score\": 0.7826086956521738, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.7096774193548387, \"f1-score\": 0.6875, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6666666666666666, \"recall\": 0.5405405405405406, \"f1-score\": 0.5970149253731343, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.6896551724137931, \"macro avg\": {\"precision\": 0.748144216894217, \"recall\": 0.774332800542478, \"f1-score\": 0.75485509909281, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.6965268926475822, \"recall\": 0.6896551724137931, \"f1-score\": 0.6874035442421258, \"support\": 116.0}}", "eval_f1_macro": 46.45262148263446, "eval_f1_micro": 68.96551724137932, "eval_f1_weighted": 68.74035442421258, "eval_loss": 1.2053927183151245, "eval_runtime": 2.062, "eval_samples_per_second": 56.256, "step": 2208 }, { "epoch": 139.0, "learning_rate": 0.00010166666666666667, "loss": 0.0001, "step": 2224 }, { "epoch": 139.0, "eval_accuracy": 69.82758620689656, "eval_average_metrics": 64.14137179915826, "eval_classification_report": "{\"accusation\": {\"precision\": 0.8181818181818182, \"recall\": 0.75, \"f1-score\": 0.7826086956521738, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.7096774193548387, \"f1-score\": 0.6875, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6774193548387096, \"recall\": 0.5675675675675675, \"f1-score\": 0.6176470588235294, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5384615384615384, \"recall\": 0.7777777777777778, \"f1-score\": 0.6363636363636364, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7681467278241472, \"recall\": 0.7777111789208564, \"f1-score\": 0.7651180984580921, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7103061936209879, \"recall\": 0.6982758620689655, \"f1-score\": 0.6982612410849579, \"support\": 116.0}}", "eval_f1_macro": 47.08419067434413, "eval_f1_micro": 69.82758620689656, "eval_f1_weighted": 69.82612410849579, "eval_loss": 1.1781898736953735, "eval_runtime": 2.1597, "eval_samples_per_second": 53.712, "step": 2224 }, { "epoch": 140.0, "learning_rate": 9.999999999999999e-05, "loss": 0.0011, "step": 2240 }, { "epoch": 140.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 65.48352678423673, "eval_classification_report": "{\"accusation\": {\"precision\": 0.6875, \"recall\": 0.9166666666666666, \"f1-score\": 0.7857142857142857, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7916666666666666, \"recall\": 0.6129032258064516, \"f1-score\": 0.6909090909090909, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6857142857142857, \"recall\": 0.6486486486486487, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7619677197802198, \"recall\": 0.7965828731957764, \"f1-score\": 0.7700948561242679, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7300318933939624, \"recall\": 0.7155172413793104, \"f1-score\": 0.714402061765145, \"support\": 116.0}}", "eval_f1_macro": 47.39045268457034, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.4402061765145, "eval_loss": 1.1871286630630493, "eval_runtime": 2.3012, "eval_samples_per_second": 50.409, "step": 2240 }, { "epoch": 141.0, "learning_rate": 9.833333333333333e-05, "loss": 0.0003, "step": 2256 }, { "epoch": 141.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 65.6120488899911, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6764705882352942, \"recall\": 0.6216216216216216, \"f1-score\": 0.6478873239436619, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7768092814048697, \"recall\": 0.7908523442797636, \"f1-score\": 0.7754492703501312, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7285715678844482, \"recall\": 0.7155172413793104, \"f1-score\": 0.7162479218563275, \"support\": 116.0}}", "eval_f1_macro": 47.71995509846961, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.62479218563274, "eval_loss": 1.1495442390441895, "eval_runtime": 2.1903, "eval_samples_per_second": 52.961, "step": 2256 }, { "epoch": 142.0, "learning_rate": 9.666666666666667e-05, "loss": 0.0009, "step": 2272 }, { "epoch": 142.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 65.6120488899911, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6764705882352942, \"recall\": 0.6216216216216216, \"f1-score\": 0.6478873239436619, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7768092814048697, \"recall\": 0.7908523442797636, \"f1-score\": 0.7754492703501312, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7285715678844482, \"recall\": 0.7155172413793104, \"f1-score\": 0.7162479218563275, \"support\": 116.0}}", "eval_f1_macro": 47.71995509846961, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.62479218563274, "eval_loss": 1.1490837335586548, "eval_runtime": 2.1815, "eval_samples_per_second": 53.174, "step": 2272 }, { "epoch": 143.0, "learning_rate": 9.499999999999999e-05, "loss": 0.0001, "step": 2288 }, { "epoch": 143.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 66.33454933252901, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6857142857142857, \"recall\": 0.6486486486486487, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7814369658119659, \"recall\": 0.794230722658142, \"f1-score\": 0.7793308787223594, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7389433602795671, \"recall\": 0.7241379310344828, \"f1-score\": 0.7255178781722812, \"support\": 116.0}}", "eval_f1_macro": 47.95882330599135, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.55178781722812, "eval_loss": 1.153124451637268, "eval_runtime": 2.1286, "eval_samples_per_second": 54.495, "step": 2288 }, { "epoch": 144.0, "learning_rate": 9.333333333333333e-05, "loss": 0.004, "step": 2304 }, { "epoch": 144.0, "eval_accuracy": 69.82758620689656, "eval_average_metrics": 64.13153922628965, "eval_classification_report": "{\"accusation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8333333333333334, \"f1-score\": 0.7407407407407408, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.6774193548387096, \"f1-score\": 0.7, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.65625, \"recall\": 0.5675675675675675, \"f1-score\": 0.6086956521739131, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7650965683023873, \"recall\": 0.7840955875230069, \"f1-score\": 0.7655060069702013, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7102843741425043, \"recall\": 0.6982758620689655, \"f1-score\": 0.6976292252396847, \"support\": 116.0}}", "eval_f1_macro": 47.108061967397, "eval_f1_micro": 69.82758620689656, "eval_f1_weighted": 69.76292252396847, "eval_loss": 1.1598789691925049, "eval_runtime": 2.2008, "eval_samples_per_second": 52.709, "step": 2304 }, { "epoch": 145.0, "learning_rate": 9.166666666666667e-05, "loss": 0.0009, "step": 2320 }, { "epoch": 145.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 65.55450050368793, "eval_classification_report": "{\"accusation\": {\"precision\": 0.6470588235294118, \"recall\": 0.9166666666666666, \"f1-score\": 0.7586206896551724, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.8, \"recall\": 0.6451612903225806, \"f1-score\": 0.7142857142857142, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.696969696969697, \"recall\": 0.6216216216216216, \"f1-score\": 0.6571428571428571, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.8888888888888888, \"f1-score\": 0.6956521739130435, \"support\": 9.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7717238031576267, \"recall\": 0.7972367528819141, \"f1-score\": 0.771630046321377, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7381072570093868, \"recall\": 0.7155172413793104, \"f1-score\": 0.7162962781142028, \"support\": 116.0}}", "eval_f1_macro": 47.48492592746936, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.62962781142028, "eval_loss": 1.2059998512268066, "eval_runtime": 2.077, "eval_samples_per_second": 55.849, "step": 2320 }, { "epoch": 146.0, "learning_rate": 8.999999999999999e-05, "loss": 0.0008, "step": 2336 }, { "epoch": 146.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 64.68990280670708, "eval_classification_report": "{\"accusation\": {\"precision\": 0.6875, \"recall\": 0.9166666666666666, \"f1-score\": 0.7857142857142857, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7916666666666666, \"recall\": 0.6129032258064516, \"f1-score\": 0.6909090909090909, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6764705882352942, \"recall\": 0.6216216216216216, \"f1-score\": 0.6478873239436619, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7486594798175681, \"recall\": 0.793204494817398, \"f1-score\": 0.7620284840355263, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7195403691980772, \"recall\": 0.7068965517241379, \"f1-score\": 0.7048624032596834, \"support\": 116.0}}", "eval_f1_macro": 46.894060556032386, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.48624032596834, "eval_loss": 1.19475519657135, "eval_runtime": 2.2669, "eval_samples_per_second": 51.171, "step": 2336 }, { "epoch": 147.0, "learning_rate": 8.833333333333333e-05, "loss": 0.0002, "step": 2352 }, { "epoch": 147.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 64.70775012641597, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.7096774193548387, \"f1-score\": 0.7096774193548389, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7, \"recall\": 0.5675675675675675, \"f1-score\": 0.626865671641791, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7515683465280241, \"recall\": 0.788127845587523, \"f1-score\": 0.7635325328392253, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7135894278135657, \"recall\": 0.7068965517241379, \"f1-score\": 0.7046507275534553, \"support\": 116.0}}", "eval_f1_macro": 46.98661740549079, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.46507275534553, "eval_loss": 1.1726635694503784, "eval_runtime": 2.2233, "eval_samples_per_second": 52.176, "step": 2352 }, { "epoch": 148.0, "learning_rate": 8.666666666666665e-05, "loss": 0.0001, "step": 2368 }, { "epoch": 148.0, "eval_accuracy": 69.82758620689656, "eval_average_metrics": 64.00347575303101, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6774193548387096, \"recall\": 0.5675675675675675, \"f1-score\": 0.6176470588235294, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7475360884635078, \"recall\": 0.7840955875230069, \"f1-score\": 0.7597361025880794, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7038007737534989, \"recall\": 0.6982758620689655, \"f1-score\": 0.6960573966983371, \"support\": 116.0}}", "eval_f1_macro": 46.7529909284972, "eval_f1_micro": 69.82758620689656, "eval_f1_weighted": 69.60573966983371, "eval_loss": 1.1674237251281738, "eval_runtime": 2.108, "eval_samples_per_second": 55.028, "step": 2368 }, { "epoch": 149.0, "learning_rate": 8.499999999999999e-05, "loss": 0.0013, "step": 2384 }, { "epoch": 149.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 65.57406977941876, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6764705882352942, \"recall\": 0.6216216216216216, \"f1-score\": 0.6478873239436619, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7715246354952238, \"recall\": 0.7908523442797636, \"f1-score\": 0.7735764699479191, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.726712435637385, \"recall\": 0.7155172413793104, \"f1-score\": 0.7158812499886412, \"support\": 116.0}}", "eval_f1_macro": 47.60470584294886, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.58812499886412, "eval_loss": 1.1712332963943481, "eval_runtime": 2.1727, "eval_samples_per_second": 53.391, "step": 2384 }, { "epoch": 150.0, "learning_rate": 8.333333333333333e-05, "loss": 0.0006, "step": 2400 }, { "epoch": 150.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 67.15508351872047, "eval_classification_report": "{\"accusation\": {\"precision\": 0.8181818181818182, \"recall\": 0.75, \"f1-score\": 0.7826086956521738, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.7741935483870968, \"f1-score\": 0.7164179104477612, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.71875, \"recall\": 0.6216216216216216, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.8022174873737374, \"recall\": 0.7925324518066453, \"f1-score\": 0.7928398963556414, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7414299242424242, \"recall\": 0.7327586206896551, \"f1-score\": 0.7327846246891143, \"support\": 116.0}}", "eval_f1_macro": 48.79014746803947, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.27846246891143, "eval_loss": 1.21891450881958, "eval_runtime": 2.3304, "eval_samples_per_second": 49.776, "step": 2400 }, { "epoch": 151.0, "learning_rate": 8.166666666666665e-05, "loss": 0.0059, "step": 2416 }, { "epoch": 151.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 67.85326013055257, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.7027027027027027, \"f1-score\": 0.7027027027027027, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7978805728805729, \"recall\": 0.8009874794148988, \"f1-score\": 0.794782459278445, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7494031830238728, \"recall\": 0.7413793103448276, \"f1-score\": 0.7422748865149431, \"support\": 116.0}}", "eval_f1_macro": 48.90968980175047, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.22748865149431, "eval_loss": 1.139691948890686, "eval_runtime": 2.1983, "eval_samples_per_second": 52.768, "step": 2416 }, { "epoch": 152.0, "learning_rate": 7.999999999999999e-05, "loss": 0.0004, "step": 2432 }, { "epoch": 152.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 67.85326013055257, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.7027027027027027, \"f1-score\": 0.7027027027027027, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7978805728805729, \"recall\": 0.8009874794148988, \"f1-score\": 0.794782459278445, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7494031830238728, \"recall\": 0.7413793103448276, \"f1-score\": 0.7422748865149431, \"support\": 116.0}}", "eval_f1_macro": 48.90968980175047, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.22748865149431, "eval_loss": 1.1354390382766724, "eval_runtime": 2.0409, "eval_samples_per_second": 56.838, "step": 2432 }, { "epoch": 153.0, "learning_rate": 7.833333333333333e-05, "loss": 0.0003, "step": 2448 }, { "epoch": 153.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 67.10759310976837, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6756756756756757, \"f1-score\": 0.684931506849315, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7899801587301587, \"recall\": 0.7976091010365204, \"f1-score\": 0.7887149059506178, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7410851122058019, \"recall\": 0.7327586206896551, \"f1-score\": 0.7334234639648906, \"support\": 116.0}}", "eval_f1_macro": 48.5363019046534, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.34234639648905, "eval_loss": 1.137083888053894, "eval_runtime": 1.9707, "eval_samples_per_second": 58.861, "step": 2448 }, { "epoch": 154.0, "learning_rate": 7.666666666666666e-05, "loss": 0.0006, "step": 2464 }, { "epoch": 154.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 67.1067536864051, "eval_classification_report": "{\"accusation\": {\"precision\": 0.6875, \"recall\": 0.9166666666666666, \"f1-score\": 0.7857142857142857, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6857142857142857, \"recall\": 0.6486486486486487, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7890128968253968, \"recall\": 0.8046473893248086, \"f1-score\": 0.7900264310200789, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7429529282977559, \"recall\": 0.7327586206896551, \"f1-score\": 0.732582794679922, \"support\": 116.0}}", "eval_f1_macro": 48.61701113969717, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.25827946799221, "eval_loss": 1.142392873764038, "eval_runtime": 2.2302, "eval_samples_per_second": 52.013, "step": 2464 }, { "epoch": 155.0, "learning_rate": 7.5e-05, "loss": 0.0002, "step": 2480 }, { "epoch": 155.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 67.10759310976837, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6756756756756757, \"f1-score\": 0.684931506849315, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7899801587301587, \"recall\": 0.7976091010365204, \"f1-score\": 0.7887149059506178, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7410851122058019, \"recall\": 0.7327586206896551, \"f1-score\": 0.7334234639648906, \"support\": 116.0}}", "eval_f1_macro": 48.5363019046534, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.34234639648905, "eval_loss": 1.127716064453125, "eval_runtime": 1.9562, "eval_samples_per_second": 59.298, "step": 2480 }, { "epoch": 156.0, "learning_rate": 7.333333333333332e-05, "loss": 0.0001, "step": 2496 }, { "epoch": 156.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 67.14425762652212, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.6774193548387096, \"f1-score\": 0.7, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6756756756756757, \"f1-score\": 0.684931506849315, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7936155319776009, \"recall\": 0.7976091010365204, \"f1-score\": 0.7910780089493139, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7398576684248503, \"recall\": 0.7327586206896551, \"f1-score\": 0.7334358274050738, \"support\": 116.0}}", "eval_f1_macro": 48.681723627650086, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.34358274050739, "eval_loss": 1.127997875213623, "eval_runtime": 2.0527, "eval_samples_per_second": 56.51, "step": 2496 }, { "epoch": 157.0, "learning_rate": 7.166666666666667e-05, "loss": 0.0002, "step": 2512 }, { "epoch": 157.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 66.4337163915427, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6857142857142857, \"recall\": 0.6486486486486487, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7895070207570207, \"recall\": 0.794230722658142, \"f1-score\": 0.7873604776969747, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7306223948465328, \"recall\": 0.7241379310344828, \"f1-score\": 0.7245432688561426, \"support\": 116.0}}", "eval_f1_macro": 48.452952473659984, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.45432688561426, "eval_loss": 1.1359306573867798, "eval_runtime": 2.1072, "eval_samples_per_second": 55.049, "step": 2512 }, { "epoch": 158.0, "learning_rate": 7e-05, "loss": 0.0001, "step": 2528 }, { "epoch": 158.0, "eval_accuracy": 69.82758620689656, "eval_average_metrics": 64.12329403555103, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.65625, \"recall\": 0.5675675675675675, \"f1-score\": 0.6086956521739131, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.762747061965812, \"recall\": 0.7840955875230069, \"f1-score\": 0.7657600338997346, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7069006962864721, \"recall\": 0.6982758620689655, \"f1-score\": 0.6971430933458117, \"support\": 116.0}}", "eval_f1_macro": 47.123694393829815, "eval_f1_micro": 69.82758620689656, "eval_f1_weighted": 69.71430933458117, "eval_loss": 1.1599950790405273, "eval_runtime": 2.2991, "eval_samples_per_second": 50.454, "step": 2528 }, { "epoch": 159.0, "learning_rate": 6.833333333333333e-05, "loss": 0.0003, "step": 2544 }, { "epoch": 159.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 64.84977905236525, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.6774193548387096, \"f1-score\": 0.7, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6666666666666666, \"recall\": 0.5945945945945946, \"f1-score\": 0.6285714285714286, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7670663866784557, \"recall\": 0.7874739659013852, \"f1-score\": 0.7696789321789321, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7166739077410896, \"recall\": 0.7068965517241379, \"f1-score\": 0.7065494849977608, \"support\": 116.0}}", "eval_f1_macro": 47.36485736485736, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.65494849977608, "eval_loss": 1.1744282245635986, "eval_runtime": 2.0565, "eval_samples_per_second": 56.406, "step": 2544 }, { "epoch": 160.0, "learning_rate": 6.666666666666666e-05, "loss": 0.0004, "step": 2560 }, { "epoch": 160.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 64.8131145356115, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6666666666666666, \"recall\": 0.5945945945945946, \"f1-score\": 0.6285714285714286, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7634310134310134, \"recall\": 0.7874739659013852, \"f1-score\": 0.767315829180236, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7179013515220412, \"recall\": 0.7068965517241379, \"f1-score\": 0.7065371215575774, \"support\": 116.0}}", "eval_f1_macro": 47.21943564186068, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.65371215575774, "eval_loss": 1.1818971633911133, "eval_runtime": 2.324, "eval_samples_per_second": 49.915, "step": 2560 }, { "epoch": 161.0, "learning_rate": 6.5e-05, "loss": 0.0001, "step": 2576 }, { "epoch": 161.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 65.54092535499936, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6764705882352942, \"recall\": 0.6216216216216216, \"f1-score\": 0.6478873239436619, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7681287258493141, \"recall\": 0.7908523442797636, \"f1-score\": 0.7712645066336179, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7284518360836819, \"recall\": 0.7155172413793104, \"f1-score\": 0.7159782196668197, \"support\": 116.0}}", "eval_f1_macro": 47.462431177453404, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.59782196668198, "eval_loss": 1.2029889822006226, "eval_runtime": 2.1328, "eval_samples_per_second": 54.388, "step": 2576 }, { "epoch": 162.0, "learning_rate": 6.333333333333333e-05, "loss": 0.0003, "step": 2592 }, { "epoch": 162.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 64.70888807942681, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6764705882352942, \"recall\": 0.6216216216216216, \"f1-score\": 0.6478873239436619, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7492032069237953, \"recall\": 0.7868200862152475, \"f1-score\": 0.7613237063570644, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7163155049818335, \"recall\": 0.7068965517241379, \"f1-score\": 0.706055523509065, \"support\": 116.0}}", "eval_f1_macro": 46.85068962197319, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.60555235090649, "eval_loss": 1.2051033973693848, "eval_runtime": 2.3524, "eval_samples_per_second": 49.312, "step": 2592 }, { "epoch": 163.0, "learning_rate": 6.166666666666666e-05, "loss": 0.0002, "step": 2608 }, { "epoch": 163.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 64.70888807942681, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6764705882352942, \"recall\": 0.6216216216216216, \"f1-score\": 0.6478873239436619, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7492032069237953, \"recall\": 0.7868200862152475, \"f1-score\": 0.7613237063570644, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7163155049818335, \"recall\": 0.7068965517241379, \"f1-score\": 0.706055523509065, \"support\": 116.0}}", "eval_f1_macro": 46.85068962197319, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.60555235090649, "eval_loss": 1.1949589252471924, "eval_runtime": 2.0434, "eval_samples_per_second": 56.768, "step": 2608 }, { "epoch": 164.0, "learning_rate": 5.9999999999999995e-05, "loss": 0.0, "step": 2624 }, { "epoch": 164.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 64.8131145356115, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6666666666666666, \"recall\": 0.5945945945945946, \"f1-score\": 0.6285714285714286, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7634310134310134, \"recall\": 0.7874739659013852, \"f1-score\": 0.767315829180236, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7179013515220412, \"recall\": 0.7068965517241379, \"f1-score\": 0.7065371215575774, \"support\": 116.0}}", "eval_f1_macro": 47.21943564186068, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.65371215575774, "eval_loss": 1.189142107963562, "eval_runtime": 1.9698, "eval_samples_per_second": 58.89, "step": 2624 }, { "epoch": 165.0, "learning_rate": 5.8333333333333326e-05, "loss": 0.0004, "step": 2640 }, { "epoch": 165.0, "eval_accuracy": 69.82758620689656, "eval_average_metrics": 63.982577048752844, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.6451612903225806, \"f1-score\": 0.689655172413793, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6666666666666666, \"recall\": 0.5945945945945946, \"f1-score\": 0.6285714285714286, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7444164631664631, \"recall\": 0.7834417078368692, \"f1-score\": 0.7573968177416452, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7055746775574362, \"recall\": 0.6982758620689655, \"f1-score\": 0.6966610084327087, \"support\": 116.0}}", "eval_f1_macro": 46.6090349379474, "eval_f1_micro": 69.82758620689656, "eval_f1_weighted": 69.66610084327087, "eval_loss": 1.203819751739502, "eval_runtime": 2.1425, "eval_samples_per_second": 54.142, "step": 2640 }, { "epoch": 166.0, "learning_rate": 5.666666666666666e-05, "loss": 0.0001, "step": 2656 }, { "epoch": 166.0, "eval_accuracy": 69.82758620689656, "eval_average_metrics": 63.982577048752844, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.6451612903225806, \"f1-score\": 0.689655172413793, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6666666666666666, \"recall\": 0.5945945945945946, \"f1-score\": 0.6285714285714286, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7444164631664631, \"recall\": 0.7834417078368692, \"f1-score\": 0.7573968177416452, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7055746775574362, \"recall\": 0.6982758620689655, \"f1-score\": 0.6966610084327087, \"support\": 116.0}}", "eval_f1_macro": 46.6090349379474, "eval_f1_micro": 69.82758620689656, "eval_f1_weighted": 69.66610084327087, "eval_loss": 1.2178500890731812, "eval_runtime": 2.0767, "eval_samples_per_second": 55.858, "step": 2656 }, { "epoch": 167.0, "learning_rate": 5.499999999999999e-05, "loss": 0.0001, "step": 2672 }, { "epoch": 167.0, "eval_accuracy": 69.82758620689656, "eval_average_metrics": 63.982577048752844, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.6451612903225806, \"f1-score\": 0.689655172413793, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6666666666666666, \"recall\": 0.5945945945945946, \"f1-score\": 0.6285714285714286, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7444164631664631, \"recall\": 0.7834417078368692, \"f1-score\": 0.7573968177416452, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7055746775574362, \"recall\": 0.6982758620689655, \"f1-score\": 0.6966610084327087, \"support\": 116.0}}", "eval_f1_macro": 46.6090349379474, "eval_f1_micro": 69.82758620689656, "eval_f1_weighted": 69.66610084327087, "eval_loss": 1.2182705402374268, "eval_runtime": 1.9593, "eval_samples_per_second": 59.206, "step": 2672 }, { "epoch": 168.0, "learning_rate": 5.333333333333333e-05, "loss": 0.0001, "step": 2688 }, { "epoch": 168.0, "eval_accuracy": 69.82758620689656, "eval_average_metrics": 63.982577048752844, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.6451612903225806, \"f1-score\": 0.689655172413793, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6666666666666666, \"recall\": 0.5945945945945946, \"f1-score\": 0.6285714285714286, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7444164631664631, \"recall\": 0.7834417078368692, \"f1-score\": 0.7573968177416452, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7055746775574362, \"recall\": 0.6982758620689655, \"f1-score\": 0.6966610084327087, \"support\": 116.0}}", "eval_f1_macro": 46.6090349379474, "eval_f1_micro": 69.82758620689656, "eval_f1_weighted": 69.66610084327087, "eval_loss": 1.2183611392974854, "eval_runtime": 2.3149, "eval_samples_per_second": 50.11, "step": 2688 }, { "epoch": 169.0, "learning_rate": 5.1666666666666664e-05, "loss": 0.0001, "step": 2704 }, { "epoch": 169.0, "eval_accuracy": 69.82758620689656, "eval_average_metrics": 63.982577048752844, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.6451612903225806, \"f1-score\": 0.689655172413793, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6666666666666666, \"recall\": 0.5945945945945946, \"f1-score\": 0.6285714285714286, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7444164631664631, \"recall\": 0.7834417078368692, \"f1-score\": 0.7573968177416452, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7055746775574362, \"recall\": 0.6982758620689655, \"f1-score\": 0.6966610084327087, \"support\": 116.0}}", "eval_f1_macro": 46.6090349379474, "eval_f1_micro": 69.82758620689656, "eval_f1_weighted": 69.66610084327087, "eval_loss": 1.216108798980713, "eval_runtime": 2.042, "eval_samples_per_second": 56.806, "step": 2704 }, { "epoch": 170.0, "learning_rate": 4.9999999999999996e-05, "loss": 0.0001, "step": 2720 }, { "epoch": 170.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 64.8131145356115, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6666666666666666, \"recall\": 0.5945945945945946, \"f1-score\": 0.6285714285714286, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7634310134310134, \"recall\": 0.7874739659013852, \"f1-score\": 0.767315829180236, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7179013515220412, \"recall\": 0.7068965517241379, \"f1-score\": 0.7065371215575774, \"support\": 116.0}}", "eval_f1_macro": 47.21943564186068, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.65371215575774, "eval_loss": 1.203769326210022, "eval_runtime": 2.0637, "eval_samples_per_second": 56.21, "step": 2720 }, { "epoch": 171.0, "learning_rate": 4.8333333333333334e-05, "loss": 0.0001, "step": 2736 }, { "epoch": 171.0, "eval_accuracy": 70.6896551724138, "eval_average_metrics": 64.8131145356115, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6666666666666666, \"recall\": 0.5945945945945946, \"f1-score\": 0.6285714285714286, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7634310134310134, \"recall\": 0.7874739659013852, \"f1-score\": 0.767315829180236, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7179013515220412, \"recall\": 0.7068965517241379, \"f1-score\": 0.7065371215575774, \"support\": 116.0}}", "eval_f1_macro": 47.21943564186068, "eval_f1_micro": 70.6896551724138, "eval_f1_weighted": 70.65371215575774, "eval_loss": 1.1848161220550537, "eval_runtime": 2.1122, "eval_samples_per_second": 54.919, "step": 2736 }, { "epoch": 172.0, "learning_rate": 4.6666666666666665e-05, "loss": 0.0001, "step": 2752 }, { "epoch": 172.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 67.89123924112491, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.7027027027027027, \"f1-score\": 0.7027027027027027, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.8031652187902187, \"recall\": 0.8009874794148988, \"f1-score\": 0.7966552596806572, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.751262315270936, \"recall\": 0.7413793103448276, \"f1-score\": 0.7426415583826294, \"support\": 116.0}}", "eval_f1_macro": 49.02493905727121, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.26415583826291, "eval_loss": 1.1390397548675537, "eval_runtime": 2.0345, "eval_samples_per_second": 57.017, "step": 2752 }, { "epoch": 173.0, "learning_rate": 4.4999999999999996e-05, "loss": 0.0001, "step": 2768 }, { "epoch": 173.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 67.89123924112491, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.7027027027027027, \"f1-score\": 0.7027027027027027, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.8031652187902187, \"recall\": 0.8009874794148988, \"f1-score\": 0.7966552596806572, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.751262315270936, \"recall\": 0.7413793103448276, \"f1-score\": 0.7426415583826294, \"support\": 116.0}}", "eval_f1_macro": 49.02493905727121, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.26415583826291, "eval_loss": 1.1333969831466675, "eval_runtime": 2.0924, "eval_samples_per_second": 55.44, "step": 2768 }, { "epoch": 174.0, "learning_rate": 4.333333333333333e-05, "loss": 0.0004, "step": 2784 }, { "epoch": 174.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 67.92790375787868, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.6774193548387096, \"f1-score\": 0.7, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.7027027027027027, \"f1-score\": 0.7027027027027027, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.806800592037661, \"recall\": 0.8009874794148988, \"f1-score\": 0.7990183626793534, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7500348714899845, \"recall\": 0.7413793103448276, \"f1-score\": 0.7426539218228126, \"support\": 116.0}}", "eval_f1_macro": 49.1703607802679, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.26539218228126, "eval_loss": 1.1344233751296997, "eval_runtime": 2.0203, "eval_samples_per_second": 57.417, "step": 2784 }, { "epoch": 175.0, "learning_rate": 4.1666666666666665e-05, "loss": 0.0002, "step": 2800 }, { "epoch": 175.0, "eval_accuracy": 74.13793103448276, "eval_average_metrics": 67.92790375787868, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.6774193548387096, \"f1-score\": 0.7, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.7027027027027027, \"f1-score\": 0.7027027027027027, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.806800592037661, \"recall\": 0.8009874794148988, \"f1-score\": 0.7990183626793534, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7500348714899845, \"recall\": 0.7413793103448276, \"f1-score\": 0.7426539218228126, \"support\": 116.0}}", "eval_f1_macro": 49.1703607802679, "eval_f1_micro": 74.13793103448276, "eval_f1_weighted": 74.26539218228126, "eval_loss": 1.136767864227295, "eval_runtime": 1.964, "eval_samples_per_second": 59.062, "step": 2800 }, { "epoch": 176.0, "learning_rate": 3.9999999999999996e-05, "loss": 0.0001, "step": 2816 }, { "epoch": 176.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 66.38902324812496, "eval_classification_report": "{\"accusation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6756756756756757, \"f1-score\": 0.684931506849315, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7881944444444444, \"recall\": 0.7871924343698538, \"f1-score\": 0.7833935827198057, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7314176245210728, \"recall\": 0.7241379310344828, \"f1-score\": 0.7251967092592287, \"support\": 116.0}}", "eval_f1_macro": 48.208835859680356, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.51967092592288, "eval_loss": 1.1415574550628662, "eval_runtime": 1.9714, "eval_samples_per_second": 58.841, "step": 2816 }, { "epoch": 177.0, "learning_rate": 3.833333333333333e-05, "loss": 0.0, "step": 2832 }, { "epoch": 177.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 65.5615084791178, "eval_classification_report": "{\"accusation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6896551724137931, \"recall\": 0.6451612903225806, \"f1-score\": 0.6666666666666667, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6756756756756757, \"f1-score\": 0.684931506849315, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7690441981390257, \"recall\": 0.7831601763053376, \"f1-score\": 0.7735184851397902, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7188008417794387, \"recall\": 0.7155172413793104, \"f1-score\": 0.7154144809354516, \"support\": 116.0}}", "eval_f1_macro": 47.601137547064006, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.54144809354516, "eval_loss": 1.1445733308792114, "eval_runtime": 1.9924, "eval_samples_per_second": 58.22, "step": 2832 }, { "epoch": 178.0, "learning_rate": 3.666666666666666e-05, "loss": 0.0, "step": 2848 }, { "epoch": 178.0, "eval_accuracy": 71.55172413793103, "eval_average_metrics": 65.5615084791178, "eval_classification_report": "{\"accusation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6896551724137931, \"recall\": 0.6451612903225806, \"f1-score\": 0.6666666666666667, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6756756756756757, \"f1-score\": 0.684931506849315, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7690441981390257, \"recall\": 0.7831601763053376, \"f1-score\": 0.7735184851397902, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7188008417794387, \"recall\": 0.7155172413793104, \"f1-score\": 0.7154144809354516, \"support\": 116.0}}", "eval_f1_macro": 47.601137547064006, "eval_f1_micro": 71.55172413793103, "eval_f1_weighted": 71.54144809354516, "eval_loss": 1.144472360610962, "eval_runtime": 2.2321, "eval_samples_per_second": 51.969, "step": 2848 }, { "epoch": 179.0, "learning_rate": 3.5e-05, "loss": 0.0098, "step": 2864 }, { "epoch": 179.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 66.38902324812496, "eval_classification_report": "{\"accusation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6756756756756757, \"f1-score\": 0.684931506849315, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7881944444444444, \"recall\": 0.7871924343698538, \"f1-score\": 0.7833935827198057, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7314176245210728, \"recall\": 0.7241379310344828, \"f1-score\": 0.7251967092592287, \"support\": 116.0}}", "eval_f1_macro": 48.208835859680356, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.51967092592288, "eval_loss": 1.142303705215454, "eval_runtime": 1.9646, "eval_samples_per_second": 59.044, "step": 2864 }, { "epoch": 180.0, "learning_rate": 3.333333333333333e-05, "loss": 0.0001, "step": 2880 }, { "epoch": 180.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 66.38902324812496, "eval_classification_report": "{\"accusation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6756756756756757, \"f1-score\": 0.684931506849315, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7881944444444444, \"recall\": 0.7871924343698538, \"f1-score\": 0.7833935827198057, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7314176245210728, \"recall\": 0.7241379310344828, \"f1-score\": 0.7251967092592287, \"support\": 116.0}}", "eval_f1_macro": 48.208835859680356, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.51967092592288, "eval_loss": 1.133960485458374, "eval_runtime": 2.0714, "eval_samples_per_second": 56.001, "step": 2880 }, { "epoch": 181.0, "learning_rate": 3.1666666666666666e-05, "loss": 0.0001, "step": 2896 }, { "epoch": 181.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 66.95417031249403, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.76, \"recall\": 0.6129032258064516, \"f1-score\": 0.6785714285714285, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7105263157894737, \"recall\": 0.7297297297297297, \"f1-score\": 0.7200000000000001, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7741431704260652, \"recall\": 0.7963013416642448, \"f1-score\": 0.7812629424239331, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7391888125486129, \"recall\": 0.7327586206896551, \"f1-score\": 0.7318723757826459, \"support\": 116.0}}", "eval_f1_macro": 48.0777195337805, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.18723757826459, "eval_loss": 1.1398240327835083, "eval_runtime": 2.1727, "eval_samples_per_second": 53.39, "step": 2896 }, { "epoch": 182.0, "learning_rate": 2.9999999999999997e-05, "loss": 0.0006, "step": 2912 }, { "epoch": 182.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 66.95417031249403, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.76, \"recall\": 0.6129032258064516, \"f1-score\": 0.6785714285714285, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7105263157894737, \"recall\": 0.7297297297297297, \"f1-score\": 0.7200000000000001, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7741431704260652, \"recall\": 0.7963013416642448, \"f1-score\": 0.7812629424239331, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7391888125486129, \"recall\": 0.7327586206896551, \"f1-score\": 0.7318723757826459, \"support\": 116.0}}", "eval_f1_macro": 48.0777195337805, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.18723757826459, "eval_loss": 1.1363673210144043, "eval_runtime": 2.1014, "eval_samples_per_second": 55.2, "step": 2912 }, { "epoch": 183.0, "learning_rate": 2.833333333333333e-05, "loss": 0.0016, "step": 2928 }, { "epoch": 183.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 66.95417031249403, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.76, \"recall\": 0.6129032258064516, \"f1-score\": 0.6785714285714285, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7105263157894737, \"recall\": 0.7297297297297297, \"f1-score\": 0.7200000000000001, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7741431704260652, \"recall\": 0.7963013416642448, \"f1-score\": 0.7812629424239331, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7391888125486129, \"recall\": 0.7327586206896551, \"f1-score\": 0.7318723757826459, \"support\": 116.0}}", "eval_f1_macro": 48.0777195337805, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.18723757826459, "eval_loss": 1.1402283906936646, "eval_runtime": 2.0837, "eval_samples_per_second": 55.671, "step": 2928 }, { "epoch": 184.0, "learning_rate": 2.6666666666666667e-05, "loss": 0.0021, "step": 2944 }, { "epoch": 184.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 66.18929941059953, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.76, \"recall\": 0.6129032258064516, \"f1-score\": 0.6785714285714285, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.7027027027027027, \"f1-score\": 0.7027027027027027, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7640743096993097, \"recall\": 0.7929229632858665, \"f1-score\": 0.7738376223670341, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7310507165248544, \"recall\": 0.7241379310344828, \"f1-score\": 0.7230883467445334, \"support\": 116.0}}", "eval_f1_macro": 47.620776761048255, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.30883467445334, "eval_loss": 1.145507574081421, "eval_runtime": 2.1519, "eval_samples_per_second": 53.906, "step": 2944 }, { "epoch": 185.0, "learning_rate": 2.4999999999999998e-05, "loss": 0.0, "step": 2960 }, { "epoch": 185.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 67.0801291345566, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6756756756756757, \"f1-score\": 0.684931506849315, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7889384920634921, \"recall\": 0.7976091010365204, \"f1-score\": 0.7859429880741948, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7457067597153804, \"recall\": 0.7327586206896551, \"f1-score\": 0.7340307005726803, \"support\": 116.0}}", "eval_f1_macro": 48.36572234302737, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.40307005726802, "eval_loss": 1.1441426277160645, "eval_runtime": 2.286, "eval_samples_per_second": 50.743, "step": 2960 }, { "epoch": 186.0, "learning_rate": 2.3333333333333332e-05, "loss": 0.0, "step": 2976 }, { "epoch": 186.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 67.0801291345566, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6756756756756757, \"f1-score\": 0.684931506849315, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7889384920634921, \"recall\": 0.7976091010365204, \"f1-score\": 0.7859429880741948, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7457067597153804, \"recall\": 0.7327586206896551, \"f1-score\": 0.7340307005726803, \"support\": 116.0}}", "eval_f1_macro": 48.36572234302737, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.40307005726802, "eval_loss": 1.1459873914718628, "eval_runtime": 2.0524, "eval_samples_per_second": 56.519, "step": 2976 }, { "epoch": 187.0, "learning_rate": 2.1666666666666664e-05, "loss": 0.0001, "step": 2992 }, { "epoch": 187.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 67.0801291345566, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6756756756756757, \"f1-score\": 0.684931506849315, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7889384920634921, \"recall\": 0.7976091010365204, \"f1-score\": 0.7859429880741948, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7457067597153804, \"recall\": 0.7327586206896551, \"f1-score\": 0.7340307005726803, \"support\": 116.0}}", "eval_f1_macro": 48.36572234302737, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.40307005726802, "eval_loss": 1.148589849472046, "eval_runtime": 2.3314, "eval_samples_per_second": 49.755, "step": 2992 }, { "epoch": 188.0, "learning_rate": 1.9999999999999998e-05, "loss": 0.0006, "step": 3008 }, { "epoch": 188.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 67.0801291345566, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6756756756756757, \"f1-score\": 0.684931506849315, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7889384920634921, \"recall\": 0.7976091010365204, \"f1-score\": 0.7859429880741948, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7457067597153804, \"recall\": 0.7327586206896551, \"f1-score\": 0.7340307005726803, \"support\": 116.0}}", "eval_f1_macro": 48.36572234302737, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.40307005726802, "eval_loss": 1.1534448862075806, "eval_runtime": 2.0097, "eval_samples_per_second": 57.72, "step": 3008 }, { "epoch": 189.0, "learning_rate": 1.833333333333333e-05, "loss": 0.0001, "step": 3024 }, { "epoch": 189.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 67.0801291345566, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6756756756756757, \"f1-score\": 0.684931506849315, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7889384920634921, \"recall\": 0.7976091010365204, \"f1-score\": 0.7859429880741948, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7457067597153804, \"recall\": 0.7327586206896551, \"f1-score\": 0.7340307005726803, \"support\": 116.0}}", "eval_f1_macro": 48.36572234302737, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.40307005726802, "eval_loss": 1.1526384353637695, "eval_runtime": 2.1314, "eval_samples_per_second": 54.425, "step": 3024 }, { "epoch": 190.0, "learning_rate": 1.6666666666666664e-05, "loss": 0.0001, "step": 3040 }, { "epoch": 190.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 66.3627208342036, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6857142857142857, \"recall\": 0.6486486486486487, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.784375, \"recall\": 0.794230722658142, \"f1-score\": 0.782125692519511, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7354987684729063, \"recall\": 0.7241379310344828, \"f1-score\": 0.7249248528256333, \"support\": 116.0}}", "eval_f1_macro": 48.13081184735453, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.49248528256334, "eval_loss": 1.1524851322174072, "eval_runtime": 2.1944, "eval_samples_per_second": 52.862, "step": 3040 }, { "epoch": 191.0, "learning_rate": 1.4999999999999999e-05, "loss": 0.0001, "step": 3056 }, { "epoch": 191.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 67.12167806340173, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6756756756756757, \"f1-score\": 0.684931506849315, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7930420274170275, \"recall\": 0.7976091010365204, \"f1-score\": 0.7891707023042469, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7429855824252376, \"recall\": 0.7327586206896551, \"f1-score\": 0.7337063720464534, \"support\": 116.0}}", "eval_f1_macro": 48.56435091103058, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.37063720464533, "eval_loss": 1.142797589302063, "eval_runtime": 1.9594, "eval_samples_per_second": 59.201, "step": 3056 }, { "epoch": 192.0, "learning_rate": 1.3333333333333333e-05, "loss": 0.0002, "step": 3072 }, { "epoch": 192.0, "eval_accuracy": 73.27586206896551, "eval_average_metrics": 67.12167806340173, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6756756756756757, \"f1-score\": 0.684931506849315, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7930420274170275, \"recall\": 0.7976091010365204, \"f1-score\": 0.7891707023042469, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7429855824252376, \"recall\": 0.7327586206896551, \"f1-score\": 0.7337063720464534, \"support\": 116.0}}", "eval_f1_macro": 48.56435091103058, "eval_f1_micro": 73.27586206896551, "eval_f1_weighted": 73.37063720464533, "eval_loss": 1.14202880859375, "eval_runtime": 2.1118, "eval_samples_per_second": 54.931, "step": 3072 }, { "epoch": 193.0, "learning_rate": 1.1666666666666666e-05, "loss": 0.0, "step": 3088 }, { "epoch": 193.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 66.3627208342036, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6857142857142857, \"recall\": 0.6486486486486487, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.784375, \"recall\": 0.794230722658142, \"f1-score\": 0.782125692519511, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7354987684729063, \"recall\": 0.7241379310344828, \"f1-score\": 0.7249248528256333, \"support\": 116.0}}", "eval_f1_macro": 48.13081184735453, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.49248528256334, "eval_loss": 1.1449861526489258, "eval_runtime": 2.0551, "eval_samples_per_second": 56.444, "step": 3088 }, { "epoch": 194.0, "learning_rate": 9.999999999999999e-06, "loss": 0.0, "step": 3104 }, { "epoch": 194.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 66.3627208342036, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6857142857142857, \"recall\": 0.6486486486486487, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.784375, \"recall\": 0.794230722658142, \"f1-score\": 0.782125692519511, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7354987684729063, \"recall\": 0.7241379310344828, \"f1-score\": 0.7249248528256333, \"support\": 116.0}}", "eval_f1_macro": 48.13081184735453, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.49248528256334, "eval_loss": 1.1457918882369995, "eval_runtime": 2.0705, "eval_samples_per_second": 56.025, "step": 3104 }, { "epoch": 195.0, "learning_rate": 8.333333333333332e-06, "loss": 0.0006, "step": 3120 }, { "epoch": 195.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 66.3627208342036, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6857142857142857, \"recall\": 0.6486486486486487, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.784375, \"recall\": 0.794230722658142, \"f1-score\": 0.782125692519511, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7354987684729063, \"recall\": 0.7241379310344828, \"f1-score\": 0.7249248528256333, \"support\": 116.0}}", "eval_f1_macro": 48.13081184735453, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.49248528256334, "eval_loss": 1.1468638181686401, "eval_runtime": 1.9389, "eval_samples_per_second": 59.828, "step": 3120 }, { "epoch": 196.0, "learning_rate": 6.666666666666667e-06, "loss": 0.0005, "step": 3136 }, { "epoch": 196.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 66.3627208342036, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6857142857142857, \"recall\": 0.6486486486486487, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.784375, \"recall\": 0.794230722658142, \"f1-score\": 0.782125692519511, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7354987684729063, \"recall\": 0.7241379310344828, \"f1-score\": 0.7249248528256333, \"support\": 116.0}}", "eval_f1_macro": 48.13081184735453, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.49248528256334, "eval_loss": 1.150787591934204, "eval_runtime": 2.0877, "eval_samples_per_second": 55.564, "step": 3136 }, { "epoch": 197.0, "learning_rate": 4.9999999999999996e-06, "loss": 0.0001, "step": 3152 }, { "epoch": 197.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 66.3627208342036, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6857142857142857, \"recall\": 0.6486486486486487, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.784375, \"recall\": 0.794230722658142, \"f1-score\": 0.782125692519511, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7354987684729063, \"recall\": 0.7241379310344828, \"f1-score\": 0.7249248528256333, \"support\": 116.0}}", "eval_f1_macro": 48.13081184735453, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.49248528256334, "eval_loss": 1.1514496803283691, "eval_runtime": 2.0764, "eval_samples_per_second": 55.866, "step": 3152 }, { "epoch": 198.0, "learning_rate": 3.3333333333333333e-06, "loss": 0.0001, "step": 3168 }, { "epoch": 198.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 66.3627208342036, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6857142857142857, \"recall\": 0.6486486486486487, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.784375, \"recall\": 0.794230722658142, \"f1-score\": 0.782125692519511, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7354987684729063, \"recall\": 0.7241379310344828, \"f1-score\": 0.7249248528256333, \"support\": 116.0}}", "eval_f1_macro": 48.13081184735453, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.49248528256334, "eval_loss": 1.1519808769226074, "eval_runtime": 2.0667, "eval_samples_per_second": 56.128, "step": 3168 }, { "epoch": 199.0, "learning_rate": 1.6666666666666667e-06, "loss": 0.0001, "step": 3184 }, { "epoch": 199.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 66.3627208342036, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6857142857142857, \"recall\": 0.6486486486486487, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.784375, \"recall\": 0.794230722658142, \"f1-score\": 0.782125692519511, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7354987684729063, \"recall\": 0.7241379310344828, \"f1-score\": 0.7249248528256333, \"support\": 116.0}}", "eval_f1_macro": 48.13081184735453, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.49248528256334, "eval_loss": 1.1532700061798096, "eval_runtime": 2.077, "eval_samples_per_second": 55.849, "step": 3184 }, { "epoch": 200.0, "learning_rate": 0.0, "loss": 0.0001, "step": 3200 }, { "epoch": 200.0, "eval_accuracy": 72.41379310344827, "eval_average_metrics": 66.3627208342036, "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6857142857142857, \"recall\": 0.6486486486486487, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.784375, \"recall\": 0.794230722658142, \"f1-score\": 0.782125692519511, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7354987684729063, \"recall\": 0.7241379310344828, \"f1-score\": 0.7249248528256333, \"support\": 116.0}}", "eval_f1_macro": 48.13081184735453, "eval_f1_micro": 72.41379310344828, "eval_f1_weighted": 72.49248528256334, "eval_loss": 1.1530934572219849, "eval_runtime": 2.0419, "eval_samples_per_second": 56.811, "step": 3200 }, { "epoch": 200.0, "step": 3200, "total_flos": 1.5716799379144704e+16, "train_runtime": 2386.5323, "train_samples_per_second": 1.341 } ], "max_steps": 3200, "num_train_epochs": 200, "total_flos": 1.5716799379144704e+16, "trial_name": null, "trial_params": null }