diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,3823 @@ +{ + "best_metric": 78.31433183733813, + "best_model_checkpoint": "/mounts/work/faeze/attempt/new_setting_outputs/source_adapter/crisis_9/3407/checkpoint-512", + "epoch": 200.0, + "global_step": 3200, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 1.0, + "learning_rate": 1.4999999999999999e-05, + "loss": 9.307, + "step": 16 + }, + { + "epoch": 1.0, + "eval_accuracy": 0.0, + "eval_average_metrics": 0.0, + "eval_classification_report": "{\"\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"\\\"media\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"How Dar\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"-\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \".\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"0:\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \":\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": D\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": Lots\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": apparently\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": good\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": please\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": very\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"????\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Anders\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Emers\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Can use passport\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"False\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"No Friday Im\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"No they have\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"PPC 2019\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"SCP-04\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Shawn\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"True\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a friend\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"canada is under\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 37.0}, \"instagram\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"vac\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"weather\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.0, \"macro avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}}", + "eval_f1_macro": 0.0, + "eval_f1_micro": 0.0, + "eval_f1_weighted": 0.0, + "eval_loss": 9.491299629211426, + "eval_runtime": 2.2405, + "eval_samples_per_second": 51.774, + "step": 16 + }, + { + "epoch": 2.0, + "learning_rate": 2.9999999999999997e-05, + "loss": 8.9068, + "step": 32 + }, + { + "epoch": 2.0, + "eval_accuracy": 0.0, + "eval_average_metrics": 0.0, + "eval_classification_report": "{\"\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"\\\"media\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"How Dar\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"-\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": Lots\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": apparently\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": good\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": please\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": very\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Anders\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Emers\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Can use passport\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Climate change is\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Ddx\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"False\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"No Friday Im\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"No they have\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"PPC 2019\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Shawn\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"True\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a friend\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"canada is under\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 37.0}, \"instagram\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"negative\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"weather\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.0, \"macro avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}}", + "eval_f1_macro": 0.0, + "eval_f1_micro": 0.0, + "eval_f1_weighted": 0.0, + "eval_loss": 9.213845252990723, + "eval_runtime": 2.3697, + "eval_samples_per_second": 48.951, + "step": 32 + }, + { + "epoch": 3.0, + "learning_rate": 4.4999999999999996e-05, + "loss": 8.1759, + "step": 48 + }, + { + "epoch": 3.0, + "eval_accuracy": 0.0, + "eval_average_metrics": 0.0, + "eval_classification_report": "{\"\\\"How Dar\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"-\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": Lots\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": good\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": please\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": very\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Can use passport\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Climate change is\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Ddx\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"False\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"No Friday Im\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"PPC 2019\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"True\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a friend\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"canada is under\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 31.0}, \"entail\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 37.0}, \"instagram\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"negative\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_e\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"weather\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.0, \"macro avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}}", + "eval_f1_macro": 0.0, + "eval_f1_micro": 0.0, + "eval_f1_weighted": 0.0, + "eval_loss": 8.574390411376953, + "eval_runtime": 2.1785, + "eval_samples_per_second": 53.247, + "step": 48 + }, + { + "epoch": 4.0, + "learning_rate": 5.9999999999999995e-05, + "loss": 6.8531, + "step": 64 + }, + { + "epoch": 4.0, + "eval_accuracy": 0.0, + "eval_average_metrics": 0.0, + "eval_classification_report": "{\"\\\"How Dar\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"Reality\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"????\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Australian\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Ddx\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"False\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"SCP-04\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"True\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a friend\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a real\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"canada\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 31.0}, \"classification: good\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"classification: very\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 37.0}, \"negative\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_e\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_en\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"panne\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"please show us\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"vac\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"weather\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.0, \"macro avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}}", + "eval_f1_macro": 0.0, + "eval_f1_micro": 0.0, + "eval_f1_weighted": 0.0, + "eval_loss": 6.66550874710083, + "eval_runtime": 2.3739, + "eval_samples_per_second": 48.865, + "step": 64 + }, + { + "epoch": 5.0, + "learning_rate": 7.5e-05, + "loss": 4.9379, + "step": 80 + }, + { + "epoch": 5.0, + "eval_accuracy": 0.0, + "eval_average_metrics": 0.0, + "eval_classification_report": "{\"\\\"Reality\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"-54\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \".\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"0\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"1\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \":D\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"?\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"????\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Australian\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Bruh\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Ddx\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Florida\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"India\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"PPC\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"SCP-04\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"True\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"canada\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 31.0}, \"classification\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"classification: good\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"dal\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"eder\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"entail\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 37.0}, \"instagram\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"microwave\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"name\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"negative\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_e\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_en\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"trud\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"u\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"vac\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"weather\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.0, \"macro avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}}", + "eval_f1_macro": 0.0, + "eval_f1_micro": 0.0, + "eval_f1_weighted": 0.0, + "eval_loss": 3.4962198734283447, + "eval_runtime": 2.0892, + "eval_samples_per_second": 55.523, + "step": 80 + }, + { + "epoch": 6.0, + "learning_rate": 8.999999999999999e-05, + "loss": 2.6265, + "step": 96 + }, + { + "epoch": 6.0, + "eval_accuracy": 12.068965517241379, + "eval_average_metrics": 10.647060714963407, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"challenge\": {\"precision\": 0.2857142857142857, \"recall\": 0.3870967741935484, \"f1-score\": 0.3287671232876712, \"support\": 31.0}, \"entail\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 37.0}, \"negative\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_en\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.10526315789473684, \"recall\": 0.2222222222222222, \"f1-score\": 0.14285714285714285, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.1206896551724138, \"macro avg\": {\"precision\": 0.03007518796992481, \"recall\": 0.04687069203198235, \"f1-score\": 0.03627878970344724, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.0845216489499611, \"recall\": 0.1206896551724138, \"f1-score\": 0.09894392334165598, \"support\": 116.0}}", + "eval_f1_macro": 5.240269623831267, + "eval_f1_micro": 15.384615384615385, + "eval_f1_weighted": 9.894392334165598, + "eval_loss": 1.4142650365829468, + "eval_runtime": 1.9052, + "eval_samples_per_second": 60.887, + "step": 96 + }, + { + "epoch": 7.0, + "learning_rate": 0.00010499999999999999, + "loss": 1.1522, + "step": 112 + }, + { + "epoch": 7.0, + "eval_accuracy": 31.896551724137932, + "eval_average_metrics": 21.324233716475096, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.3333333333333333, \"recall\": 1.0, \"f1-score\": 0.5, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"accuracy\": 0.31896551724137934, \"macro avg\": {\"precision\": 0.041666666666666664, \"recall\": 0.125, \"f1-score\": 0.0625, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.10632183908045977, \"recall\": 0.31896551724137934, \"f1-score\": 0.15948275862068967, \"support\": 116.0}}", + "eval_f1_macro": 5.555555555555555, + "eval_f1_micro": 31.896551724137932, + "eval_f1_weighted": 15.948275862068966, + "eval_loss": 0.7661405205726624, + "eval_runtime": 2.2446, + "eval_samples_per_second": 51.68, + "step": 112 + }, + { + "epoch": 8.0, + "learning_rate": 0.00011999999999999999, + "loss": 0.7426, + "step": 128 + }, + { + "epoch": 8.0, + "eval_accuracy": 38.793103448275865, + "eval_average_metrics": 30.810095419319552, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6, \"recall\": 0.3333333333333333, \"f1-score\": 0.42857142857142855, \"support\": 9.0}, \"challenge\": {\"precision\": 0.3611111111111111, \"recall\": 0.41935483870967744, \"f1-score\": 0.3880597014925373, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.4027777777777778, \"recall\": 0.7837837837837838, \"f1-score\": 0.5321100917431194, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"accuracy\": 0.3879310344827586, \"macro avg\": {\"precision\": 0.17048611111111112, \"recall\": 0.19205899447834932, \"f1-score\": 0.16859265272588567, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.2715277777777778, \"recall\": 0.3879310344827586, \"f1-score\": 0.3066816120509218, \"support\": 116.0}}", + "eval_f1_macro": 14.986013575634283, + "eval_f1_micro": 38.793103448275865, + "eval_f1_weighted": 30.668161205092183, + "eval_loss": 0.6721057891845703, + "eval_runtime": 2.0959, + "eval_samples_per_second": 55.347, + "step": 128 + }, + { + "epoch": 9.0, + "learning_rate": 0.000135, + "loss": 0.7745, + "step": 144 + }, + { + "epoch": 9.0, + "eval_accuracy": 31.03448275862069, + "eval_average_metrics": 22.98618128359508, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.3333333333333333, \"f1-score\": 0.46153846153846156, \"support\": 9.0}, \"challenge\": {\"precision\": 0.28440366972477066, \"recall\": 1.0, \"f1-score\": 0.44285714285714284, \"support\": 31.0}, \"informing statement\": {\"precision\": 1.0, \"recall\": 0.05405405405405406, \"f1-score\": 0.10256410256410257, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"accuracy\": 0.3103448275862069, \"macro avg\": {\"precision\": 0.25430045871559637, \"recall\": 0.17342342342342343, \"f1-score\": 0.12586996336996337, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.45315960139196454, \"recall\": 0.3103448275862069, \"f1-score\": 0.18687318428697738, \"support\": 116.0}}", + "eval_f1_macro": 11.188441188441187, + "eval_f1_micro": 31.03448275862069, + "eval_f1_weighted": 18.68731842869774, + "eval_loss": 0.7225409150123596, + "eval_runtime": 2.094, + "eval_samples_per_second": 55.396, + "step": 144 + }, + { + "epoch": 10.0, + "learning_rate": 0.00015, + "loss": 0.725, + "step": 160 + }, + { + "epoch": 10.0, + "eval_accuracy": 33.62068965517241, + "eval_average_metrics": 26.30665126493825, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.3333333333333333, \"f1-score\": 0.46153846153846156, \"support\": 9.0}, \"challenge\": {\"precision\": 0.3118279569892473, \"recall\": 0.9354838709677419, \"f1-score\": 0.46774193548387094, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.4666666666666667, \"recall\": 0.1891891891891892, \"f1-score\": 0.2692307692307693, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"accuracy\": 0.33620689655172414, \"macro avg\": {\"precision\": 0.19106182795698923, \"recall\": 0.18225079918628306, \"f1-score\": 0.14981389578163773, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.29037356321839075, \"recall\": 0.33620689655172414, \"f1-score\": 0.24668435013262602, \"support\": 116.0}}", + "eval_f1_macro": 13.316790736145576, + "eval_f1_micro": 33.62068965517241, + "eval_f1_weighted": 24.668435013262602, + "eval_loss": 0.6647224426269531, + "eval_runtime": 2.1313, + "eval_samples_per_second": 54.427, + "step": 160 + }, + { + "epoch": 11.0, + "learning_rate": 0.000165, + "loss": 0.7576, + "step": 176 + }, + { + "epoch": 11.0, + "eval_accuracy": 33.62068965517241, + "eval_average_metrics": 25.33956739840334, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.5, \"recall\": 0.5555555555555556, \"f1-score\": 0.5263157894736842, \"support\": 9.0}, \"challenge\": {\"precision\": 0.3163265306122449, \"recall\": 1.0, \"f1-score\": 0.4806201550387597, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.5, \"recall\": 0.08108108108108109, \"f1-score\": 0.13953488372093023, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"accuracy\": 0.33620689655172414, \"macro avg\": {\"precision\": 0.16454081632653061, \"recall\": 0.2045795795795796, \"f1-score\": 0.14330885352917178, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.28281140042223785, \"recall\": 0.33620689655172414, \"f1-score\": 0.21378325525119937, \"support\": 116.0}}", + "eval_f1_macro": 12.738564758148602, + "eval_f1_micro": 33.62068965517241, + "eval_f1_weighted": 21.378325525119937, + "eval_loss": 0.6687206029891968, + "eval_runtime": 2.1419, + "eval_samples_per_second": 54.157, + "step": 176 + }, + { + "epoch": 12.0, + "learning_rate": 0.00017999999999999998, + "loss": 0.6978, + "step": 192 + }, + { + "epoch": 12.0, + "eval_accuracy": 36.206896551724135, + "eval_average_metrics": 28.099909132802715, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6, \"recall\": 0.3333333333333333, \"f1-score\": 0.42857142857142855, \"support\": 9.0}, \"challenge\": {\"precision\": 0.5, \"recall\": 0.16129032258064516, \"f1-score\": 0.24390243902439024, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.3695652173913043, \"recall\": 0.918918918918919, \"f1-score\": 0.5271317829457364, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"accuracy\": 0.3620689655172414, \"macro avg\": {\"precision\": 0.18369565217391304, \"recall\": 0.17669282185411217, \"f1-score\": 0.14995070631769442, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.2980509745127436, \"recall\": 0.3620689655172414, \"f1-score\": 0.2665689175507862, \"support\": 116.0}}", + "eval_f1_macro": 13.328951672683944, + "eval_f1_micro": 36.20689655172414, + "eval_f1_weighted": 26.656891755078625, + "eval_loss": 0.6495026350021362, + "eval_runtime": 2.1678, + "eval_samples_per_second": 53.51, + "step": 192 + }, + { + "epoch": 13.0, + "learning_rate": 0.000195, + "loss": 0.7123, + "step": 208 + }, + { + "epoch": 13.0, + "eval_accuracy": 36.206896551724135, + "eval_average_metrics": 31.268398801131113, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.07407407407407407, \"recall\": 0.16666666666666666, \"f1-score\": 0.10256410256410256, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.4146341463414634, \"recall\": 0.5483870967741935, \"f1-score\": 0.4722222222222222, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.47058823529411764, \"recall\": 0.43243243243243246, \"f1-score\": 0.4507042253521127, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"accuracy\": 0.3620689655172414, \"macro avg\": {\"precision\": 0.19945751150916147, \"recall\": 0.2406579967063838, \"f1-score\": 0.2156863187673047, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.31794469705714945, \"recall\": 0.3620689655172414, \"f1-score\": 0.3348768487731576, \"support\": 116.0}}", + "eval_f1_macro": 19.172117223760416, + "eval_f1_micro": 36.20689655172414, + "eval_f1_weighted": 33.48768487731576, + "eval_loss": 0.6175509691238403, + "eval_runtime": 2.2362, + "eval_samples_per_second": 51.874, + "step": 208 + }, + { + "epoch": 14.0, + "learning_rate": 0.00020999999999999998, + "loss": 0.6115, + "step": 224 + }, + { + "epoch": 14.0, + "eval_accuracy": 45.689655172413794, + "eval_average_metrics": 37.07872659990352, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.4262295081967213, \"recall\": 0.8387096774193549, \"f1-score\": 0.5652173913043479, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.5, \"recall\": 0.5405405405405406, \"f1-score\": 0.5194805194805195, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"accuracy\": 0.45689655172413796, \"macro avg\": {\"precision\": 0.1953241430700447, \"recall\": 0.2696284994672091, \"f1-score\": 0.22308723884810844, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.3227619610463025, \"recall\": 0.45689655172413796, \"f1-score\": 0.3710561926828794, \"support\": 116.0}}", + "eval_f1_macro": 19.829976786498527, + "eval_f1_micro": 45.689655172413794, + "eval_f1_weighted": 37.105619268287946, + "eval_loss": 0.5956699252128601, + "eval_runtime": 2.0913, + "eval_samples_per_second": 55.468, + "step": 224 + }, + { + "epoch": 15.0, + "learning_rate": 0.000225, + "loss": 0.632, + "step": 240 + }, + { + "epoch": 15.0, + "eval_accuracy": 43.103448275862064, + "eval_average_metrics": 35.833906806036026, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.5333333333333333, \"recall\": 0.8888888888888888, \"f1-score\": 0.6666666666666667, \"support\": 9.0}, \"challenge\": {\"precision\": 0.3888888888888889, \"recall\": 0.9032258064516129, \"f1-score\": 0.5436893203883495, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.5217391304347826, \"recall\": 0.32432432432432434, \"f1-score\": 0.4, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.1111111111111111, \"f1-score\": 0.19999999999999998, \"support\": 9.0}, \"request\": {\"precision\": 0.25, \"recall\": 0.125, \"f1-score\": 0.16666666666666666, \"support\": 8.0}, \"accuracy\": 0.43103448275862066, \"macro avg\": {\"precision\": 0.3367451690821256, \"recall\": 0.2940687663469922, \"f1-score\": 0.24712783171521036, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.4065508912210561, \"recall\": 0.43103448275862066, \"f1-score\": 0.3516181229773463, \"support\": 116.0}}", + "eval_f1_macro": 21.966918374685367, + "eval_f1_micro": 43.103448275862064, + "eval_f1_weighted": 35.161812297734635, + "eval_loss": 0.5922260880470276, + "eval_runtime": 2.1966, + "eval_samples_per_second": 52.809, + "step": 240 + }, + { + "epoch": 16.0, + "learning_rate": 0.00023999999999999998, + "loss": 0.613, + "step": 256 + }, + { + "epoch": 16.0, + "eval_accuracy": 46.55172413793103, + "eval_average_metrics": 37.72845991532228, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"challenge\": {\"precision\": 0.5238095238095238, \"recall\": 0.7096774193548387, \"f1-score\": 0.6027397260273972, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.4, \"recall\": 0.6486486486486487, \"f1-score\": 0.49484536082474234, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"accuracy\": 0.46551724137931033, \"macro avg\": {\"precision\": 0.19239926739926738, \"recall\": 0.28090186961154706, \"f1-score\": 0.22810722676560835, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.31531514462548943, \"recall\": 0.46551724137931033, \"f1-score\": 0.37534193450706316, \"support\": 116.0}}", + "eval_f1_macro": 20.27619793472074, + "eval_f1_micro": 46.55172413793103, + "eval_f1_weighted": 37.53419345070632, + "eval_loss": 0.5560970306396484, + "eval_runtime": 2.0931, + "eval_samples_per_second": 55.419, + "step": 256 + }, + { + "epoch": 17.0, + "learning_rate": 0.00025499999999999996, + "loss": 0.544, + "step": 272 + }, + { + "epoch": 17.0, + "eval_accuracy": 52.58620689655172, + "eval_average_metrics": 45.3769094880828, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"challenge\": {\"precision\": 0.5217391304347826, \"recall\": 0.7741935483870968, \"f1-score\": 0.6233766233766235, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.4897959183673469, \"recall\": 0.6486486486486487, \"f1-score\": 0.5581395348837208, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5, \"recall\": 0.5555555555555556, \"f1-score\": 0.5263157894736842, \"support\": 9.0}, \"request\": {\"precision\": 0.3333333333333333, \"recall\": 0.125, \"f1-score\": 0.18181818181818182, \"support\": 8.0}, \"accuracy\": 0.5258620689655172, \"macro avg\": {\"precision\": 0.33998354776693285, \"recall\": 0.3601469412961349, \"f1-score\": 0.3391474426646145, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.42532783353221343, \"recall\": 0.5258620689655172, \"f1-score\": 0.46188784811262035, \"support\": 116.0}}", + "eval_f1_macro": 30.14643934796573, + "eval_f1_micro": 52.58620689655172, + "eval_f1_weighted": 46.18878481126203, + "eval_loss": 0.5054384469985962, + "eval_runtime": 2.1595, + "eval_samples_per_second": 53.717, + "step": 272 + }, + { + "epoch": 18.0, + "learning_rate": 0.00027, + "loss": 0.5119, + "step": 288 + }, + { + "epoch": 18.0, + "eval_accuracy": 52.58620689655172, + "eval_average_metrics": 46.69862925181969, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.5333333333333333, \"recall\": 0.8888888888888888, \"f1-score\": 0.6666666666666667, \"support\": 9.0}, \"challenge\": {\"precision\": 0.5609756097560976, \"recall\": 0.7419354838709677, \"f1-score\": 0.638888888888889, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.5277777777777778, \"recall\": 0.5135135135135135, \"f1-score\": 0.5205479452054794, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.5555555555555556, \"f1-score\": 0.7142857142857143, \"support\": 9.0}, \"request\": {\"precision\": 0.375, \"recall\": 0.75, \"f1-score\": 0.5, \"support\": 8.0}, \"accuracy\": 0.5258620689655172, \"macro avg\": {\"precision\": 0.37463584010840106, \"recall\": 0.4312366802286157, \"f1-score\": 0.3800486518808437, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.46308639379497246, \"recall\": 0.5258620689655172, \"f1-score\": 0.47840000824767004, \"support\": 116.0}}", + "eval_f1_macro": 33.78210238940833, + "eval_f1_micro": 52.58620689655172, + "eval_f1_weighted": 47.840000824767, + "eval_loss": 0.5165005922317505, + "eval_runtime": 2.2159, + "eval_samples_per_second": 52.349, + "step": 288 + }, + { + "epoch": 19.0, + "learning_rate": 0.000285, + "loss": 0.4492, + "step": 304 + }, + { + "epoch": 19.0, + "eval_accuracy": 60.3448275862069, + "eval_average_metrics": 55.316034256647285, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5384615384615384, \"recall\": 0.5833333333333334, \"f1-score\": 0.5599999999999999, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.5, \"recall\": 0.967741935483871, \"f1-score\": 0.6593406593406593, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.35135135135135137, \"f1-score\": 0.48148148148148157, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.603448275862069, \"macro avg\": {\"precision\": 0.5309514831573654, \"recall\": 0.5242616608544028, \"f1-score\": 0.5012169312169312, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.6142728974879077, \"recall\": 0.603448275862069, \"f1-score\": 0.5602186574600367, \"support\": 116.0}}", + "eval_f1_macro": 44.55261610817167, + "eval_f1_micro": 60.3448275862069, + "eval_f1_weighted": 56.02186574600368, + "eval_loss": 0.4782406687736511, + "eval_runtime": 2.2036, + "eval_samples_per_second": 52.641, + "step": 304 + }, + { + "epoch": 20.0, + "learning_rate": 0.0003, + "loss": 0.3965, + "step": 320 + }, + { + "epoch": 20.0, + "eval_accuracy": 66.37931034482759, + "eval_average_metrics": 61.760117229791256, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.5, \"recall\": 0.2222222222222222, \"f1-score\": 0.30769230769230765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.75, \"recall\": 0.5, \"f1-score\": 0.6, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6486486486486487, \"recall\": 0.7741935483870968, \"f1-score\": 0.7058823529411764, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7058823529411765, \"recall\": 0.6486486486486487, \"f1-score\": 0.676056338028169, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5333333333333333, \"recall\": 0.8888888888888888, \"f1-score\": 0.6666666666666667, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.6637931034482759, \"macro avg\": {\"precision\": 0.5797330418653948, \"recall\": 0.5684802746294682, \"f1-score\": 0.5540610176898495, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.665451912358606, \"recall\": 0.6637931034482759, \"f1-score\": 0.6503197999041211, \"support\": 116.0}}", + "eval_f1_macro": 49.24986823909773, + "eval_f1_micro": 66.37931034482759, + "eval_f1_weighted": 65.03197999041211, + "eval_loss": 0.39811432361602783, + "eval_runtime": 2.0694, + "eval_samples_per_second": 56.054, + "step": 320 + }, + { + "epoch": 21.0, + "learning_rate": 0.00029833333333333334, + "loss": 0.3235, + "step": 336 + }, + { + "epoch": 21.0, + "eval_accuracy": 68.10344827586206, + "eval_average_metrics": 64.5612969778262, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.4444444444444444, \"f1-score\": 0.5333333333333333, \"support\": 9.0}, \"accusation\": {\"precision\": 0.7272727272727273, \"recall\": 0.6666666666666666, \"f1-score\": 0.6956521739130435, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.5490196078431373, \"recall\": 0.9032258064516129, \"f1-score\": 0.6829268292682927, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.85, \"recall\": 0.4594594594594595, \"f1-score\": 0.5964912280701754, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.6810344827586207, \"macro avg\": {\"precision\": 0.650707176810118, \"recall\": 0.6251967693499951, \"f1-score\": 0.6198996519223119, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7246036502375245, \"recall\": 0.6810344827586207, \"f1-score\": 0.6693610007759736, \"support\": 116.0}}", + "eval_f1_macro": 55.10219128198328, + "eval_f1_micro": 68.10344827586206, + "eval_f1_weighted": 66.93610007759737, + "eval_loss": 0.4031699597835541, + "eval_runtime": 2.1347, + "eval_samples_per_second": 54.34, + "step": 336 + }, + { + "epoch": 22.0, + "learning_rate": 0.00029666666666666665, + "loss": 0.2945, + "step": 352 + }, + { + "epoch": 22.0, + "eval_accuracy": 69.82758620689656, + "eval_average_metrics": 66.08601331128551, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.5, \"recall\": 0.5555555555555556, \"f1-score\": 0.5263157894736842, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8333333333333334, \"f1-score\": 0.7407407407407408, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.7096774193548387, \"f1-score\": 0.6875, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7666666666666667, \"recall\": 0.6216216216216216, \"f1-score\": 0.6865671641791045, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.6242424242424243, \"recall\": 0.6403707134553909, \"f1-score\": 0.6215324031911326, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7075757575757575, \"recall\": 0.6982758620689655, \"f1-score\": 0.6944155610324821, \"support\": 116.0}}", + "eval_f1_macro": 55.24732472810068, + "eval_f1_micro": 69.82758620689656, + "eval_f1_weighted": 69.44155610324822, + "eval_loss": 0.3617841899394989, + "eval_runtime": 2.0239, + "eval_samples_per_second": 57.314, + "step": 352 + }, + { + "epoch": 23.0, + "learning_rate": 0.00029499999999999996, + "loss": 0.3049, + "step": 368 + }, + { + "epoch": 23.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 71.50408533829969, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.7142857142857143, \"recall\": 0.5555555555555556, \"f1-score\": 0.6250000000000001, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8333333333333334, \"f1-score\": 0.7407407407407408, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.7741935483870968, \"f1-score\": 0.7164179104477612, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7692307692307693, \"recall\": 0.5405405405405406, \"f1-score\": 0.6349206349206349, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7720872877122877, \"recall\": 0.8084389833381769, \"f1-score\": 0.7817480060017373, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7320725251759735, \"recall\": 0.7241379310344828, \"f1-score\": 0.717000435017033, \"support\": 116.0}}", + "eval_f1_macro": 69.4887116445989, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 71.7000435017033, + "eval_loss": 0.3993362486362457, + "eval_runtime": 2.1528, + "eval_samples_per_second": 53.884, + "step": 368 + }, + { + "epoch": 24.0, + "learning_rate": 0.00029333333333333327, + "loss": 0.2268, + "step": 384 + }, + { + "epoch": 24.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 72.61508757752452, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.7142857142857143, \"recall\": 0.5555555555555556, \"f1-score\": 0.6250000000000001, \"support\": 9.0}, \"accusation\": {\"precision\": 0.9, \"recall\": 0.75, \"f1-score\": 0.8181818181818182, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6, \"recall\": 0.8709677419354839, \"f1-score\": 0.7105263157894737, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.8, \"recall\": 0.5405405405405406, \"f1-score\": 0.6451612903225806, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.875, \"f1-score\": 0.9333333333333333, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.8212301587301587, \"recall\": 0.7962302019761698, \"f1-score\": 0.7978745510526071, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7623152709359605, \"recall\": 0.7327586206896551, \"f1-score\": 0.7298644385637975, \"support\": 116.0}}", + "eval_f1_macro": 70.9221823157873, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 72.98644385637975, + "eval_loss": 0.3922426402568817, + "eval_runtime": 1.9795, + "eval_samples_per_second": 58.601, + "step": 384 + }, + { + "epoch": 25.0, + "learning_rate": 0.00029166666666666664, + "loss": 0.2401, + "step": 400 + }, + { + "epoch": 25.0, + "eval_accuracy": 75.86206896551724, + "eval_average_metrics": 74.81572695778885, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.7142857142857143, \"recall\": 0.5555555555555556, \"f1-score\": 0.6250000000000001, \"support\": 9.0}, \"accusation\": {\"precision\": 0.9, \"recall\": 0.75, \"f1-score\": 0.8181818181818182, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6428571428571429, \"recall\": 0.8709677419354839, \"f1-score\": 0.7397260273972603, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.8214285714285714, \"recall\": 0.6216216216216216, \"f1-score\": 0.7076923076923075, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.875, \"f1-score\": 0.9333333333333333, \"support\": 8.0}, \"accuracy\": 0.7586206896551724, \"macro avg\": {\"precision\": 0.8257305194805196, \"recall\": 0.8063653371113049, \"f1-score\": 0.8082548437203267, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7784090909090909, \"recall\": 0.7586206896551724, \"f1-score\": 0.7569389490275861, \"support\": 116.0}}", + "eval_f1_macro": 71.84487499736237, + "eval_f1_micro": 75.86206896551724, + "eval_f1_weighted": 75.69389490275861, + "eval_loss": 0.38488852977752686, + "eval_runtime": 2.103, + "eval_samples_per_second": 55.158, + "step": 400 + }, + { + "epoch": 26.0, + "learning_rate": 0.00029, + "loss": 0.1774, + "step": 416 + }, + { + "epoch": 26.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 69.79597725103807, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.625, \"recall\": 0.5555555555555556, \"f1-score\": 0.5882352941176471, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8333333333333334, \"f1-score\": 0.7407407407407408, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.5333333333333333, \"recall\": 0.8888888888888888, \"f1-score\": 0.6666666666666667, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.7419354838709677, \"f1-score\": 0.7419354838709677, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7307692307692307, \"recall\": 0.5135135135135135, \"f1-score\": 0.6031746031746031, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6923076923076923, \"recall\": 1.0, \"f1-score\": 0.8181818181818181, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.875, \"f1-score\": 0.9333333333333333, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7487515508684863, \"recall\": 0.8010283468952824, \"f1-score\": 0.7615334925107221, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7215019893899203, \"recall\": 0.7068965517241379, \"f1-score\": 0.701127326583716, \"support\": 116.0}}", + "eval_f1_macro": 67.69186600095307, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.1127326583716, + "eval_loss": 0.4694823622703552, + "eval_runtime": 2.2038, + "eval_samples_per_second": 52.636, + "step": 416 + }, + { + "epoch": 27.0, + "learning_rate": 0.0002883333333333333, + "loss": 0.1601, + "step": 432 + }, + { + "epoch": 27.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 71.6286036100653, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.7142857142857143, \"recall\": 0.5555555555555556, \"f1-score\": 0.6250000000000001, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.7741935483870968, \"f1-score\": 0.7164179104477612, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7777777777777778, \"recall\": 0.5675675675675675, \"f1-score\": 0.65625, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.875, \"f1-score\": 0.9333333333333333, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7842797725150666, \"recall\": 0.7979284728276663, \"f1-score\": 0.7828452099190978, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7385973616196739, \"recall\": 0.7241379310344828, \"f1-score\": 0.7210058735166711, \"support\": 116.0}}", + "eval_f1_macro": 69.58624088169758, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.10058735166709, + "eval_loss": 0.4516548216342926, + "eval_runtime": 2.1426, + "eval_samples_per_second": 54.14, + "step": 432 + }, + { + "epoch": 28.0, + "learning_rate": 0.0002866666666666667, + "loss": 0.1426, + "step": 448 + }, + { + "epoch": 28.0, + "eval_accuracy": 75.86206896551724, + "eval_average_metrics": 75.02176780786866, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.7741935483870968, \"f1-score\": 0.7164179104477612, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.8, \"recall\": 0.6486486486486487, \"f1-score\": 0.7164179104477612, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.875, \"f1-score\": 0.9333333333333333, \"support\": 8.0}, \"accuracy\": 0.7586206896551724, \"macro avg\": {\"precision\": 0.8166666666666667, \"recall\": 0.8254247190739126, \"f1-score\": 0.8164155862659283, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7682471264367817, \"recall\": 0.7586206896551724, \"f1-score\": 0.75792658965691, \"support\": 116.0}}", + "eval_f1_macro": 72.57027433474917, + "eval_f1_micro": 75.86206896551724, + "eval_f1_weighted": 75.79265896569099, + "eval_loss": 0.4462680220603943, + "eval_runtime": 2.0575, + "eval_samples_per_second": 56.379, + "step": 448 + }, + { + "epoch": 29.0, + "learning_rate": 0.000285, + "loss": 0.1153, + "step": 464 + }, + { + "epoch": 29.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 69.98010595460308, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.7142857142857143, \"recall\": 0.5555555555555556, \"f1-score\": 0.6250000000000001, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6428571428571429, \"recall\": 0.75, \"f1-score\": 0.6923076923076924, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.5, \"recall\": 0.8888888888888888, \"f1-score\": 0.64, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6410256410256411, \"recall\": 0.8064516129032258, \"f1-score\": 0.7142857142857142, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.8260869565217391, \"recall\": 0.5135135135135135, \"f1-score\": 0.6333333333333333, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.875, \"f1-score\": 0.9333333333333333, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7766430429473907, \"recall\": 0.784787307468759, \"f1-score\": 0.7658936202686203, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7420678397065203, \"recall\": 0.7068965517241379, \"f1-score\": 0.704616805608185, \"support\": 116.0}}", + "eval_f1_macro": 68.07943291276625, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.4616805608185, + "eval_loss": 0.5256942510604858, + "eval_runtime": 2.1565, + "eval_samples_per_second": 53.791, + "step": 464 + }, + { + "epoch": 30.0, + "learning_rate": 0.0002833333333333333, + "loss": 0.0824, + "step": 480 + }, + { + "epoch": 30.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 70.1937005308774, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.6451612903225806, \"f1-score\": 0.6557377049180327, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7777777777777778, \"recall\": 0.5675675675675675, \"f1-score\": 0.65625, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.875, \"f1-score\": 0.9333333333333333, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7606060606060606, \"recall\": 0.8095772183473796, \"f1-score\": 0.775683926773902, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7230581678857542, \"recall\": 0.7068965517241379, \"f1-score\": 0.7044580939877959, \"support\": 116.0}}", + "eval_f1_macro": 68.9496823799024, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.4458093987796, + "eval_loss": 0.5310544371604919, + "eval_runtime": 2.0193, + "eval_samples_per_second": 57.447, + "step": 480 + }, + { + "epoch": 31.0, + "learning_rate": 0.00028166666666666666, + "loss": 0.0898, + "step": 496 + }, + { + "epoch": 31.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 71.05355939177225, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.5714285714285714, \"recall\": 0.8888888888888888, \"f1-score\": 0.6956521739130435, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.5675675675675675, \"f1-score\": 0.6461538461538462, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.875, \"f1-score\": 0.9333333333333333, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7776785714285714, \"recall\": 0.8072250678097452, \"f1-score\": 0.7837074960829841, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7310806650246304, \"recall\": 0.7155172413793104, \"f1-score\": 0.7144790075051725, \"support\": 116.0}}", + "eval_f1_macro": 69.6628885407097, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.44790075051725, + "eval_loss": 0.5583775043487549, + "eval_runtime": 2.0569, + "eval_samples_per_second": 56.394, + "step": 496 + }, + { + "epoch": 32.0, + "learning_rate": 0.00028, + "loss": 0.0709, + "step": 512 + }, + { + "epoch": 32.0, + "eval_accuracy": 78.44827586206897, + "eval_average_metrics": 77.31764307402186, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.9, \"recall\": 0.75, \"f1-score\": 0.8181818181818182, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6923076923076923, \"recall\": 0.8709677419354839, \"f1-score\": 0.7714285714285716, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.8275862068965517, \"recall\": 0.6486486486486487, \"f1-score\": 0.7272727272727273, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.875, \"f1-score\": 0.9333333333333333, \"support\": 8.0}, \"accuracy\": 0.7844827586206896, \"macro avg\": {\"precision\": 0.8420700707338638, \"recall\": 0.8375214932674611, \"f1-score\": 0.8331714982643775, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7994157596268179, \"recall\": 0.7844827586206896, \"f1-score\": 0.7831433183733814, \"support\": 116.0}}", + "eval_f1_macro": 74.05968873461136, + "eval_f1_micro": 78.44827586206897, + "eval_f1_weighted": 78.31433183733813, + "eval_loss": 0.5467008948326111, + "eval_runtime": 1.9594, + "eval_samples_per_second": 59.203, + "step": 512 + }, + { + "epoch": 33.0, + "learning_rate": 0.00027833333333333334, + "loss": 0.084, + "step": 528 + }, + { + "epoch": 33.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 72.73042654857636, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6428571428571429, \"recall\": 0.75, \"f1-score\": 0.6923076923076924, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6216216216216216, \"recall\": 0.7419354838709677, \"f1-score\": 0.676470588235294, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.8148148148148148, \"recall\": 0.5945945945945946, \"f1-score\": 0.6875, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.875, \"f1-score\": 0.9333333333333333, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7994950307450308, \"recall\": 0.8146357042526398, \"f1-score\": 0.8005958937218612, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7498508976957252, \"recall\": 0.7327586206896551, \"f1-score\": 0.7320590261443122, \"support\": 116.0}}", + "eval_f1_macro": 71.16407944194322, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.20590261443122, + "eval_loss": 0.5764071345329285, + "eval_runtime": 2.116, + "eval_samples_per_second": 54.82, + "step": 528 + }, + { + "epoch": 34.0, + "learning_rate": 0.00027666666666666665, + "loss": 0.0838, + "step": 544 + }, + { + "epoch": 34.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 70.64634709659882, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6923076923076923, \"recall\": 0.75, \"f1-score\": 0.7199999999999999, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6486486486486487, \"recall\": 0.7741935483870968, \"f1-score\": 0.7058823529411764, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7777777777777778, \"recall\": 0.5675675675675675, \"f1-score\": 0.65625, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7754478254478254, \"recall\": 0.7840395839387775, \"f1-score\": 0.7681058539306411, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7323897806656428, \"recall\": 0.7155172413793104, \"f1-score\": 0.7120586420558731, \"support\": 116.0}}", + "eval_f1_macro": 68.27607590494587, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.2058642055873, + "eval_loss": 0.6548714637756348, + "eval_runtime": 2.138, + "eval_samples_per_second": 54.256, + "step": 544 + }, + { + "epoch": 35.0, + "learning_rate": 0.00027499999999999996, + "loss": 0.0924, + "step": 560 + }, + { + "epoch": 35.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 72.46959003253342, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6428571428571429, \"recall\": 0.75, \"f1-score\": 0.6923076923076924, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6451612903225806, \"recall\": 0.6451612903225806, \"f1-score\": 0.6451612903225806, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.8333333333333334, \"recall\": 0.6756756756756757, \"f1-score\": 0.746268656716418, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 9.0}, \"request\": {\"precision\": 0.7777777777777778, \"recall\": 0.875, \"f1-score\": 0.823529411764706, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7707245263696876, \"recall\": 0.8126740651942264, \"f1-score\": 0.7875938450480724, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7428434592227696, \"recall\": 0.7327586206896551, \"f1-score\": 0.7331829421015178, \"support\": 116.0}}", + "eval_f1_macro": 70.00834178205089, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.31829421015178, + "eval_loss": 0.5695762038230896, + "eval_runtime": 2.0764, + "eval_samples_per_second": 55.866, + "step": 560 + }, + { + "epoch": 36.0, + "learning_rate": 0.00027333333333333333, + "loss": 0.0452, + "step": 576 + }, + { + "epoch": 36.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 70.22480821008993, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.7142857142857143, \"recall\": 0.5555555555555556, \"f1-score\": 0.6250000000000001, \"support\": 9.0}, \"accusation\": {\"precision\": 0.43478260869565216, \"recall\": 0.8333333333333334, \"f1-score\": 0.5714285714285714, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7037037037037037, \"recall\": 0.6129032258064516, \"f1-score\": 0.6551724137931035, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.8, \"recall\": 0.6486486486486487, \"f1-score\": 0.7164179104477612, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.875, \"f1-score\": 0.9333333333333333, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7736419578810884, \"recall\": 0.7879023176402209, \"f1-score\": 0.7685900812569252, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7404142265086794, \"recall\": 0.7068965517241379, \"f1-score\": 0.7120080416158326, \"support\": 116.0}}", + "eval_f1_macro": 68.3191183339489, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 71.20080416158326, + "eval_loss": 0.6615910530090332, + "eval_runtime": 2.0721, + "eval_samples_per_second": 55.982, + "step": 576 + }, + { + "epoch": 37.0, + "learning_rate": 0.00027166666666666664, + "loss": 0.0662, + "step": 592 + }, + { + "epoch": 37.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 70.7853828197613, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.7142857142857143, \"recall\": 0.5555555555555556, \"f1-score\": 0.6250000000000001, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5, \"recall\": 0.8333333333333334, \"f1-score\": 0.625, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6774193548387096, \"recall\": 0.6774193548387096, \"f1-score\": 0.6774193548387096, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.8214285714285714, \"recall\": 0.6216216216216216, \"f1-score\": 0.7076923076923075, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.774747765675185, \"recall\": 0.7908523442797636, \"f1-score\": 0.7692699101973295, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7429746977160769, \"recall\": 0.7155172413793104, \"f1-score\": 0.7165853543008714, \"support\": 116.0}}", + "eval_f1_macro": 68.37954757309596, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.65853543008714, + "eval_loss": 0.686450719833374, + "eval_runtime": 1.9539, + "eval_samples_per_second": 59.368, + "step": 592 + }, + { + "epoch": 38.0, + "learning_rate": 0.00027, + "loss": 0.0408, + "step": 608 + }, + { + "epoch": 38.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 71.66157684555884, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6571428571428571, \"recall\": 0.7419354838709677, \"f1-score\": 0.6969696969696969, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.8148148148148148, \"recall\": 0.5945945945945946, \"f1-score\": 0.6875, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7836632696007696, \"recall\": 0.7990107042526398, \"f1-score\": 0.7815321014769544, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7446828714070093, \"recall\": 0.7241379310344828, \"f1-score\": 0.7234920104405398, \"support\": 116.0}}", + "eval_f1_macro": 69.46952013128484, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.34920104405398, + "eval_loss": 0.7136876583099365, + "eval_runtime": 2.1177, + "eval_samples_per_second": 54.777, + "step": 608 + }, + { + "epoch": 39.0, + "learning_rate": 0.0002683333333333333, + "loss": 0.0315, + "step": 624 + }, + { + "epoch": 39.0, + "eval_accuracy": 68.10344827586206, + "eval_average_metrics": 67.42979960880837, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.4444444444444444, \"f1-score\": 0.5333333333333333, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6052631578947368, \"recall\": 0.7419354838709677, \"f1-score\": 0.6666666666666666, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7777777777777778, \"recall\": 0.5675675675675675, \"f1-score\": 0.65625, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"accuracy\": 0.6810344827586207, \"macro avg\": {\"precision\": 0.7563779239766082, \"recall\": 0.7383406592075947, \"f1-score\": 0.7353737797113752, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7091096995361968, \"recall\": 0.6810344827586207, \"f1-score\": 0.6814574368694261, \"support\": 116.0}}", + "eval_f1_macro": 65.36655819656669, + "eval_f1_micro": 68.10344827586206, + "eval_f1_weighted": 68.14574368694261, + "eval_loss": 0.8272783756256104, + "eval_runtime": 2.1248, + "eval_samples_per_second": 54.594, + "step": 624 + }, + { + "epoch": 40.0, + "learning_rate": 0.0002666666666666666, + "loss": 0.0537, + "step": 640 + }, + { + "epoch": 40.0, + "eval_accuracy": 69.82758620689656, + "eval_average_metrics": 69.24541954934139, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6363636363636364, \"recall\": 0.6774193548387096, \"f1-score\": 0.65625, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.8076923076923077, \"recall\": 0.5675675675675675, \"f1-score\": 0.6666666666666666, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7550152972027973, \"recall\": 0.787567809745229, \"f1-score\": 0.7602210426634769, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7212412587412587, \"recall\": 0.6982758620689655, \"f1-score\": 0.6975130199126343, \"support\": 116.0}}", + "eval_f1_macro": 67.57520379230904, + "eval_f1_micro": 69.82758620689656, + "eval_f1_weighted": 69.75130199126343, + "eval_loss": 0.724757969379425, + "eval_runtime": 2.3811, + "eval_samples_per_second": 48.716, + "step": 640 + }, + { + "epoch": 41.0, + "learning_rate": 0.000265, + "loss": 0.0215, + "step": 656 + }, + { + "epoch": 41.0, + "eval_accuracy": 69.82758620689656, + "eval_average_metrics": 69.33402692747292, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6388888888888888, \"recall\": 0.7419354838709677, \"f1-score\": 0.6865671641791045, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7407407407407407, \"recall\": 0.5405405405405406, \"f1-score\": 0.625, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7615477693602694, \"recall\": 0.7801011697181053, \"f1-score\": 0.7655244436082891, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7106916869847903, \"recall\": 0.6982758620689655, \"f1-score\": 0.6963431808647288, \"support\": 116.0}}", + "eval_f1_macro": 68.0466172096257, + "eval_f1_micro": 69.82758620689656, + "eval_f1_weighted": 69.63431808647287, + "eval_loss": 0.7391438484191895, + "eval_runtime": 2.157, + "eval_samples_per_second": 53.778, + "step": 656 + }, + { + "epoch": 42.0, + "learning_rate": 0.0002633333333333333, + "loss": 0.0267, + "step": 672 + }, + { + "epoch": 42.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 71.60599983270731, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6923076923076923, \"recall\": 0.75, \"f1-score\": 0.7199999999999999, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6388888888888888, \"recall\": 0.7419354838709677, \"f1-score\": 0.6865671641791045, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7857142857142857, \"recall\": 0.5945945945945946, \"f1-score\": 0.676923076923077, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7761127761127761, \"recall\": 0.7990107042526398, \"f1-score\": 0.781409669493515, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7340987556504798, \"recall\": 0.7241379310344828, \"f1-score\": 0.7213777583562028, \"support\": 116.0}}", + "eval_f1_macro": 69.45863728831245, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.13777583562027, + "eval_loss": 0.730360209941864, + "eval_runtime": 2.2556, + "eval_samples_per_second": 51.428, + "step": 672 + }, + { + "epoch": 43.0, + "learning_rate": 0.00026166666666666667, + "loss": 0.0238, + "step": 688 + }, + { + "epoch": 43.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 71.03986860621411, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.43478260869565216, \"recall\": 0.8333333333333334, \"f1-score\": 0.5714285714285714, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.5806451612903226, \"f1-score\": 0.6545454545454547, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7878787878787878, \"recall\": 0.7027027027027027, \"f1-score\": 0.7428571428571429, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.875, \"f1-score\": 0.9333333333333333, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7809892402283706, \"recall\": 0.7906268163324615, \"f1-score\": 0.7746169679993209, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7508291308891009, \"recall\": 0.7155172413793104, \"f1-score\": 0.7220118454905474, \"support\": 116.0}}", + "eval_f1_macro": 68.85484159993965, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 72.20118454905474, + "eval_loss": 0.7982239127159119, + "eval_runtime": 2.2527, + "eval_samples_per_second": 51.494, + "step": 688 + }, + { + "epoch": 44.0, + "learning_rate": 0.00026, + "loss": 0.0209, + "step": 704 + }, + { + "epoch": 44.0, + "eval_accuracy": 67.24137931034483, + "eval_average_metrics": 66.37757277211017, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.625, \"recall\": 0.8333333333333334, \"f1-score\": 0.7142857142857143, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6451612903225806, \"recall\": 0.6451612903225806, \"f1-score\": 0.6451612903225806, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7333333333333333, \"recall\": 0.5945945945945946, \"f1-score\": 0.6567164179104478, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.6724137931034483, \"macro avg\": {\"precision\": 0.7129368279569892, \"recall\": 0.7400389300590913, \"f1-score\": 0.7194389741807576, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.6818965517241379, \"recall\": 0.6724137931034483, \"f1-score\": 0.6707740142946146, \"support\": 116.0}}", + "eval_f1_macro": 63.95013103828957, + "eval_f1_micro": 67.24137931034483, + "eval_f1_weighted": 67.07740142946146, + "eval_loss": 0.788184404373169, + "eval_runtime": 2.2001, + "eval_samples_per_second": 52.725, + "step": 704 + }, + { + "epoch": 45.0, + "learning_rate": 0.00025833333333333334, + "loss": 0.0349, + "step": 720 + }, + { + "epoch": 45.0, + "eval_accuracy": 69.82758620689656, + "eval_average_metrics": 68.73739006091948, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6428571428571429, \"recall\": 0.75, \"f1-score\": 0.6923076923076924, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6571428571428571, \"recall\": 0.7419354838709677, \"f1-score\": 0.6969696969696969, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7586206896551724, \"recall\": 0.5945945945945946, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 8.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7613339964633068, \"recall\": 0.7538718153637509, \"f1-score\": 0.7397088205911735, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7193374253570448, \"recall\": 0.6982758620689655, \"f1-score\": 0.6954249266622493, \"support\": 116.0}}", + "eval_f1_macro": 65.75189516365988, + "eval_f1_micro": 69.82758620689656, + "eval_f1_weighted": 69.54249266622493, + "eval_loss": 0.9080767035484314, + "eval_runtime": 2.3303, + "eval_samples_per_second": 49.778, + "step": 720 + }, + { + "epoch": 46.0, + "learning_rate": 0.00025666666666666665, + "loss": 0.0085, + "step": 736 + }, + { + "epoch": 46.0, + "eval_accuracy": 68.96551724137932, + "eval_average_metrics": 68.39529379047518, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.7142857142857143, \"recall\": 0.5555555555555556, \"f1-score\": 0.6250000000000001, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6470588235294118, \"recall\": 0.7096774193548387, \"f1-score\": 0.6769230769230768, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7333333333333333, \"recall\": 0.5945945945945946, \"f1-score\": 0.6567164179104478, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"accuracy\": 0.6896551724137931, \"macro avg\": {\"precision\": 0.761999207577768, \"recall\": 0.7515756684103458, \"f1-score\": 0.7481724665655658, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7097975669396616, \"recall\": 0.6896551724137931, \"f1-score\": 0.6914592142886957, \"support\": 116.0}}", + "eval_f1_macro": 66.50421925027251, + "eval_f1_micro": 68.96551724137932, + "eval_f1_weighted": 69.14592142886957, + "eval_loss": 0.9103646278381348, + "eval_runtime": 2.3131, + "eval_samples_per_second": 50.149, + "step": 736 + }, + { + "epoch": 47.0, + "learning_rate": 0.00025499999999999996, + "loss": 0.0252, + "step": 752 + }, + { + "epoch": 47.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 69.88211421894832, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6428571428571429, \"recall\": 0.75, \"f1-score\": 0.6923076923076924, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6486486486486487, \"recall\": 0.7741935483870968, \"f1-score\": 0.7058823529411764, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.5675675675675675, \"f1-score\": 0.6461538461538462, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7606265356265356, \"recall\": 0.7718868061609997, \"f1-score\": 0.7615414384377233, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7164380121276674, \"recall\": 0.7068965517241379, \"f1-score\": 0.7045657422539028, \"support\": 116.0}}", + "eval_f1_macro": 67.69257230557541, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.45657422539028, + "eval_loss": 0.8836216330528259, + "eval_runtime": 2.3628, + "eval_samples_per_second": 49.094, + "step": 752 + }, + { + "epoch": 48.0, + "learning_rate": 0.00025333333333333333, + "loss": 0.0068, + "step": 768 + }, + { + "epoch": 48.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 70.60514862150394, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6774193548387096, \"recall\": 0.6774193548387096, \"f1-score\": 0.6774193548387096, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.7027027027027027, \"f1-score\": 0.7222222222222223, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7785762288786482, \"recall\": 0.7610569238593432, \"f1-score\": 0.7609549816618129, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7289716748768472, \"recall\": 0.7155172413793104, \"f1-score\": 0.7167670339577035, \"support\": 116.0}}", + "eval_f1_macro": 67.64044281438335, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.67670339577035, + "eval_loss": 0.7966245412826538, + "eval_runtime": 2.368, + "eval_samples_per_second": 48.987, + "step": 768 + }, + { + "epoch": 49.0, + "learning_rate": 0.00025166666666666664, + "loss": 0.0115, + "step": 784 + }, + { + "epoch": 49.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 71.30009730584443, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5, \"recall\": 0.9166666666666666, \"f1-score\": 0.6470588235294118, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7586206896551724, \"recall\": 0.7096774193548387, \"f1-score\": 0.7333333333333333, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.8461538461538461, \"recall\": 0.5945945945945946, \"f1-score\": 0.6984126984126985, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 1.0, \"f1-score\": 0.8571428571428571, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7630968169761274, \"recall\": 0.800186779521457, \"f1-score\": 0.7647525676937441, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7585482789109425, \"recall\": 0.7241379310344828, \"f1-score\": 0.7239479699925947, \"support\": 116.0}}", + "eval_f1_macro": 67.9780060172217, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.39479699925948, + "eval_loss": 0.9425663352012634, + "eval_runtime": 2.4101, + "eval_samples_per_second": 48.13, + "step": 784 + }, + { + "epoch": 50.0, + "learning_rate": 0.00025, + "loss": 0.0116, + "step": 800 + }, + { + "epoch": 50.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 70.36753700765975, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7741935483870968, \"recall\": 0.6486486486486487, \"f1-score\": 0.7058823529411764, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"request\": {\"precision\": 0.7142857142857143, \"recall\": 0.625, \"f1-score\": 0.6666666666666666, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.74224677652097, \"recall\": 0.778605722658142, \"f1-score\": 0.7523584009875137, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7290652432476681, \"recall\": 0.7155172413793104, \"f1-score\": 0.7149039744477569, \"support\": 116.0}}", + "eval_f1_macro": 66.87630231000122, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.49039744477568, + "eval_loss": 0.9411255121231079, + "eval_runtime": 2.3319, + "eval_samples_per_second": 49.744, + "step": 800 + }, + { + "epoch": 51.0, + "learning_rate": 0.0002483333333333333, + "loss": 0.0087, + "step": 816 + }, + { + "epoch": 51.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 69.62433585996804, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6571428571428571, \"recall\": 0.7419354838709677, \"f1-score\": 0.6969696969696969, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7931034482758621, \"recall\": 0.6216216216216216, \"f1-score\": 0.696969696969697, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7493363437328955, \"recall\": 0.7589863048532404, \"f1-score\": 0.7476551226551227, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7205424381405356, \"recall\": 0.7068965517241379, \"f1-score\": 0.706597999701448, \"support\": 116.0}}", + "eval_f1_macro": 66.45823312489979, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.6597999701448, + "eval_loss": 0.9394950866699219, + "eval_runtime": 2.3963, + "eval_samples_per_second": 48.408, + "step": 816 + }, + { + "epoch": 52.0, + "learning_rate": 0.0002466666666666666, + "loss": 0.0126, + "step": 832 + }, + { + "epoch": 52.0, + "eval_accuracy": 69.82758620689656, + "eval_average_metrics": 69.02533360824437, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6428571428571429, \"recall\": 0.75, \"f1-score\": 0.6923076923076924, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6216216216216216, \"recall\": 0.7419354838709677, \"f1-score\": 0.676470588235294, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7419354838709677, \"recall\": 0.6216216216216216, \"f1-score\": 0.676470588235294, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7626389022558377, \"recall\": 0.7450974159643514, \"f1-score\": 0.7486182934712347, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7108198989803551, \"recall\": 0.6982758620689655, \"f1-score\": 0.6990231371063014, \"support\": 116.0}}", + "eval_f1_macro": 66.54384830855419, + "eval_f1_micro": 69.82758620689656, + "eval_f1_weighted": 69.90231371063015, + "eval_loss": 0.9984409809112549, + "eval_runtime": 2.2343, + "eval_samples_per_second": 51.918, + "step": 832 + }, + { + "epoch": 53.0, + "learning_rate": 0.000245, + "loss": 0.0192, + "step": 848 + }, + { + "epoch": 53.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 69.02199924752503, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5238095238095238, \"recall\": 0.9166666666666666, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.6774193548387096, \"f1-score\": 0.7, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7857142857142857, \"recall\": 0.5945945945945946, \"f1-score\": 0.676923076923077, \"support\": 37.0}, \"question\": {\"precision\": 0.5, \"recall\": 1.0, \"f1-score\": 0.6666666666666666, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.6971460292580982, \"recall\": 0.7840017436791631, \"f1-score\": 0.7197805032944352, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7313778819278225, \"recall\": 0.7068965517241379, \"f1-score\": 0.7072819746354496, \"support\": 116.0}}", + "eval_f1_macro": 63.98048918172757, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.72819746354496, + "eval_loss": 1.050161600112915, + "eval_runtime": 2.3204, + "eval_samples_per_second": 49.991, + "step": 848 + }, + { + "epoch": 54.0, + "learning_rate": 0.0002433333333333333, + "loss": 0.0124, + "step": 864 + }, + { + "epoch": 54.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 72.27556400444958, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6923076923076923, \"recall\": 0.75, \"f1-score\": 0.7199999999999999, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.7096774193548387, \"f1-score\": 0.7096774193548389, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.7027027027027027, \"f1-score\": 0.7123287671232876, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7841271072117846, \"recall\": 0.7807141819238593, \"f1-score\": 0.7791653391360963, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7371729085091153, \"recall\": 0.7327586206896551, \"f1-score\": 0.7329139062332538, \"support\": 116.0}}", + "eval_f1_macro": 69.2591412565419, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.29139062332538, + "eval_loss": 0.9056476950645447, + "eval_runtime": 2.3786, + "eval_samples_per_second": 48.768, + "step": 864 + }, + { + "epoch": 55.0, + "learning_rate": 0.00024166666666666664, + "loss": 0.0034, + "step": 880 + }, + { + "epoch": 55.0, + "eval_accuracy": 68.10344827586206, + "eval_average_metrics": 67.43628999917863, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.5, \"recall\": 0.6666666666666666, \"f1-score\": 0.5714285714285715, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.6451612903225806, \"f1-score\": 0.6557377049180327, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.5675675675675675, \"f1-score\": 0.6461538461538462, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.6810344827586207, \"macro avg\": {\"precision\": 0.7209686147186147, \"recall\": 0.769646662791824, \"f1-score\": 0.7377364438839848, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.6941315868040006, \"recall\": 0.6810344827586207, \"f1-score\": 0.6796169065530286, \"support\": 116.0}}", + "eval_f1_macro": 65.57657278968753, + "eval_f1_micro": 68.10344827586206, + "eval_f1_weighted": 67.96169065530286, + "eval_loss": 1.0662262439727783, + "eval_runtime": 2.2695, + "eval_samples_per_second": 51.112, + "step": 880 + }, + { + "epoch": 56.0, + "learning_rate": 0.00023999999999999998, + "loss": 0.0158, + "step": 896 + }, + { + "epoch": 56.0, + "eval_accuracy": 68.96551724137932, + "eval_average_metrics": 68.23178854815569, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.46153846153846156, \"recall\": 0.6666666666666666, \"f1-score\": 0.5454545454545455, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5789473684210527, \"recall\": 0.9166666666666666, \"f1-score\": 0.7096774193548387, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.72, \"recall\": 0.5806451612903226, \"f1-score\": 0.6428571428571428, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7777777777777778, \"recall\": 0.5675675675675675, \"f1-score\": 0.65625, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 1.0, \"f1-score\": 0.8571428571428571, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.6896551724137931, \"macro avg\": {\"precision\": 0.726509141443352, \"recall\": 0.796304368885014, \"f1-score\": 0.7466608408392683, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7138459931880984, \"recall\": 0.6896551724137931, \"f1-score\": 0.6862626719081808, \"support\": 116.0}}", + "eval_f1_macro": 66.36985251904606, + "eval_f1_micro": 68.96551724137932, + "eval_f1_weighted": 68.62626719081808, + "eval_loss": 1.1591931581497192, + "eval_runtime": 2.2123, + "eval_samples_per_second": 52.434, + "step": 896 + }, + { + "epoch": 57.0, + "learning_rate": 0.0002383333333333333, + "loss": 0.0376, + "step": 912 + }, + { + "epoch": 57.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 69.91520968255182, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6428571428571429, \"recall\": 0.75, \"f1-score\": 0.6923076923076924, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.7096774193548387, \"f1-score\": 0.6875, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7333333333333333, \"recall\": 0.5945945945945946, \"f1-score\": 0.6567164179104478, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7554924242424242, \"recall\": 0.7810895572992347, \"f1-score\": 0.7630389031330098, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7126455441110613, \"recall\": 0.7068965517241379, \"f1-score\": 0.7045584810688993, \"support\": 116.0}}", + "eval_f1_macro": 67.82568027848977, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.45584810688995, + "eval_loss": 0.9944426417350769, + "eval_runtime": 2.3478, + "eval_samples_per_second": 49.409, + "step": 912 + }, + { + "epoch": 58.0, + "learning_rate": 0.00023666666666666663, + "loss": 0.0072, + "step": 928 + }, + { + "epoch": 58.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 70.84391114732561, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6764705882352942, \"recall\": 0.7419354838709677, \"f1-score\": 0.7076923076923077, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7586206896551724, \"recall\": 0.5945945945945946, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.768480475392874, \"recall\": 0.786857926474862, \"f1-score\": 0.7733734198807728, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7264381593086916, \"recall\": 0.7155172413793104, \"f1-score\": 0.7152789232403838, \"support\": 116.0}}", + "eval_f1_macro": 68.74430398940203, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.52789232403838, + "eval_loss": 0.9998703002929688, + "eval_runtime": 2.4048, + "eval_samples_per_second": 48.237, + "step": 928 + }, + { + "epoch": 59.0, + "learning_rate": 0.00023499999999999997, + "loss": 0.0045, + "step": 944 + }, + { + "epoch": 59.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 71.44905544977263, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.7096774193548387, \"f1-score\": 0.7096774193548389, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7741935483870968, \"recall\": 0.6486486486486487, \"f1-score\": 0.7058823529411764, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7650225614439323, \"recall\": 0.7878463140559915, \"f1-score\": 0.7710386616446239, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.734502621960909, \"recall\": 0.7241379310344828, \"f1-score\": 0.7243186566822737, \"support\": 116.0}}", + "eval_f1_macro": 68.53676992396657, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.43186566822737, + "eval_loss": 0.9982974529266357, + "eval_runtime": 2.2032, + "eval_samples_per_second": 52.652, + "step": 944 + }, + { + "epoch": 60.0, + "learning_rate": 0.0002333333333333333, + "loss": 0.0095, + "step": 960 + }, + { + "epoch": 60.0, + "eval_accuracy": 69.82758620689656, + "eval_average_metrics": 69.03488330679734, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.5, \"recall\": 0.6666666666666666, \"f1-score\": 0.5714285714285715, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6896551724137931, \"recall\": 0.6451612903225806, \"f1-score\": 0.6666666666666667, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7666666666666667, \"recall\": 0.6216216216216216, \"f1-score\": 0.6865671641791045, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7350164203612479, \"recall\": 0.7764034195485807, \"f1-score\": 0.7494173867504583, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7112337919710096, \"recall\": 0.6982758620689655, \"f1-score\": 0.6986948199113326, \"support\": 116.0}}", + "eval_f1_macro": 66.61487882226295, + "eval_f1_micro": 69.82758620689656, + "eval_f1_weighted": 69.86948199113327, + "eval_loss": 0.9828293919563293, + "eval_runtime": 2.1348, + "eval_samples_per_second": 54.338, + "step": 960 + }, + { + "epoch": 61.0, + "learning_rate": 0.00023166666666666667, + "loss": 0.0037, + "step": 976 + }, + { + "epoch": 61.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 69.7600508509814, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6486486486486487, \"f1-score\": 0.6956521739130435, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7619791666666667, \"recall\": 0.7681890559914752, \"f1-score\": 0.7532445691144488, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7237068965517242, \"recall\": 0.7068965517241379, \"f1-score\": 0.7070582024892478, \"support\": 116.0}}", + "eval_f1_macro": 66.95507281017323, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.70582024892478, + "eval_loss": 1.0811266899108887, + "eval_runtime": 2.041, + "eval_samples_per_second": 56.836, + "step": 976 + }, + { + "epoch": 62.0, + "learning_rate": 0.00023, + "loss": 0.0035, + "step": 992 + }, + { + "epoch": 62.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 70.90889013651437, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6756756756756757, \"f1-score\": 0.6944444444444444, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7938311688311688, \"recall\": 0.7634469146565921, \"f1-score\": 0.7717174369747899, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7324857254814152, \"recall\": 0.7155172413793104, \"f1-score\": 0.7193500676132522, \"support\": 116.0}}", + "eval_f1_macro": 68.59710550887021, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.93500676132521, + "eval_loss": 1.030908465385437, + "eval_runtime": 2.3005, + "eval_samples_per_second": 50.424, + "step": 992 + }, + { + "epoch": 63.0, + "learning_rate": 0.0002283333333333333, + "loss": 0.0086, + "step": 1008 + }, + { + "epoch": 63.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 71.35518309105177, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6756756756756757, \"f1-score\": 0.7246376811594203, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7590061745852187, \"recall\": 0.7935768429720043, \"f1-score\": 0.766072320982366, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7432002359962553, \"recall\": 0.7241379310344828, \"f1-score\": 0.724978287366558, \"support\": 116.0}}", + "eval_f1_macro": 68.09531742065475, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.49782873665578, + "eval_loss": 1.0668530464172363, + "eval_runtime": 2.4387, + "eval_samples_per_second": 47.567, + "step": 1008 + }, + { + "epoch": 64.0, + "learning_rate": 0.00022666666666666663, + "loss": 0.0016, + "step": 1024 + }, + { + "epoch": 64.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 72.52369399223397, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.7096774193548387, \"f1-score\": 0.7096774193548389, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7666666666666667, \"recall\": 0.6216216216216216, \"f1-score\": 0.6865671641791045, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.875, \"f1-score\": 0.9333333333333333, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7851557558507274, \"recall\": 0.8105096023442797, \"f1-score\": 0.7911406834106129, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7446162947937796, \"recall\": 0.7327586206896551, \"f1-score\": 0.7321943552783927, \"support\": 116.0}}", + "eval_f1_macro": 70.32361630316558, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.21943552783927, + "eval_loss": 1.0520766973495483, + "eval_runtime": 2.3594, + "eval_samples_per_second": 49.164, + "step": 1024 + }, + { + "epoch": 65.0, + "learning_rate": 0.000225, + "loss": 0.0091, + "step": 1040 + }, + { + "epoch": 65.0, + "eval_accuracy": 68.96551724137932, + "eval_average_metrics": 68.31228885061847, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.5454545454545454, \"recall\": 0.6666666666666666, \"f1-score\": 0.6, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6774193548387096, \"recall\": 0.6774193548387096, \"f1-score\": 0.6774193548387096, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.5675675675675675, \"f1-score\": 0.6461538461538462, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"request\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"accuracy\": 0.6896551724137931, \"macro avg\": {\"precision\": 0.7318522930922124, \"recall\": 0.7754150319674513, \"f1-score\": 0.7476391104415299, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7018025078369906, \"recall\": 0.6896551724137931, \"f1-score\": 0.6886131110269041, \"support\": 116.0}}", + "eval_f1_macro": 66.45680981702488, + "eval_f1_micro": 68.96551724137932, + "eval_f1_weighted": 68.86131110269041, + "eval_loss": 1.0356357097625732, + "eval_runtime": 2.3808, + "eval_samples_per_second": 48.723, + "step": 1040 + }, + { + "epoch": 66.0, + "learning_rate": 0.00022333333333333333, + "loss": 0.0028, + "step": 1056 + }, + { + "epoch": 66.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 72.2270820572477, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6756756756756757, \"f1-score\": 0.7246376811594203, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7725114193404634, \"recall\": 0.7976091010365204, \"f1-score\": 0.776483943258395, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7479355868522958, \"recall\": 0.7327586206896551, \"f1-score\": 0.7333580913475801, \"support\": 116.0}}", + "eval_f1_macro": 69.02079495630178, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.335809134758, + "eval_loss": 1.0336496829986572, + "eval_runtime": 2.2453, + "eval_samples_per_second": 51.664, + "step": 1056 + }, + { + "epoch": 67.0, + "learning_rate": 0.00022166666666666667, + "loss": 0.0028, + "step": 1072 + }, + { + "epoch": 67.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 71.6024787948385, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6756756756756757, \"f1-score\": 0.7142857142857142, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.787405303030303, \"recall\": 0.7773358035454809, \"f1-score\": 0.7755018674136321, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7389465778474399, \"recall\": 0.7241379310344828, \"f1-score\": 0.7264882964680124, \"support\": 116.0}}", + "eval_f1_macro": 68.93349932565619, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.64882964680125, + "eval_loss": 1.002924919128418, + "eval_runtime": 2.1768, + "eval_samples_per_second": 53.289, + "step": 1072 + }, + { + "epoch": 68.0, + "learning_rate": 0.00021999999999999995, + "loss": 0.0037, + "step": 1088 + }, + { + "epoch": 68.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 72.42795900402467, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.7096774193548387, \"f1-score\": 0.7213114754098361, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6486486486486487, \"f1-score\": 0.6956521739130435, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.875, \"f1-score\": 0.9333333333333333, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7832516339869281, \"recall\": 0.7999990918337692, \"f1-score\": 0.7856896309617154, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7438979039891818, \"recall\": 0.7327586206896551, \"f1-score\": 0.7332103357045962, \"support\": 116.0}}", + "eval_f1_macro": 69.83907830770802, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.32103357045962, + "eval_loss": 1.0405433177947998, + "eval_runtime": 2.4018, + "eval_samples_per_second": 48.298, + "step": 1088 + }, + { + "epoch": 69.0, + "learning_rate": 0.0002183333333333333, + "loss": 0.0023, + "step": 1104 + }, + { + "epoch": 69.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 70.44974265216908, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.8, \"recall\": 0.6451612903225806, \"f1-score\": 0.7142857142857142, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.7027027027027027, \"f1-score\": 0.7123287671232876, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.76635101010101, \"recall\": 0.7674413324614937, \"f1-score\": 0.753787379019047, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7385013932427725, \"recall\": 0.7155172413793104, \"f1-score\": 0.7169219975334343, \"support\": 116.0}}", + "eval_f1_macro": 67.00332257947085, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.69219975334343, + "eval_loss": 1.0624644756317139, + "eval_runtime": 2.39, + "eval_samples_per_second": 48.535, + "step": 1104 + }, + { + "epoch": 70.0, + "learning_rate": 0.00021666666666666666, + "loss": 0.0023, + "step": 1120 + }, + { + "epoch": 70.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 69.99088604087092, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.5675675675675675, \"f1-score\": 0.6461538461538462, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7510416666666666, \"recall\": 0.7933361789208564, \"f1-score\": 0.7666375473941263, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7156250000000001, \"recall\": 0.7068965517241379, \"f1-score\": 0.704386740502893, \"support\": 116.0}}", + "eval_f1_macro": 68.1455597683668, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.4386740502893, + "eval_loss": 1.127900242805481, + "eval_runtime": 2.2682, + "eval_samples_per_second": 51.141, + "step": 1120 + }, + { + "epoch": 71.0, + "learning_rate": 0.000215, + "loss": 0.0277, + "step": 1136 + }, + { + "epoch": 71.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 73.17305223934821, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7586206896551724, \"recall\": 0.7096774193548387, \"f1-score\": 0.7333333333333333, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6756756756756757, \"f1-score\": 0.7142857142857142, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.875, \"f1-score\": 0.9333333333333333, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7873595232241277, \"recall\": 0.8033774702121476, \"f1-score\": 0.7895215557487363, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7530721374053893, \"recall\": 0.7413793103448276, \"f1-score\": 0.7423665304409517, \"support\": 116.0}}", + "eval_f1_macro": 70.17969384433212, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.23665304409515, + "eval_loss": 1.0505918264389038, + "eval_runtime": 2.3583, + "eval_samples_per_second": 49.188, + "step": 1136 + }, + { + "epoch": 72.0, + "learning_rate": 0.00021333333333333333, + "loss": 0.0013, + "step": 1152 + }, + { + "epoch": 72.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 70.18153002432922, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6470588235294118, \"recall\": 0.7096774193548387, \"f1-score\": 0.6769230769230768, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7419354838709677, \"recall\": 0.6216216216216216, \"f1-score\": 0.676470588235294, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.875, \"f1-score\": 0.9333333333333333, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.776607242970502, \"recall\": 0.7723151578998353, \"f1-score\": 0.7698744882568411, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7207992168957784, \"recall\": 0.7068965517241379, \"f1-score\": 0.7091352190743672, \"support\": 116.0}}", + "eval_f1_macro": 68.43328784505256, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.91352190743672, + "eval_loss": 1.0885376930236816, + "eval_runtime": 2.3047, + "eval_samples_per_second": 50.332, + "step": 1152 + }, + { + "epoch": 73.0, + "learning_rate": 0.00021166666666666667, + "loss": 0.0062, + "step": 1168 + }, + { + "epoch": 73.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 69.98285623150008, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.5675675675675675, \"f1-score\": 0.6461538461538462, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7474877450980392, \"recall\": 0.7997205875230069, \"f1-score\": 0.7668653226453315, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7164553752535497, \"recall\": 0.7068965517241379, \"f1-score\": 0.7038630812380989, \"support\": 116.0}}", + "eval_f1_macro": 68.1658064573628, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.38630812380991, + "eval_loss": 1.1306726932525635, + "eval_runtime": 2.2798, + "eval_samples_per_second": 50.881, + "step": 1168 + }, + { + "epoch": 74.0, + "learning_rate": 0.00020999999999999998, + "loss": 0.0045, + "step": 1184 + }, + { + "epoch": 74.0, + "eval_accuracy": 75.0, + "eval_average_metrics": 74.14612206166869, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.7297297297297297, \"f1-score\": 0.7397260273972601, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.875, \"f1-score\": 0.9333333333333333, \"support\": 8.0}, \"accuracy\": 0.75, \"macro avg\": {\"precision\": 0.8087121212121212, \"recall\": 0.8061019689043882, \"f1-score\": 0.8013653477451443, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7672849181469871, \"recall\": 0.75, \"f1-score\": 0.7535201289155082, \"support\": 116.0}}", + "eval_f1_macro": 71.23247535512394, + "eval_f1_micro": 75.0, + "eval_f1_weighted": 75.35201289155083, + "eval_loss": 1.092178225517273, + "eval_runtime": 2.2337, + "eval_samples_per_second": 51.932, + "step": 1184 + }, + { + "epoch": 75.0, + "learning_rate": 0.00020833333333333332, + "loss": 0.0082, + "step": 1200 + }, + { + "epoch": 75.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 71.86428236984912, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6470588235294118, \"recall\": 0.7096774193548387, \"f1-score\": 0.6769230769230768, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7666666666666667, \"recall\": 0.6216216216216216, \"f1-score\": 0.6865671641791045, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"request\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7881740196078432, \"recall\": 0.8000929356776131, \"f1-score\": 0.7892650616503777, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7359956051386072, \"recall\": 0.7241379310344828, \"f1-score\": 0.7247264890357742, \"support\": 116.0}}", + "eval_f1_macro": 70.15689436892248, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.47264890357744, + "eval_loss": 1.0961277484893799, + "eval_runtime": 2.248, + "eval_samples_per_second": 51.601, + "step": 1200 + }, + { + "epoch": 76.0, + "learning_rate": 0.00020666666666666666, + "loss": 0.004, + "step": 1216 + }, + { + "epoch": 76.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 70.9403461206404, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.47619047619047616, \"recall\": 0.8333333333333334, \"f1-score\": 0.6060606060606061, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7391304347826086, \"recall\": 0.5483870967741935, \"f1-score\": 0.6296296296296297, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.7297297297297297, \"f1-score\": 0.7397260273972601, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"request\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7783355684170902, \"recall\": 0.7899729366463237, \"f1-score\": 0.7746843858271134, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.740125878619132, \"recall\": 0.7155172413793104, \"f1-score\": 0.7179710191095614, \"support\": 116.0}}", + "eval_f1_macro": 68.86083429574342, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.79710191095613, + "eval_loss": 1.1594845056533813, + "eval_runtime": 2.0779, + "eval_samples_per_second": 55.827, + "step": 1216 + }, + { + "epoch": 77.0, + "learning_rate": 0.000205, + "loss": 0.0064, + "step": 1232 + }, + { + "epoch": 77.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 69.73375044753716, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.6451612903225806, \"f1-score\": 0.689655172413793, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6486486486486487, \"f1-score\": 0.6956521739130435, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7420965608465608, \"recall\": 0.776309575704737, \"f1-score\": 0.7521734433535112, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7201445904032111, \"recall\": 0.7068965517241379, \"f1-score\": 0.7069582981389781, \"support\": 116.0}}", + "eval_f1_macro": 66.85986163142323, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.69582981389783, + "eval_loss": 1.1205899715423584, + "eval_runtime": 2.1939, + "eval_samples_per_second": 52.875, + "step": 1232 + }, + { + "epoch": 78.0, + "learning_rate": 0.00020333333333333333, + "loss": 0.0179, + "step": 1248 + }, + { + "epoch": 78.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 70.08500699570796, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.696969696969697, \"recall\": 0.6216216216216216, \"f1-score\": 0.6571428571428571, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7792585784313726, \"recall\": 0.767106824566502, \"f1-score\": 0.766470668727261, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7199265090048558, \"recall\": 0.7068965517241379, \"f1-score\": 0.7082999152891438, \"support\": 116.0}}", + "eval_f1_macro": 68.13072610908988, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.82999152891438, + "eval_loss": 1.1212310791015625, + "eval_runtime": 2.3495, + "eval_samples_per_second": 49.372, + "step": 1248 + }, + { + "epoch": 79.0, + "learning_rate": 0.00020166666666666667, + "loss": 0.002, + "step": 1264 + }, + { + "epoch": 79.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 71.4467055723817, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.8095238095238095, \"recall\": 0.5483870967741935, \"f1-score\": 0.6538461538461537, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7297297297297297, \"recall\": 0.7297297297297297, \"f1-score\": 0.7297297297297297, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 0.7777777777777778, \"recall\": 0.875, \"f1-score\": 0.823529411764706, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7618511368511369, \"recall\": 0.8038618255352126, \"f1-score\": 0.7739040349334467, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7412424740010948, \"recall\": 0.7241379310344828, \"f1-score\": 0.7216776631076833, \"support\": 116.0}}", + "eval_f1_macro": 68.79146977186194, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.16776631076833, + "eval_loss": 1.1796600818634033, + "eval_runtime": 2.3002, + "eval_samples_per_second": 50.43, + "step": 1264 + }, + { + "epoch": 80.0, + "learning_rate": 0.00019999999999999998, + "loss": 0.0039, + "step": 1280 + }, + { + "epoch": 80.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 69.91079996424841, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7333333333333333, \"recall\": 0.5945945945945946, \"f1-score\": 0.6567164179104478, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7565718901196843, \"recall\": 0.7776173350770126, \"f1-score\": 0.7615964881045013, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7158540728203506, \"recall\": 0.7068965517241379, \"f1-score\": 0.7056642390287705, \"support\": 116.0}}", + "eval_f1_macro": 67.69746560928901, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.56642390287705, + "eval_loss": 1.1252835988998413, + "eval_runtime": 2.5769, + "eval_samples_per_second": 45.014, + "step": 1280 + }, + { + "epoch": 81.0, + "learning_rate": 0.00019833333333333332, + "loss": 0.0033, + "step": 1296 + }, + { + "epoch": 81.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 72.81511711278446, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.8333333333333334, \"recall\": 0.6451612903225806, \"f1-score\": 0.7272727272727272, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7567567567567568, \"recall\": 0.7567567567567568, \"f1-score\": 0.7567567567567567, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.78619812994813, \"recall\": 0.7880869781071393, \"f1-score\": 0.7734642543466073, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7654780564263323, \"recall\": 0.7413793103448276, \"f1-score\": 0.7423222821802943, \"support\": 116.0}}", + "eval_f1_macro": 68.75237816414287, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.23222821802943, + "eval_loss": 1.1057864427566528, + "eval_runtime": 2.4718, + "eval_samples_per_second": 46.929, + "step": 1296 + }, + { + "epoch": 82.0, + "learning_rate": 0.00019666666666666666, + "loss": 0.0009, + "step": 1312 + }, + { + "epoch": 82.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 71.35367532053087, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.6774193548387096, \"f1-score\": 0.7, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.7027027027027027, \"f1-score\": 0.7222222222222223, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7801992505466137, \"recall\": 0.77147359052601, \"f1-score\": 0.766136972495732, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7385352276247219, \"recall\": 0.7241379310344828, \"f1-score\": 0.7248605085338408, \"support\": 116.0}}", + "eval_f1_macro": 68.10106422184285, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.48605085338409, + "eval_loss": 1.0622020959854126, + "eval_runtime": 2.3973, + "eval_samples_per_second": 48.389, + "step": 1312 + }, + { + "epoch": 83.0, + "learning_rate": 0.000195, + "loss": 0.0018, + "step": 1328 + }, + { + "epoch": 83.0, + "eval_accuracy": 68.10344827586206, + "eval_average_metrics": 67.65207391031592, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.47619047619047616, \"recall\": 0.8333333333333334, \"f1-score\": 0.6060606060606061, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.6451612903225806, \"f1-score\": 0.6557377049180327, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.5675675675675675, \"f1-score\": 0.6461538461538462, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.6810344827586207, \"macro avg\": {\"precision\": 0.7420454545454545, \"recall\": 0.7661744405696018, \"f1-score\": 0.7444924716539445, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7040117928048962, \"recall\": 0.6810344827586207, \"f1-score\": 0.6822429049807779, \"support\": 116.0}}", + "eval_f1_macro": 66.17710859146173, + "eval_f1_micro": 68.10344827586206, + "eval_f1_weighted": 68.22429049807779, + "eval_loss": 1.2004237174987793, + "eval_runtime": 2.4489, + "eval_samples_per_second": 47.369, + "step": 1328 + }, + { + "epoch": 84.0, + "learning_rate": 0.00019333333333333333, + "loss": 0.0012, + "step": 1344 + }, + { + "epoch": 84.0, + "eval_accuracy": 68.96551724137932, + "eval_average_metrics": 68.13078144338341, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.5675675675675675, \"f1-score\": 0.6461538461538462, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.6896551724137931, \"macro avg\": {\"precision\": 0.7366228070175438, \"recall\": 0.7684705875230069, \"f1-score\": 0.7404771647095969, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7084694494857835, \"recall\": 0.6896551724137931, \"f1-score\": 0.6877189887214419, \"support\": 116.0}}", + "eval_f1_macro": 65.82019241863084, + "eval_f1_micro": 68.96551724137932, + "eval_f1_weighted": 68.77189887214419, + "eval_loss": 1.1874313354492188, + "eval_runtime": 2.4341, + "eval_samples_per_second": 47.656, + "step": 1344 + }, + { + "epoch": 85.0, + "learning_rate": 0.00019166666666666665, + "loss": 0.0044, + "step": 1360 + }, + { + "epoch": 85.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 72.94211721069689, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.8076923076923077, \"recall\": 0.6774193548387096, \"f1-score\": 0.7368421052631579, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.717948717948718, \"recall\": 0.7567567567567568, \"f1-score\": 0.736842105263158, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7986411056263998, \"recall\": 0.7782303472827666, \"f1-score\": 0.778417674250118, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7598160375797294, \"recall\": 0.7413793103448276, \"f1-score\": 0.7429992461825603, \"support\": 116.0}}", + "eval_f1_macro": 69.19268215556605, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.29992461825603, + "eval_loss": 1.1297322511672974, + "eval_runtime": 2.3474, + "eval_samples_per_second": 49.416, + "step": 1360 + }, + { + "epoch": 86.0, + "learning_rate": 0.00018999999999999998, + "loss": 0.0179, + "step": 1376 + }, + { + "epoch": 86.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 70.03168438592736, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6388888888888888, \"recall\": 0.7419354838709677, \"f1-score\": 0.6865671641791045, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7419354838709677, \"recall\": 0.6216216216216216, \"f1-score\": 0.676470588235294, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.78548183447377, \"recall\": 0.7589863048532404, \"f1-score\": 0.7633472906958421, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7257279412590869, \"recall\": 0.7068965517241379, \"f1-score\": 0.7089433469258483, \"support\": 116.0}}", + "eval_f1_macro": 67.85309250629707, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.89433469258482, + "eval_loss": 1.117130994796753, + "eval_runtime": 2.394, + "eval_samples_per_second": 48.455, + "step": 1376 + }, + { + "epoch": 87.0, + "learning_rate": 0.00018833333333333332, + "loss": 0.0013, + "step": 1392 + }, + { + "epoch": 87.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 70.59785642558124, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.8, \"recall\": 0.5161290322580645, \"f1-score\": 0.6274509803921569, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7368421052631579, \"recall\": 0.7567567567567568, \"f1-score\": 0.7466666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7662280701754387, \"recall\": 0.7875829458490748, \"f1-score\": 0.7640013458694174, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7411978221415607, \"recall\": 0.7155172413793104, \"f1-score\": 0.7137674668251471, \"support\": 116.0}}", + "eval_f1_macro": 67.91123074394821, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.3767466825147, + "eval_loss": 1.1777045726776123, + "eval_runtime": 2.2717, + "eval_samples_per_second": 51.064, + "step": 1392 + }, + { + "epoch": 88.0, + "learning_rate": 0.00018666666666666666, + "loss": 0.0047, + "step": 1408 + }, + { + "epoch": 88.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 70.77918320692154, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6571428571428571, \"recall\": 0.7419354838709677, \"f1-score\": 0.6969696969696969, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7857142857142857, \"recall\": 0.5945945945945946, \"f1-score\": 0.676923076923077, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"request\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7657738095238095, \"recall\": 0.786857926474862, \"f1-score\": 0.7709575351119469, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7286022167487685, \"recall\": 0.7155172413793104, \"f1-score\": 0.714837258752066, \"support\": 116.0}}", + "eval_f1_macro": 68.5295586766175, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.4837258752066, + "eval_loss": 1.106935739517212, + "eval_runtime": 2.2634, + "eval_samples_per_second": 51.251, + "step": 1408 + }, + { + "epoch": 89.0, + "learning_rate": 0.000185, + "loss": 0.0042, + "step": 1424 + }, + { + "epoch": 89.0, + "eval_accuracy": 69.82758620689656, + "eval_average_metrics": 69.00121160981696, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.47619047619047616, \"recall\": 0.8333333333333334, \"f1-score\": 0.6060606060606061, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7619047619047619, \"recall\": 0.5161290322580645, \"f1-score\": 0.6153846153846153, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.7027027027027027, \"f1-score\": 0.7323943661971832, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7408263305322129, \"recall\": 0.780826189092318, \"f1-score\": 0.7486785700091854, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7248671882546122, \"recall\": 0.6982758620689655, \"f1-score\": 0.6980046780243607, \"support\": 116.0}}", + "eval_f1_macro": 66.5492062230387, + "eval_f1_micro": 69.82758620689656, + "eval_f1_weighted": 69.80046780243605, + "eval_loss": 1.2019020318984985, + "eval_runtime": 2.4101, + "eval_samples_per_second": 48.131, + "step": 1424 + }, + { + "epoch": 90.0, + "learning_rate": 0.00018333333333333334, + "loss": 0.0019, + "step": 1440 + }, + { + "epoch": 90.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 72.61043030355066, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7931034482758621, \"recall\": 0.6216216216216216, \"f1-score\": 0.696969696969697, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"request\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7909782088122606, \"recall\": 0.8105096023442797, \"f1-score\": 0.7932515809354045, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7506378814902894, \"recall\": 0.7327586206896551, \"f1-score\": 0.7337874543756896, \"support\": 116.0}}", + "eval_f1_macro": 70.51125163870262, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.37874543756897, + "eval_loss": 1.147979974746704, + "eval_runtime": 2.2001, + "eval_samples_per_second": 52.726, + "step": 1440 + }, + { + "epoch": 91.0, + "learning_rate": 0.00018166666666666665, + "loss": 0.0084, + "step": 1456 + }, + { + "epoch": 91.0, + "eval_accuracy": 69.82758620689656, + "eval_average_metrics": 69.01145238292938, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7307692307692307, \"recall\": 0.6129032258064516, \"f1-score\": 0.6666666666666667, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7419354838709677, \"recall\": 0.6216216216216216, \"f1-score\": 0.676470588235294, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7349178151597506, \"recall\": 0.7827878281507313, \"f1-score\": 0.7504109538552264, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7118510614339314, \"recall\": 0.6982758620689655, \"f1-score\": 0.6968744121968207, \"support\": 116.0}}", + "eval_f1_macro": 66.70319589824234, + "eval_f1_micro": 69.82758620689656, + "eval_f1_weighted": 69.68744121968207, + "eval_loss": 1.1545029878616333, + "eval_runtime": 2.3358, + "eval_samples_per_second": 49.662, + "step": 1456 + }, + { + "epoch": 92.0, + "learning_rate": 0.00017999999999999998, + "loss": 0.0013, + "step": 1472 + }, + { + "epoch": 92.0, + "eval_accuracy": 69.82758620689656, + "eval_average_metrics": 69.01145238292938, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7307692307692307, \"recall\": 0.6129032258064516, \"f1-score\": 0.6666666666666667, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7419354838709677, \"recall\": 0.6216216216216216, \"f1-score\": 0.676470588235294, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7349178151597506, \"recall\": 0.7827878281507313, \"f1-score\": 0.7504109538552264, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7118510614339314, \"recall\": 0.6982758620689655, \"f1-score\": 0.6968744121968207, \"support\": 116.0}}", + "eval_f1_macro": 66.70319589824234, + "eval_f1_micro": 69.82758620689656, + "eval_f1_weighted": 69.68744121968207, + "eval_loss": 1.1468178033828735, + "eval_runtime": 2.2416, + "eval_samples_per_second": 51.749, + "step": 1472 + }, + { + "epoch": 93.0, + "learning_rate": 0.00017833333333333332, + "loss": 0.0014, + "step": 1488 + }, + { + "epoch": 93.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 70.0316545956725, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.7096774193548387, \"f1-score\": 0.6875, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.5675675675675675, \"f1-score\": 0.6461538461538462, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7582831996434938, \"recall\": 0.7898639566986341, \"f1-score\": 0.7682870793333388, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7168341631323376, \"recall\": 0.7068965517241379, \"f1-score\": 0.7045512320823231, \"support\": 116.0}}", + "eval_f1_macro": 68.29218482963012, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.45512320823231, + "eval_loss": 1.1516327857971191, + "eval_runtime": 2.4305, + "eval_samples_per_second": 47.727, + "step": 1488 + }, + { + "epoch": 94.0, + "learning_rate": 0.00017666666666666666, + "loss": 0.002, + "step": 1504 + }, + { + "epoch": 94.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 69.98519551620497, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.6451612903225806, \"f1-score\": 0.689655172413793, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.696969696969697, \"recall\": 0.6216216216216216, \"f1-score\": 0.6571428571428571, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7504970043572985, \"recall\": 0.7885561973263586, \"f1-score\": 0.7650281414743756, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7132494655821227, \"recall\": 0.7068965517241379, \"f1-score\": 0.7055897025560334, \"support\": 116.0}}", + "eval_f1_macro": 68.00250146438894, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.55897025560334, + "eval_loss": 1.162688136100769, + "eval_runtime": 2.1458, + "eval_samples_per_second": 54.059, + "step": 1504 + }, + { + "epoch": 95.0, + "learning_rate": 0.000175, + "loss": 0.0055, + "step": 1520 + }, + { + "epoch": 95.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 71.0756914574162, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.625, \"recall\": 0.8333333333333334, \"f1-score\": 0.7142857142857143, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6470588235294118, \"recall\": 0.7096774193548387, \"f1-score\": 0.6769230769230768, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7333333333333333, \"recall\": 0.5945945945945946, \"f1-score\": 0.6567164179104478, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"request\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7833444741532977, \"recall\": 0.7932423350770126, \"f1-score\": 0.7837480040810814, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7255985309484294, \"recall\": 0.7155172413793104, \"f1-score\": 0.7153282830215106, \"support\": 116.0}}", + "eval_f1_macro": 69.66648925165167, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.53282830215106, + "eval_loss": 1.1537128686904907, + "eval_runtime": 2.3897, + "eval_samples_per_second": 48.543, + "step": 1520 + }, + { + "epoch": 96.0, + "learning_rate": 0.0001733333333333333, + "loss": 0.0049, + "step": 1536 + }, + { + "epoch": 96.0, + "eval_accuracy": 69.82758620689656, + "eval_average_metrics": 68.8282167495548, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7619047619047619, \"recall\": 0.5161290322580645, \"f1-score\": 0.6153846153846153, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.717948717948718, \"recall\": 0.7567567567567568, \"f1-score\": 0.736842105263158, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.75531661321135, \"recall\": 0.758069056960186, \"f1-score\": 0.74253787384093, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.722639761931958, \"recall\": 0.6982758620689655, \"f1-score\": 0.6965432802078783, \"support\": 116.0}}", + "eval_f1_macro": 66.00336656363822, + "eval_f1_micro": 69.82758620689656, + "eval_f1_weighted": 69.65432802078783, + "eval_loss": 1.1960561275482178, + "eval_runtime": 2.4859, + "eval_samples_per_second": 46.663, + "step": 1536 + }, + { + "epoch": 97.0, + "learning_rate": 0.00017166666666666665, + "loss": 0.0142, + "step": 1552 + }, + { + "epoch": 97.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 70.00234719378676, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6774193548387096, \"recall\": 0.6774193548387096, \"f1-score\": 0.6774193548387096, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7058823529411765, \"recall\": 0.6486486486486487, \"f1-score\": 0.676056338028169, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7808486908937573, \"recall\": 0.7647168337692531, \"f1-score\": 0.7633916095426883, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7211529596164484, \"recall\": 0.7068965517241379, \"f1-score\": 0.7077304647096941, \"support\": 116.0}}", + "eval_f1_macro": 67.85703195935007, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.77304647096942, + "eval_loss": 1.1356544494628906, + "eval_runtime": 2.2827, + "eval_samples_per_second": 50.817, + "step": 1552 + }, + { + "epoch": 98.0, + "learning_rate": 0.00016999999999999999, + "loss": 0.0011, + "step": 1568 + }, + { + "epoch": 98.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 70.02228110967859, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5, \"recall\": 0.8333333333333334, \"f1-score\": 0.625, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.72, \"recall\": 0.5806451612903226, \"f1-score\": 0.6428571428571428, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6486486486486487, \"f1-score\": 0.6956521739130435, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7577489177489177, \"recall\": 0.7960228373534826, \"f1-score\": 0.7664120539520156, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7244894760412003, \"recall\": 0.7068965517241379, \"f1-score\": 0.7058429818704093, \"support\": 116.0}}", + "eval_f1_macro": 68.12551590684583, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.58429818704093, + "eval_loss": 1.159824252128601, + "eval_runtime": 2.3383, + "eval_samples_per_second": 49.609, + "step": 1568 + }, + { + "epoch": 99.0, + "learning_rate": 0.00016833333333333332, + "loss": 0.001, + "step": 1584 + }, + { + "epoch": 99.0, + "eval_accuracy": 69.82758620689656, + "eval_average_metrics": 68.93085179723111, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.8095238095238095, \"recall\": 0.5483870967741935, \"f1-score\": 0.6538461538461537, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.6756756756756757, \"recall\": 0.6756756756756757, \"f1-score\": 0.6756756756756757, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 1.0, \"f1-score\": 0.8571428571428571, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7446777134277135, \"recall\": 0.7797439576673447, \"f1-score\": 0.7495235620235621, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7170566502463054, \"recall\": 0.6982758620689655, \"f1-score\": 0.694439181508147, \"support\": 116.0}}", + "eval_f1_macro": 66.62431662431662, + "eval_f1_micro": 69.82758620689656, + "eval_f1_weighted": 69.4439181508147, + "eval_loss": 1.1499238014221191, + "eval_runtime": 2.5501, + "eval_samples_per_second": 45.488, + "step": 1584 + }, + { + "epoch": 100.0, + "learning_rate": 0.00016666666666666666, + "loss": 0.007, + "step": 1600 + }, + { + "epoch": 100.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 71.4194327673127, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.625, \"recall\": 0.8333333333333334, \"f1-score\": 0.7142857142857143, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7058823529411765, \"recall\": 0.6486486486486487, \"f1-score\": 0.676056338028169, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7686330213903743, \"recall\": 0.7924946115470308, \"f1-score\": 0.7730349110281496, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7297237076648841, \"recall\": 0.7241379310344828, \"f1-score\": 0.7213593054874093, \"support\": 116.0}}", + "eval_f1_macro": 68.71421431361331, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.13593054874093, + "eval_loss": 1.0965989828109741, + "eval_runtime": 2.3288, + "eval_samples_per_second": 49.81, + "step": 1600 + }, + { + "epoch": 101.0, + "learning_rate": 0.000165, + "loss": 0.0002, + "step": 1616 + }, + { + "epoch": 101.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 71.41445986904613, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7352941176470589, \"recall\": 0.6756756756756757, \"f1-score\": 0.7042253521126761, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7724673202614379, \"recall\": 0.7819841010365204, \"f1-score\": 0.7692147679474433, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7374070317782286, \"recall\": 0.7241379310344828, \"f1-score\": 0.7245560722951524, \"support\": 116.0}}", + "eval_f1_macro": 68.37464603977274, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.45560722951524, + "eval_loss": 1.1273354291915894, + "eval_runtime": 2.3586, + "eval_samples_per_second": 49.182, + "step": 1616 + }, + { + "epoch": 102.0, + "learning_rate": 0.0001633333333333333, + "loss": 0.0002, + "step": 1632 + }, + { + "epoch": 102.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 72.15508814394651, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.7027027027027027, \"f1-score\": 0.7222222222222223, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7768849206349207, \"recall\": 0.7853624794148988, \"f1-score\": 0.7729985672429891, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7472427476737821, \"recall\": 0.7327586206896551, \"f1-score\": 0.7335764468292271, \"support\": 116.0}}", + "eval_f1_macro": 68.71098375493236, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.35764468292271, + "eval_loss": 1.1395922899246216, + "eval_runtime": 2.2846, + "eval_samples_per_second": 50.775, + "step": 1632 + }, + { + "epoch": 103.0, + "learning_rate": 0.00016166666666666665, + "loss": 0.0023, + "step": 1648 + }, + { + "epoch": 103.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 70.42410021572994, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.8260869565217391, \"recall\": 0.6129032258064516, \"f1-score\": 0.7037037037037037, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6756756756756757, \"f1-score\": 0.6944444444444444, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 1.0, \"f1-score\": 0.8571428571428571, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7473259956156375, \"recall\": 0.787808473796377, \"f1-score\": 0.756589450157962, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7320065282484809, \"recall\": 0.7155172413793104, \"f1-score\": 0.7134055701746107, \"support\": 116.0}}", + "eval_f1_macro": 67.25239556959663, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.34055701746107, + "eval_loss": 1.1940408945083618, + "eval_runtime": 2.2464, + "eval_samples_per_second": 51.637, + "step": 1648 + }, + { + "epoch": 104.0, + "learning_rate": 0.00015999999999999999, + "loss": 0.0078, + "step": 1664 + }, + { + "epoch": 104.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 70.8781299077741, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7419354838709677, \"recall\": 0.6216216216216216, \"f1-score\": 0.676470588235294, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7823500704045059, \"recall\": 0.7809957134553909, \"f1-score\": 0.7730421335200748, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7330855381640977, \"recall\": 0.7155172413793104, \"f1-score\": 0.7169421504233877, \"support\": 116.0}}", + "eval_f1_macro": 68.71485631289553, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.69421504233877, + "eval_loss": 1.1880056858062744, + "eval_runtime": 2.3241, + "eval_samples_per_second": 49.913, + "step": 1664 + }, + { + "epoch": 105.0, + "learning_rate": 0.00015833333333333332, + "loss": 0.0007, + "step": 1680 + }, + { + "epoch": 105.0, + "eval_accuracy": 68.10344827586206, + "eval_average_metrics": 67.81830142186269, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.44, \"recall\": 0.9166666666666666, \"f1-score\": 0.5945945945945945, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.68, \"recall\": 0.5483870967741935, \"f1-score\": 0.6071428571428571, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6486486486486487, \"f1-score\": 0.6956521739130435, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.6810344827586207, \"macro avg\": {\"precision\": 0.7641883116883117, \"recall\": 0.7607405792889663, \"f1-score\": 0.747283997323959, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7193484102104792, \"recall\": 0.6810344827586207, \"f1-score\": 0.6864106492915251, \"support\": 116.0}}", + "eval_f1_macro": 66.42524420657413, + "eval_f1_micro": 68.10344827586206, + "eval_f1_weighted": 68.6410649291525, + "eval_loss": 1.272948980331421, + "eval_runtime": 2.2187, + "eval_samples_per_second": 52.283, + "step": 1680 + }, + { + "epoch": 106.0, + "learning_rate": 0.00015666666666666666, + "loss": 0.0006, + "step": 1696 + }, + { + "epoch": 106.0, + "eval_accuracy": 68.96551724137932, + "eval_average_metrics": 68.14354636082183, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7096774193548387, \"recall\": 0.5945945945945946, \"f1-score\": 0.6470588235294118, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.6896551724137931, \"macro avg\": {\"precision\": 0.7390253339850115, \"recall\": 0.7579600770124963, \"f1-score\": 0.7403022684483629, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7009838372602555, \"recall\": 0.6896551724137931, \"f1-score\": 0.6883850487622973, \"support\": 116.0}}", + "eval_f1_macro": 65.80464608429892, + "eval_f1_micro": 68.96551724137932, + "eval_f1_weighted": 68.83850487622973, + "eval_loss": 1.1888731718063354, + "eval_runtime": 2.3283, + "eval_samples_per_second": 49.821, + "step": 1696 + }, + { + "epoch": 107.0, + "learning_rate": 0.000155, + "loss": 0.0001, + "step": 1712 + }, + { + "epoch": 107.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 71.35838392975816, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.84, \"recall\": 0.6774193548387096, \"f1-score\": 0.75, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.6842105263157895, \"recall\": 0.7027027027027027, \"f1-score\": 0.6933333333333334, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.778831240031898, \"recall\": 0.774945812748232, \"f1-score\": 0.7654010450333979, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7448383105098169, \"recall\": 0.7241379310344828, \"f1-score\": 0.7257030106472298, \"support\": 116.0}}", + "eval_f1_macro": 68.03564844741315, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.57030106472297, + "eval_loss": 1.1652153730392456, + "eval_runtime": 2.2279, + "eval_samples_per_second": 52.067, + "step": 1712 + }, + { + "epoch": 108.0, + "learning_rate": 0.0001533333333333333, + "loss": 0.0011, + "step": 1728 + }, + { + "epoch": 108.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 69.89276815025954, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.71875, \"recall\": 0.6216216216216216, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7675398284313726, \"recall\": 0.7752273442797636, \"f1-score\": 0.7605942362755609, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7202808062880325, \"recall\": 0.7068965517241379, \"f1-score\": 0.7058338569838294, \"support\": 116.0}}", + "eval_f1_macro": 67.60837655782763, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.58338569838294, + "eval_loss": 1.1787962913513184, + "eval_runtime": 2.3198, + "eval_samples_per_second": 50.005, + "step": 1728 + }, + { + "epoch": 109.0, + "learning_rate": 0.00015166666666666665, + "loss": 0.0003, + "step": 1744 + }, + { + "epoch": 109.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 70.57402020858294, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.5, \"recall\": 0.6666666666666666, \"f1-score\": 0.5714285714285715, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7916666666666666, \"recall\": 0.6129032258064516, \"f1-score\": 0.6909090909090909, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7352941176470589, \"recall\": 0.6756756756756757, \"f1-score\": 0.7042253521126761, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7625055704099822, \"recall\": 0.787808473796377, \"f1-score\": 0.7609192147499579, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7385345595918618, \"recall\": 0.7155172413793104, \"f1-score\": 0.7155536902514009, \"support\": 116.0}}", + "eval_f1_macro": 67.63726353332959, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.5553690251401, + "eval_loss": 1.215308427810669, + "eval_runtime": 2.4262, + "eval_samples_per_second": 47.812, + "step": 1744 + }, + { + "epoch": 110.0, + "learning_rate": 0.00015, + "loss": 0.0007, + "step": 1760 + }, + { + "epoch": 110.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 70.4562231169158, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7916666666666666, \"recall\": 0.6129032258064516, \"f1-score\": 0.6909090909090909, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.7027027027027027, \"f1-score\": 0.7222222222222223, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7605023448773449, \"recall\": 0.7772979632858665, \"f1-score\": 0.7553140572877415, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7382715703836394, \"recall\": 0.7155172413793104, \"f1-score\": 0.7158241687733521, \"support\": 116.0}}", + "eval_f1_macro": 67.1390273144659, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.5824168773352, + "eval_loss": 1.2416104078292847, + "eval_runtime": 2.2686, + "eval_samples_per_second": 51.133, + "step": 1760 + }, + { + "epoch": 111.0, + "learning_rate": 0.00014833333333333332, + "loss": 0.0006, + "step": 1776 + }, + { + "epoch": 111.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 70.4444570130092, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.8, \"recall\": 0.6451612903225806, \"f1-score\": 0.7142857142857142, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.7027027027027027, \"f1-score\": 0.7123287671232876, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7635101010101011, \"recall\": 0.7674413324614937, \"f1-score\": 0.7536473229966381, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.73673807035876, \"recall\": 0.7155172413793104, \"f1-score\": 0.7168350662091805, \"support\": 116.0}}", + "eval_f1_macro": 66.99087315525671, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.68350662091805, + "eval_loss": 1.234352946281433, + "eval_runtime": 2.3796, + "eval_samples_per_second": 48.747, + "step": 1776 + }, + { + "epoch": 112.0, + "learning_rate": 0.00014666666666666664, + "loss": 0.0005, + "step": 1792 + }, + { + "epoch": 112.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 71.34661390855857, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.7096774193548387, \"f1-score\": 0.7213114754098361, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6756756756756757, \"f1-score\": 0.7142857142857142, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7791035353535354, \"recall\": 0.7721274702121477, \"f1-score\": 0.7649352802115073, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7423066875653083, \"recall\": 0.7241379310344828, \"f1-score\": 0.7256462229742598, \"support\": 116.0}}", + "eval_f1_macro": 67.99424712991176, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.56462229742598, + "eval_loss": 1.2403353452682495, + "eval_runtime": 2.3489, + "eval_samples_per_second": 49.385, + "step": 1792 + }, + { + "epoch": 113.0, + "learning_rate": 0.000145, + "loss": 0.0002, + "step": 1808 + }, + { + "epoch": 113.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 70.37924043117656, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.7096774193548387, \"f1-score\": 0.7213114754098361, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7741935483870968, \"recall\": 0.6486486486486487, \"f1-score\": 0.7058823529411764, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7512565167807104, \"recall\": 0.7687490918337693, \"f1-score\": 0.7517535702805713, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7304703037044528, \"recall\": 0.7155172413793104, \"f1-score\": 0.7159097386834893, \"support\": 116.0}}", + "eval_f1_macro": 66.82253958049523, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.59097386834891, + "eval_loss": 1.2404866218566895, + "eval_runtime": 2.2047, + "eval_samples_per_second": 52.614, + "step": 1808 + }, + { + "epoch": 114.0, + "learning_rate": 0.00014333333333333334, + "loss": 0.0001, + "step": 1824 + }, + { + "epoch": 114.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 70.44727017147791, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5, \"recall\": 0.8333333333333334, \"f1-score\": 0.625, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7727272727272727, \"recall\": 0.5483870967741935, \"f1-score\": 0.6415094339622641, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7941176470588235, \"recall\": 0.7297297297297297, \"f1-score\": 0.7605633802816901, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 1.0, \"f1-score\": 0.8571428571428571, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7604389483065954, \"recall\": 0.7865007144241015, \"f1-score\": 0.7558662687363454, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7455767103079475, \"recall\": 0.7155172413793104, \"f1-score\": 0.7149751963348554, \"support\": 116.0}}", + "eval_f1_macro": 67.18811277656404, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.49751963348554, + "eval_loss": 1.2582147121429443, + "eval_runtime": 2.2366, + "eval_samples_per_second": 51.864, + "step": 1824 + }, + { + "epoch": 115.0, + "learning_rate": 0.00014166666666666665, + "loss": 0.0014, + "step": 1840 + }, + { + "epoch": 115.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 71.7426864751411, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.6428571428571429, \"recall\": 0.75, \"f1-score\": 0.6923076923076924, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6388888888888888, \"recall\": 0.7419354838709677, \"f1-score\": 0.6865671641791045, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7857142857142857, \"recall\": 0.5945945945945946, \"f1-score\": 0.676923076923077, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7810019841269841, \"recall\": 0.8007468153637509, \"f1-score\": 0.7859342421430876, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7357006020799124, \"recall\": 0.7241379310344828, \"f1-score\": 0.7228233816983779, \"support\": 116.0}}", + "eval_f1_macro": 69.86082152383003, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.2823381698378, + "eval_loss": 1.202860713005066, + "eval_runtime": 2.3445, + "eval_samples_per_second": 49.477, + "step": 1840 + }, + { + "epoch": 116.0, + "learning_rate": 0.00014, + "loss": 0.0003, + "step": 1856 + }, + { + "epoch": 116.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 70.5549298094937, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7916666666666666, \"recall\": 0.6129032258064516, \"f1-score\": 0.6909090909090909, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.7027027027027027, \"f1-score\": 0.7027027027027027, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6923076923076923, \"recall\": 1.0, \"f1-score\": 0.8181818181818181, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.770730466042966, \"recall\": 0.7807701855080886, \"f1-score\": 0.7614585798409328, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7351072060123783, \"recall\": 0.7155172413793104, \"f1-score\": 0.7143106386514095, \"support\": 116.0}}", + "eval_f1_macro": 67.6852070969718, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.43106386514096, + "eval_loss": 1.2238714694976807, + "eval_runtime": 2.1973, + "eval_samples_per_second": 52.793, + "step": 1856 + }, + { + "epoch": 117.0, + "learning_rate": 0.00013833333333333333, + "loss": 0.0028, + "step": 1872 + }, + { + "epoch": 117.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 70.59935701207709, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6486486486486487, \"f1-score\": 0.6956521739130435, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 1.0, \"f1-score\": 0.8571428571428571, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7718248663101605, \"recall\": 0.778605722658142, \"f1-score\": 0.7632609894756968, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7304835884196939, \"recall\": 0.7155172413793104, \"f1-score\": 0.7144855848571772, \"support\": 116.0}}", + "eval_f1_macro": 67.84542128672861, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.44855848571771, + "eval_loss": 1.1936085224151611, + "eval_runtime": 2.2205, + "eval_samples_per_second": 52.241, + "step": 1872 + }, + { + "epoch": 118.0, + "learning_rate": 0.00013666666666666666, + "loss": 0.0033, + "step": 1888 + }, + { + "epoch": 118.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 70.58604104216032, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7741935483870968, \"recall\": 0.6486486486486487, \"f1-score\": 0.7058823529411764, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 1.0, \"f1-score\": 0.8571428571428571, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7707640925382861, \"recall\": 0.778605722658142, \"f1-score\": 0.7616661986358225, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.734819833485017, \"recall\": 0.7155172413793104, \"f1-score\": 0.715370537918172, \"support\": 116.0}}", + "eval_f1_macro": 67.703662100962, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.5370537918172, + "eval_loss": 1.1629191637039185, + "eval_runtime": 2.2141, + "eval_samples_per_second": 52.391, + "step": 1888 + }, + { + "epoch": 119.0, + "learning_rate": 0.000135, + "loss": 0.0002, + "step": 1904 + }, + { + "epoch": 119.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 69.70267547622983, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7037037037037037, \"recall\": 0.6129032258064516, \"f1-score\": 0.6551724137931035, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6756756756756757, \"f1-score\": 0.7142857142857142, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 1.0, \"f1-score\": 0.8571428571428571, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7544847989700931, \"recall\": 0.7739195849074881, \"f1-score\": 0.7521332342793323, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7222518457051925, \"recall\": 0.7068965517241379, \"f1-score\": 0.7057510406859552, \"support\": 116.0}}", + "eval_f1_macro": 66.8562874914962, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.57510406859552, + "eval_loss": 1.1673834323883057, + "eval_runtime": 2.2399, + "eval_samples_per_second": 51.788, + "step": 1904 + }, + { + "epoch": 120.0, + "learning_rate": 0.0001333333333333333, + "loss": 0.0008, + "step": 1920 + }, + { + "epoch": 120.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 69.87312178530907, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7272727272727273, \"recall\": 0.6486486486486487, \"f1-score\": 0.6857142857142857, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.77896148989899, \"recall\": 0.7548602029448803, \"f1-score\": 0.7561795230912878, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7262114681295715, \"recall\": 0.7068965517241379, \"f1-score\": 0.7089721918829424, \"support\": 116.0}}", + "eval_f1_macro": 67.21595760811448, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.89721918829423, + "eval_loss": 1.1733537912368774, + "eval_runtime": 2.3365, + "eval_samples_per_second": 49.648, + "step": 1920 + }, + { + "epoch": 121.0, + "learning_rate": 0.00013166666666666665, + "loss": 0.0007, + "step": 1936 + }, + { + "epoch": 121.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 71.45106394557037, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.6774193548387096, \"f1-score\": 0.7, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.7027027027027027, \"f1-score\": 0.7222222222222223, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7872253943374633, \"recall\": 0.77147359052601, \"f1-score\": 0.7691113624937154, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7420511167389884, \"recall\": 0.7241379310344828, \"f1-score\": 0.7261121513149911, \"support\": 116.0}}", + "eval_f1_macro": 68.36545444388584, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.61121513149911, + "eval_loss": 1.1702574491500854, + "eval_runtime": 2.2206, + "eval_samples_per_second": 52.239, + "step": 1936 + }, + { + "epoch": 122.0, + "learning_rate": 0.00013, + "loss": 0.0003, + "step": 1952 + }, + { + "epoch": 122.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 71.22487176013084, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.6451612903225806, \"f1-score\": 0.689655172413793, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.7027027027027027, \"f1-score\": 0.7323943661971832, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7575435729847495, \"recall\": 0.7813302213503825, \"f1-score\": 0.7620833896299759, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7343325069491398, \"recall\": 0.7241379310344828, \"f1-score\": 0.723311550887401, \"support\": 116.0}}", + "eval_f1_macro": 67.74074574488674, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.33115508874009, + "eval_loss": 1.178951382637024, + "eval_runtime": 2.1307, + "eval_samples_per_second": 54.442, + "step": 1952 + }, + { + "epoch": 123.0, + "learning_rate": 0.00012833333333333333, + "loss": 0.0006, + "step": 1968 + }, + { + "epoch": 123.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 71.50574768110899, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6764705882352942, \"recall\": 0.7419354838709677, \"f1-score\": 0.7076923076923077, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.8148148148148148, \"recall\": 0.5945945945945946, \"f1-score\": 0.6875, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7682220931301813, \"recall\": 0.7990107042526398, \"f1-score\": 0.7757295706744236, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7399958241743231, \"recall\": 0.7241379310344828, \"f1-score\": 0.7224166490203509, \"support\": 116.0}}", + "eval_f1_macro": 68.95373961550433, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.24166490203508, + "eval_loss": 1.2310043573379517, + "eval_runtime": 2.426, + "eval_samples_per_second": 47.816, + "step": 1968 + }, + { + "epoch": 124.0, + "learning_rate": 0.00012666666666666666, + "loss": 0.0023, + "step": 1984 + }, + { + "epoch": 124.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 71.59508898183678, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7931034482758621, \"recall\": 0.6216216216216216, \"f1-score\": 0.696969696969697, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7748571770662288, \"recall\": 0.7948846023442797, \"f1-score\": 0.7773406332229862, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7407856647415209, \"recall\": 0.7241379310344828, \"f1-score\": 0.7245582454507403, \"support\": 116.0}}", + "eval_f1_macro": 69.09694517537655, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.45582454507404, + "eval_loss": 1.1732732057571411, + "eval_runtime": 2.2592, + "eval_samples_per_second": 51.345, + "step": 1984 + }, + { + "epoch": 125.0, + "learning_rate": 0.000125, + "loss": 0.0014, + "step": 2000 + }, + { + "epoch": 125.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 71.59508898183678, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7931034482758621, \"recall\": 0.6216216216216216, \"f1-score\": 0.696969696969697, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7748571770662288, \"recall\": 0.7948846023442797, \"f1-score\": 0.7773406332229862, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7407856647415209, \"recall\": 0.7241379310344828, \"f1-score\": 0.7245582454507403, \"support\": 116.0}}", + "eval_f1_macro": 69.09694517537655, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.45582454507404, + "eval_loss": 1.2506325244903564, + "eval_runtime": 2.2452, + "eval_samples_per_second": 51.666, + "step": 2000 + }, + { + "epoch": 126.0, + "learning_rate": 0.0001233333333333333, + "loss": 0.0002, + "step": 2016 + }, + { + "epoch": 126.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 71.59508898183678, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7931034482758621, \"recall\": 0.6216216216216216, \"f1-score\": 0.696969696969697, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7748571770662288, \"recall\": 0.7948846023442797, \"f1-score\": 0.7773406332229862, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7407856647415209, \"recall\": 0.7241379310344828, \"f1-score\": 0.7245582454507403, \"support\": 116.0}}", + "eval_f1_macro": 69.09694517537655, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.45582454507404, + "eval_loss": 1.2512825727462769, + "eval_runtime": 2.344, + "eval_samples_per_second": 49.489, + "step": 2016 + }, + { + "epoch": 127.0, + "learning_rate": 0.00012166666666666665, + "loss": 0.001, + "step": 2032 + }, + { + "epoch": 127.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 70.6663288540288, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7666666666666667, \"recall\": 0.6216216216216216, \"f1-score\": 0.6865671641791045, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7685763888888889, \"recall\": 0.7792596023442797, \"f1-score\": 0.7653260309098764, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7307112068965518, \"recall\": 0.7155172413793104, \"f1-score\": 0.7153288661493075, \"support\": 116.0}}", + "eval_f1_macro": 68.02898052532235, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.53288661493076, + "eval_loss": 1.2251347303390503, + "eval_runtime": 2.2231, + "eval_samples_per_second": 52.179, + "step": 2032 + }, + { + "epoch": 128.0, + "learning_rate": 0.00011999999999999999, + "loss": 0.0001, + "step": 2048 + }, + { + "epoch": 128.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 72.28350002650896, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.6774193548387096, \"f1-score\": 0.7, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.7027027027027027, \"f1-score\": 0.7323943661971832, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7937443655623169, \"recall\": 0.7853624794148988, \"f1-score\": 0.7781209757286809, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7513712084120211, \"recall\": 0.7327586206896551, \"f1-score\": 0.7341596701444429, \"support\": 116.0}}", + "eval_f1_macro": 69.16630895366053, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.41596701444429, + "eval_loss": 1.1956950426101685, + "eval_runtime": 2.2387, + "eval_samples_per_second": 51.816, + "step": 2048 + }, + { + "epoch": 129.0, + "learning_rate": 0.00011833333333333331, + "loss": 0.0002, + "step": 2064 + }, + { + "epoch": 129.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 72.19392646661663, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.7297297297297297, \"f1-score\": 0.7397260273972601, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7871260683760684, \"recall\": 0.7847085997287611, \"f1-score\": 0.7743547708399907, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7522656940760389, \"recall\": 0.7327586206896551, \"f1-score\": 0.7339244654275855, \"support\": 116.0}}", + "eval_f1_macro": 68.83153518577694, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.39244654275855, + "eval_loss": 1.2062249183654785, + "eval_runtime": 2.0692, + "eval_samples_per_second": 56.061, + "step": 2064 + }, + { + "epoch": 130.0, + "learning_rate": 0.00011666666666666665, + "loss": 0.0016, + "step": 2080 + }, + { + "epoch": 130.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 69.63518777151155, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7142857142857143, \"recall\": 0.6451612903225806, \"f1-score\": 0.6779661016949152, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7931034482758621, \"recall\": 0.6216216216216216, \"f1-score\": 0.696969696969697, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6923076923076923, \"recall\": 1.0, \"f1-score\": 0.8181818181818181, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7402398846364364, \"recall\": 0.7850839751041363, \"f1-score\": 0.7503301782962799, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7245843732169535, \"recall\": 0.7068965517241379, \"f1-score\": 0.7046542489266042, \"support\": 116.0}}", + "eval_f1_macro": 66.6960158485582, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.46542489266044, + "eval_loss": 1.2418709993362427, + "eval_runtime": 2.0305, + "eval_samples_per_second": 57.129, + "step": 2080 + }, + { + "epoch": 131.0, + "learning_rate": 0.000115, + "loss": 0.0005, + "step": 2096 + }, + { + "epoch": 131.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 69.90597139032764, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7586206896551724, \"recall\": 0.5945945945945946, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7639156710022181, \"recall\": 0.7758812239659014, \"f1-score\": 0.7601502966778109, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7251200142895884, \"recall\": 0.7068965517241379, \"f1-score\": 0.7067565995623314, \"support\": 116.0}}", + "eval_f1_macro": 67.56891526024985, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.67565995623313, + "eval_loss": 1.2399200201034546, + "eval_runtime": 2.1094, + "eval_samples_per_second": 54.992, + "step": 2096 + }, + { + "epoch": 132.0, + "learning_rate": 0.00011333333333333331, + "loss": 0.0004, + "step": 2112 + }, + { + "epoch": 132.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 69.96898151137239, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5, \"recall\": 0.8333333333333334, \"f1-score\": 0.625, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.6451612903225806, \"f1-score\": 0.689655172413793, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6756756756756757, \"f1-score\": 0.6944444444444444, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7787848725348725, \"recall\": 0.7640629540831154, \"f1-score\": 0.759137703489884, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7316276475759235, \"recall\": 0.7068965517241379, \"f1-score\": 0.7101770872378342, \"support\": 116.0}}", + "eval_f1_macro": 67.47890697687858, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 71.01770872378343, + "eval_loss": 1.286320686340332, + "eval_runtime": 2.359, + "eval_samples_per_second": 49.174, + "step": 2112 + }, + { + "epoch": 133.0, + "learning_rate": 0.00011166666666666667, + "loss": 0.0001, + "step": 2128 + }, + { + "epoch": 133.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 69.96898151137239, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5, \"recall\": 0.8333333333333334, \"f1-score\": 0.625, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.6451612903225806, \"f1-score\": 0.689655172413793, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6756756756756757, \"f1-score\": 0.6944444444444444, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7787848725348725, \"recall\": 0.7640629540831154, \"f1-score\": 0.759137703489884, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7316276475759235, \"recall\": 0.7068965517241379, \"f1-score\": 0.7101770872378342, \"support\": 116.0}}", + "eval_f1_macro": 67.47890697687858, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 71.01770872378343, + "eval_loss": 1.2882046699523926, + "eval_runtime": 2.2026, + "eval_samples_per_second": 52.665, + "step": 2128 + }, + { + "epoch": 134.0, + "learning_rate": 0.00010999999999999998, + "loss": 0.0041, + "step": 2144 + }, + { + "epoch": 134.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 70.70045157638818, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.6451612903225806, \"f1-score\": 0.689655172413793, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.7027027027027027, \"f1-score\": 0.7123287671232876, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7830664097111466, \"recall\": 0.7674413324614937, \"f1-score\": 0.763893405115062, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7368814429839838, \"recall\": 0.7155172413793104, \"f1-score\": 0.7179672201946288, \"support\": 116.0}}", + "eval_f1_macro": 67.90163601022773, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.79672201946288, + "eval_loss": 1.2369537353515625, + "eval_runtime": 2.2854, + "eval_samples_per_second": 50.757, + "step": 2144 + }, + { + "epoch": 135.0, + "learning_rate": 0.00010833333333333333, + "loss": 0.0003, + "step": 2160 + }, + { + "epoch": 135.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 72.28635318822373, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.7027027027027027, \"f1-score\": 0.7222222222222223, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7942460317460318, \"recall\": 0.7853624794148988, \"f1-score\": 0.7783325085792683, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7513136288998358, \"recall\": 0.7327586206896551, \"f1-score\": 0.7340857674125114, \"support\": 116.0}}", + "eval_f1_macro": 69.18511187371273, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.40857674125114, + "eval_loss": 1.1938015222549438, + "eval_runtime": 2.5226, + "eval_samples_per_second": 45.985, + "step": 2160 + }, + { + "epoch": 136.0, + "learning_rate": 0.00010666666666666667, + "loss": 0.0002, + "step": 2176 + }, + { + "epoch": 136.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 71.45726817405861, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.7096774193548387, \"f1-score\": 0.7096774193548389, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6756756756756757, \"f1-score\": 0.7142857142857142, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7872576572173346, \"recall\": 0.7721274702121477, \"f1-score\": 0.7693289764210067, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7428814002089864, \"recall\": 0.7241379310344828, \"f1-score\": 0.7261668858524842, \"support\": 116.0}}", + "eval_f1_macro": 68.3847979040895, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.61668858524843, + "eval_loss": 1.1832021474838257, + "eval_runtime": 2.498, + "eval_samples_per_second": 46.438, + "step": 2176 + }, + { + "epoch": 137.0, + "learning_rate": 0.00010499999999999999, + "loss": 0.0023, + "step": 2192 + }, + { + "epoch": 137.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 72.24180724602901, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.7096774193548387, \"f1-score\": 0.7096774193548389, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7741935483870968, \"recall\": 0.6486486486486487, \"f1-score\": 0.7058823529411764, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7777632827324479, \"recall\": 0.7965268696115471, \"f1-score\": 0.7793320243617616, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7432814783310432, \"recall\": 0.7327586206896551, \"f1-score\": 0.7314154712513956, \"support\": 116.0}}", + "eval_f1_macro": 69.27395772104548, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.14154712513957, + "eval_loss": 1.1572390794754028, + "eval_runtime": 2.3161, + "eval_samples_per_second": 50.085, + "step": 2192 + }, + { + "epoch": 138.0, + "learning_rate": 0.00010333333333333333, + "loss": 0.0015, + "step": 2208 + }, + { + "epoch": 138.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 73.31427426365428, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.625, \"recall\": 0.8333333333333334, \"f1-score\": 0.7142857142857143, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.7096774193548387, \"f1-score\": 0.7096774193548389, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7741935483870968, \"recall\": 0.6486486486486487, \"f1-score\": 0.7058823529411764, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"request\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7896505376344086, \"recall\": 0.8138879807226581, \"f1-score\": 0.7968241732176742, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7512513904338153, \"recall\": 0.7413793103448276, \"f1-score\": 0.7415241958852505, \"support\": 116.0}}", + "eval_f1_macro": 70.8288153971266, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.15241958852505, + "eval_loss": 1.127516746520996, + "eval_runtime": 2.26, + "eval_samples_per_second": 51.328, + "step": 2208 + }, + { + "epoch": 139.0, + "learning_rate": 0.00010166666666666667, + "loss": 0.0002, + "step": 2224 + }, + { + "epoch": 139.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 71.14146434132823, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.8095238095238095, \"recall\": 0.5483870967741935, \"f1-score\": 0.6538461538461537, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7567567567567568, \"recall\": 0.7567567567567568, \"f1-score\": 0.7567567567567567, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 0.7142857142857143, \"recall\": 0.625, \"f1-score\": 0.6666666666666666, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7503485628485629, \"recall\": 0.7898790928024799, \"f1-score\": 0.7605985467827573, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7411740558292282, \"recall\": 0.7241379310344828, \"f1-score\": 0.7212951144439347, \"support\": 116.0}}", + "eval_f1_macro": 67.60875971402288, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.12951144439347, + "eval_loss": 1.218409538269043, + "eval_runtime": 2.3805, + "eval_samples_per_second": 48.73, + "step": 2224 + }, + { + "epoch": 140.0, + "learning_rate": 9.999999999999999e-05, + "loss": 0.0007, + "step": 2240 + }, + { + "epoch": 140.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 71.34715800657082, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7105263157894737, \"recall\": 0.7297297297297297, \"f1-score\": 0.7200000000000001, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7881529106858054, \"recall\": 0.7609630800154994, \"f1-score\": 0.7636808914547252, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7449958981832848, \"recall\": 0.7241379310344828, \"f1-score\": 0.7267829991230006, \"support\": 116.0}}", + "eval_f1_macro": 67.88274590708669, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.67829991230006, + "eval_loss": 1.2071099281311035, + "eval_runtime": 2.2973, + "eval_samples_per_second": 50.493, + "step": 2240 + }, + { + "epoch": 141.0, + "learning_rate": 9.833333333333333e-05, + "loss": 0.0001, + "step": 2256 + }, + { + "epoch": 141.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 69.61180365216576, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.7096774193548387, \"f1-score\": 0.7096774193548389, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6756756756756757, \"f1-score\": 0.6944444444444444, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7801102906949682, \"recall\": 0.7426135813232587, \"f1-score\": 0.7458583702298124, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7279957829526795, \"recall\": 0.7068965517241379, \"f1-score\": 0.7076938246562994, \"support\": 116.0}}", + "eval_f1_macro": 66.29852179820554, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.76938246562993, + "eval_loss": 1.2121734619140625, + "eval_runtime": 2.2118, + "eval_samples_per_second": 52.447, + "step": 2256 + }, + { + "epoch": 142.0, + "learning_rate": 9.666666666666667e-05, + "loss": 0.0008, + "step": 2272 + }, + { + "epoch": 142.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 73.08583159191204, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.7297297297297297, \"f1-score\": 0.7397260273972601, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.798931623931624, \"recall\": 0.7985974886176499, \"f1-score\": 0.7865666731447722, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7595932802829355, \"recall\": 0.7413793103448276, \"f1-score\": 0.7415042668581396, \"support\": 116.0}}", + "eval_f1_macro": 69.91703761286865, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.15042668581397, + "eval_loss": 1.2578798532485962, + "eval_runtime": 2.5066, + "eval_samples_per_second": 46.278, + "step": 2272 + }, + { + "epoch": 143.0, + "learning_rate": 9.499999999999999e-05, + "loss": 0.0003, + "step": 2288 + }, + { + "epoch": 143.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 73.08094197020597, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7714285714285715, \"recall\": 0.7297297297297297, \"f1-score\": 0.75, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7979552245999615, \"recall\": 0.7985974886176499, \"f1-score\": 0.785162747677104, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7634034522646138, \"recall\": 0.7413793103448276, \"f1-score\": 0.7425566157389356, \"support\": 116.0}}", + "eval_f1_macro": 69.79224423796481, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.25566157389356, + "eval_loss": 1.2859212160110474, + "eval_runtime": 2.1106, + "eval_samples_per_second": 54.96, + "step": 2288 + }, + { + "epoch": 144.0, + "learning_rate": 9.333333333333333e-05, + "loss": 0.0004, + "step": 2304 + }, + { + "epoch": 144.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 72.21325519082055, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.7027027027027027, \"f1-score\": 0.7323943661971832, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7762815551321743, \"recall\": 0.7952191102392715, \"f1-score\": 0.7760939115836201, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7497648933927892, \"recall\": 0.7327586206896551, \"f1-score\": 0.7331517115125162, \"support\": 116.0}}", + "eval_f1_macro": 68.98612547409955, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.31517115125162, + "eval_loss": 1.2795659303665161, + "eval_runtime": 2.3108, + "eval_samples_per_second": 50.198, + "step": 2304 + }, + { + "epoch": 145.0, + "learning_rate": 9.166666666666667e-05, + "loss": 0.0001, + "step": 2320 + }, + { + "epoch": 145.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 72.97136593324288, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.7027027027027027, \"f1-score\": 0.7323943661971832, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7810049019607843, \"recall\": 0.7992513683037876, \"f1-score\": 0.7815800267603271, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7550738111336488, \"recall\": 0.7413793103448276, \"f1-score\": 0.7413582150753241, \"support\": 116.0}}", + "eval_f1_macro": 69.47378015647352, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.13582150753241, + "eval_loss": 1.259643793106079, + "eval_runtime": 2.2374, + "eval_samples_per_second": 51.845, + "step": 2320 + }, + { + "epoch": 146.0, + "learning_rate": 8.999999999999999e-05, + "loss": 0.0143, + "step": 2336 + }, + { + "epoch": 146.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 71.74999381388027, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.7096774193548387, \"f1-score\": 0.6875, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7666666666666667, \"recall\": 0.6216216216216216, \"f1-score\": 0.6865671641791045, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.786922268907563, \"recall\": 0.7948846023442797, \"f1-score\": 0.7840857402533838, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7387870665507582, \"recall\": 0.7241379310344828, \"f1-score\": 0.7247587880387926, \"support\": 116.0}}", + "eval_f1_macro": 69.69651024474524, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.47587880387925, + "eval_loss": 1.2264900207519531, + "eval_runtime": 2.4493, + "eval_samples_per_second": 47.361, + "step": 2336 + }, + { + "epoch": 147.0, + "learning_rate": 8.833333333333333e-05, + "loss": 0.0053, + "step": 2352 + }, + { + "epoch": 147.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 71.42834861077034, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7741935483870968, \"recall\": 0.6486486486486487, \"f1-score\": 0.7058823529411764, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7736022164242041, \"recall\": 0.7826379807226581, \"f1-score\": 0.7706139927235263, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7364926852275949, \"recall\": 0.7241379310344828, \"f1-score\": 0.7238678666076022, \"support\": 116.0}}", + "eval_f1_macro": 68.49902157542456, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.38678666076022, + "eval_loss": 1.2261444330215454, + "eval_runtime": 2.3209, + "eval_samples_per_second": 49.981, + "step": 2352 + }, + { + "epoch": 148.0, + "learning_rate": 8.666666666666665e-05, + "loss": 0.0014, + "step": 2368 + }, + { + "epoch": 148.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 72.94342125396005, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7857142857142857, \"recall\": 0.7096774193548387, \"f1-score\": 0.7457627118644068, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6756756756756757, \"f1-score\": 0.7142857142857142, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7853800611153553, \"recall\": 0.7999052479899253, \"f1-score\": 0.780938224545764, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.756981946383569, \"recall\": 0.7413793103448276, \"f1-score\": 0.7408109187614014, \"support\": 116.0}}", + "eval_f1_macro": 69.41673107073457, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.08109187614014, + "eval_loss": 1.25702702999115, + "eval_runtime": 2.2587, + "eval_samples_per_second": 51.357, + "step": 2368 + }, + { + "epoch": 149.0, + "learning_rate": 8.499999999999999e-05, + "loss": 0.0008, + "step": 2384 + }, + { + "epoch": 149.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 70.75520071430205, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.625, \"recall\": 0.8333333333333334, \"f1-score\": 0.7142857142857143, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6470588235294118, \"recall\": 0.7096774193548387, \"f1-score\": 0.6769230769230768, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6486486486486487, \"f1-score\": 0.6956521739130435, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7892156862745098, \"recall\": 0.7687490918337692, \"f1-score\": 0.768426032620406, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7321881338742394, \"recall\": 0.7155172413793104, \"f1-score\": 0.7161281834842113, \"support\": 116.0}}", + "eval_f1_macro": 68.30453623292497, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.61281834842114, + "eval_loss": 1.2749736309051514, + "eval_runtime": 2.441, + "eval_samples_per_second": 47.522, + "step": 2384 + }, + { + "epoch": 150.0, + "learning_rate": 8.333333333333333e-05, + "loss": 0.0153, + "step": 2400 + }, + { + "epoch": 150.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 73.04663385761674, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6756756756756757, \"f1-score\": 0.7246376811594203, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7713840744631186, \"recall\": 0.8114979899254092, \"f1-score\": 0.7849103643666888, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7537355818764288, \"recall\": 0.7413793103448276, \"f1-score\": 0.7414086319557359, \"support\": 116.0}}", + "eval_f1_macro": 69.7698101659279, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.14086319557359, + "eval_loss": 1.1729803085327148, + "eval_runtime": 2.3845, + "eval_samples_per_second": 48.648, + "step": 2400 + }, + { + "epoch": 151.0, + "learning_rate": 8.166666666666665e-05, + "loss": 0.0002, + "step": 2416 + }, + { + "epoch": 151.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 72.2925741490733, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6756756756756757, \"f1-score\": 0.7142857142857142, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7673564180917123, \"recall\": 0.8074657318608931, \"f1-score\": 0.7808184253737793, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7439002005432026, \"recall\": 0.7327586206896551, \"f1-score\": 0.7321249020291514, \"support\": 116.0}}", + "eval_f1_macro": 69.40608225544706, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.21249020291513, + "eval_loss": 1.2129653692245483, + "eval_runtime": 2.2652, + "eval_samples_per_second": 51.211, + "step": 2416 + }, + { + "epoch": 152.0, + "learning_rate": 7.999999999999999e-05, + "loss": 0.0016, + "step": 2432 + }, + { + "epoch": 152.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 72.41728534320436, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.8214285714285714, \"recall\": 0.6216216216216216, \"f1-score\": 0.7076923076923075, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7797867063492063, \"recall\": 0.8087734912331686, \"f1-score\": 0.7859909010837803, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7506824712643678, \"recall\": 0.7327586206896551, \"f1-score\": 0.7325155936077256, \"support\": 116.0}}", + "eval_f1_macro": 69.86585787411381, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.25155936077256, + "eval_loss": 1.2629878520965576, + "eval_runtime": 2.2761, + "eval_samples_per_second": 50.964, + "step": 2432 + }, + { + "epoch": 153.0, + "learning_rate": 7.833333333333333e-05, + "loss": 0.0011, + "step": 2448 + }, + { + "epoch": 153.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 70.92028682644413, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7857142857142857, \"recall\": 0.5945945945945946, \"f1-score\": 0.676923076923077, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7759829260651628, \"recall\": 0.7932423350770125, \"f1-score\": 0.7748835617787231, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7387906728026963, \"recall\": 0.7155172413793104, \"f1-score\": 0.7169916020513906, \"support\": 116.0}}", + "eval_f1_macro": 68.87853882477538, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.69916020513907, + "eval_loss": 1.2745832204818726, + "eval_runtime": 2.4207, + "eval_samples_per_second": 47.919, + "step": 2448 + }, + { + "epoch": 154.0, + "learning_rate": 7.666666666666666e-05, + "loss": 0.012, + "step": 2464 + }, + { + "epoch": 154.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 72.41728534320436, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.8214285714285714, \"recall\": 0.6216216216216216, \"f1-score\": 0.7076923076923075, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7797867063492063, \"recall\": 0.8087734912331686, \"f1-score\": 0.7859909010837803, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7506824712643678, \"recall\": 0.7327586206896551, \"f1-score\": 0.7325155936077256, \"support\": 116.0}}", + "eval_f1_macro": 69.86585787411381, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.25155936077256, + "eval_loss": 1.2336031198501587, + "eval_runtime": 2.4363, + "eval_samples_per_second": 47.613, + "step": 2464 + }, + { + "epoch": 155.0, + "learning_rate": 7.5e-05, + "loss": 0.0001, + "step": 2480 + }, + { + "epoch": 155.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 73.16539772739387, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.7096774193548387, \"f1-score\": 0.7096774193548389, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.8275862068965517, \"recall\": 0.6486486486486487, \"f1-score\": 0.7272727272727273, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7833285882020588, \"recall\": 0.8121518696115471, \"f1-score\": 0.7898465436491005, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7585732687843271, \"recall\": 0.7413793103448276, \"f1-score\": 0.7417714718291212, \"support\": 116.0}}", + "eval_f1_macro": 70.20858165769782, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.17714718291212, + "eval_loss": 1.221439003944397, + "eval_runtime": 2.3711, + "eval_samples_per_second": 48.922, + "step": 2480 + }, + { + "epoch": 156.0, + "learning_rate": 7.333333333333332e-05, + "loss": 0.002, + "step": 2496 + }, + { + "epoch": 156.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 73.1455226720313, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.7027027027027027, \"f1-score\": 0.7323943661971832, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7829127163685987, \"recall\": 0.8108441102392715, \"f1-score\": 0.7894963693409165, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7544317329053637, \"recall\": 0.7413793103448276, \"f1-score\": 0.7412877356663379, \"support\": 116.0}}", + "eval_f1_macro": 70.17745505252591, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.12877356663378, + "eval_loss": 1.2159744501113892, + "eval_runtime": 2.149, + "eval_samples_per_second": 53.978, + "step": 2496 + }, + { + "epoch": 157.0, + "learning_rate": 7.166666666666667e-05, + "loss": 0.0002, + "step": 2512 + }, + { + "epoch": 157.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 72.30710211656402, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.7027027027027027, \"f1-score\": 0.7323943661971832, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.775469866820486, \"recall\": 0.7969552213503827, \"f1-score\": 0.7790701020598105, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7490558323389006, \"recall\": 0.7327586206896551, \"f1-score\": 0.7342600858967526, \"support\": 116.0}}", + "eval_f1_macro": 69.25067573864982, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.42600858967526, + "eval_loss": 1.2059881687164307, + "eval_runtime": 2.2612, + "eval_samples_per_second": 51.3, + "step": 2512 + }, + { + "epoch": 158.0, + "learning_rate": 7e-05, + "loss": 0.0013, + "step": 2528 + }, + { + "epoch": 158.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 73.05038784859155, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6756756756756757, \"f1-score\": 0.7246376811594203, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7758218344155844, \"recall\": 0.8114979899254092, \"f1-score\": 0.7854036742133135, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7556448879185949, \"recall\": 0.7413793103448276, \"f1-score\": 0.7411202939532842, \"support\": 116.0}}", + "eval_f1_macro": 69.81365993007232, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.11202939532842, + "eval_loss": 1.2273198366165161, + "eval_runtime": 2.2819, + "eval_samples_per_second": 50.836, + "step": 2528 + }, + { + "epoch": 159.0, + "learning_rate": 6.833333333333333e-05, + "loss": 0.0006, + "step": 2544 + }, + { + "epoch": 159.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 72.30265892598246, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.6774193548387096, \"f1-score\": 0.7, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.8, \"recall\": 0.6486486486486487, \"f1-score\": 0.7164179104477612, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7714606035726725, \"recall\": 0.8081196115470308, \"f1-score\": 0.7813589614950458, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.747290704737197, \"recall\": 0.7327586206896551, \"f1-score\": 0.7320478165532811, \"support\": 116.0}}", + "eval_f1_macro": 69.45412991067073, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.2047816553281, + "eval_loss": 1.2536399364471436, + "eval_runtime": 2.1983, + "eval_samples_per_second": 52.768, + "step": 2544 + }, + { + "epoch": 160.0, + "learning_rate": 6.666666666666666e-05, + "loss": 0.0001, + "step": 2560 + }, + { + "epoch": 160.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 71.21555209201495, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6756756756756757, \"f1-score\": 0.7142857142857142, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 1.0, \"f1-score\": 0.8571428571428571, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7539635609488551, \"recall\": 0.7918407318608931, \"f1-score\": 0.7632119090830525, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7357926471770285, \"recall\": 0.7241379310344828, \"f1-score\": 0.7219356357600302, \"support\": 116.0}}", + "eval_f1_macro": 67.84105858516021, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.19356357600302, + "eval_loss": 1.2445656061172485, + "eval_runtime": 2.2477, + "eval_samples_per_second": 51.608, + "step": 2560 + }, + { + "epoch": 161.0, + "learning_rate": 6.5e-05, + "loss": 0.0005, + "step": 2576 + }, + { + "epoch": 161.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 72.0886472797046, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.7027027027027027, \"f1-score\": 0.7323943661971832, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 1.0, \"f1-score\": 0.8571428571428571, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7756881598793364, \"recall\": 0.7952191102392715, \"f1-score\": 0.7723436224401179, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7495611639881417, \"recall\": 0.7327586206896551, \"f1-score\": 0.7315009854176577, \"support\": 116.0}}", + "eval_f1_macro": 68.65276643912159, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.15009854176577, + "eval_loss": 1.2382774353027344, + "eval_runtime": 2.1847, + "eval_samples_per_second": 53.096, + "step": 2576 + }, + { + "epoch": 162.0, + "learning_rate": 6.333333333333333e-05, + "loss": 0.0002, + "step": 2592 + }, + { + "epoch": 162.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 71.21555209201495, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6756756756756757, \"f1-score\": 0.7142857142857142, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 1.0, \"f1-score\": 0.8571428571428571, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.625, \"f1-score\": 0.7142857142857143, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7539635609488551, \"recall\": 0.7918407318608931, \"f1-score\": 0.7632119090830525, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7357926471770285, \"recall\": 0.7241379310344828, \"f1-score\": 0.7219356357600302, \"support\": 116.0}}", + "eval_f1_macro": 67.84105858516021, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.19356357600302, + "eval_loss": 1.2521488666534424, + "eval_runtime": 2.2733, + "eval_samples_per_second": 51.028, + "step": 2592 + }, + { + "epoch": 163.0, + "learning_rate": 6.166666666666666e-05, + "loss": 0.0001, + "step": 2608 + }, + { + "epoch": 163.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 71.31191410231975, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.7027027027027027, \"f1-score\": 0.7323943661971832, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 1.0, \"f1-score\": 0.8571428571428571, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7745770096776289, \"recall\": 0.7813302213503827, \"f1-score\": 0.7639457527249349, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7472701180531863, \"recall\": 0.7241379310344828, \"f1-score\": 0.7251378107127714, \"support\": 116.0}}", + "eval_f1_macro": 67.90628913110532, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.51378107127714, + "eval_loss": 1.268195629119873, + "eval_runtime": 2.3211, + "eval_samples_per_second": 49.975, + "step": 2608 + }, + { + "epoch": 164.0, + "learning_rate": 5.9999999999999995e-05, + "loss": 0.0002, + "step": 2624 + }, + { + "epoch": 164.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 70.9771975436676, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5, \"recall\": 0.8333333333333334, \"f1-score\": 0.625, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.6896551724137931, \"recall\": 0.6451612903225806, \"f1-score\": 0.6666666666666667, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7741935483870968, \"recall\": 0.6486486486486487, \"f1-score\": 0.7058823529411764, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7812765446455657, \"recall\": 0.790198464593626, \"f1-score\": 0.7759928313430635, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7379460720898525, \"recall\": 0.7155172413793104, \"f1-score\": 0.7182820133498045, \"support\": 116.0}}", + "eval_f1_macro": 68.97714056382786, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.82820133498046, + "eval_loss": 1.2899616956710815, + "eval_runtime": 2.1951, + "eval_samples_per_second": 52.844, + "step": 2624 + }, + { + "epoch": 165.0, + "learning_rate": 5.8333333333333326e-05, + "loss": 0.0004, + "step": 2640 + }, + { + "epoch": 165.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 70.83086708646859, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5, \"recall\": 0.8333333333333334, \"f1-score\": 0.625, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7931034482758621, \"recall\": 0.6216216216216216, \"f1-score\": 0.696969696969697, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7670762427227944, \"recall\": 0.7908523442797637, \"f1-score\": 0.77046813264093, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7368900273329524, \"recall\": 0.7155172413793104, \"f1-score\": 0.7173396383526295, \"support\": 116.0}}", + "eval_f1_macro": 68.48605623474933, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.73396383526295, + "eval_loss": 1.30678391456604, + "eval_runtime": 2.3185, + "eval_samples_per_second": 50.032, + "step": 2640 + }, + { + "epoch": 166.0, + "learning_rate": 5.666666666666666e-05, + "loss": 0.0005, + "step": 2656 + }, + { + "epoch": 166.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 70.83086708646859, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5, \"recall\": 0.8333333333333334, \"f1-score\": 0.625, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7931034482758621, \"recall\": 0.6216216216216216, \"f1-score\": 0.696969696969697, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7670762427227944, \"recall\": 0.7908523442797637, \"f1-score\": 0.77046813264093, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7368900273329524, \"recall\": 0.7155172413793104, \"f1-score\": 0.7173396383526295, \"support\": 116.0}}", + "eval_f1_macro": 68.48605623474933, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.73396383526295, + "eval_loss": 1.3133952617645264, + "eval_runtime": 2.377, + "eval_samples_per_second": 48.802, + "step": 2656 + }, + { + "epoch": 167.0, + "learning_rate": 5.499999999999999e-05, + "loss": 0.0002, + "step": 2672 + }, + { + "epoch": 167.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 69.96647365085389, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.47619047619047616, \"recall\": 0.8333333333333334, \"f1-score\": 0.6060606060606061, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.76, \"recall\": 0.6129032258064516, \"f1-score\": 0.6785714285714285, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6486486486486487, \"f1-score\": 0.6956521739130435, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7559848484848485, \"recall\": 0.7861662065291097, \"f1-score\": 0.7607708201857818, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7303649798477384, \"recall\": 0.7068965517241379, \"f1-score\": 0.7086251135318513, \"support\": 116.0}}", + "eval_f1_macro": 67.62407290540283, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.86251135318513, + "eval_loss": 1.300965666770935, + "eval_runtime": 2.2349, + "eval_samples_per_second": 51.904, + "step": 2672 + }, + { + "epoch": 168.0, + "learning_rate": 5.333333333333333e-05, + "loss": 0.0001, + "step": 2688 + }, + { + "epoch": 168.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 72.3130865910927, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.7027027027027027, \"f1-score\": 0.7222222222222223, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7763937451437452, \"recall\": 0.7969552213503827, \"f1-score\": 0.7804867561059512, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.745111641232331, \"recall\": 0.7327586206896551, \"f1-score\": 0.7332402168368859, \"support\": 116.0}}", + "eval_f1_macro": 69.37660054275122, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.32402168368858, + "eval_loss": 1.237136721611023, + "eval_runtime": 2.2752, + "eval_samples_per_second": 50.985, + "step": 2688 + }, + { + "epoch": 169.0, + "learning_rate": 5.1666666666666664e-05, + "loss": 0.0, + "step": 2704 + }, + { + "epoch": 169.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 73.07915281690298, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.7027027027027027, \"f1-score\": 0.7222222222222223, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7815470885323826, \"recall\": 0.8009874794148988, \"f1-score\": 0.7861582624580383, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7507764181952823, \"recall\": 0.7413793103448276, \"f1-score\": 0.7416001475793187, \"support\": 116.0}}", + "eval_f1_macro": 69.88073444071452, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.16001475793188, + "eval_loss": 1.2186721563339233, + "eval_runtime": 2.4159, + "eval_samples_per_second": 48.016, + "step": 2704 + }, + { + "epoch": 170.0, + "learning_rate": 4.9999999999999996e-05, + "loss": 0.0001, + "step": 2720 + }, + { + "epoch": 170.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 73.07915281690298, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.7027027027027027, \"f1-score\": 0.7222222222222223, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7815470885323826, \"recall\": 0.8009874794148988, \"f1-score\": 0.7861582624580383, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7507764181952823, \"recall\": 0.7413793103448276, \"f1-score\": 0.7416001475793187, \"support\": 116.0}}", + "eval_f1_macro": 69.88073444071452, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.16001475793188, + "eval_loss": 1.214310646057129, + "eval_runtime": 2.3883, + "eval_samples_per_second": 48.57, + "step": 2720 + }, + { + "epoch": 171.0, + "learning_rate": 4.8333333333333334e-05, + "loss": 0.0001, + "step": 2736 + }, + { + "epoch": 171.0, + "eval_accuracy": 75.0, + "eval_average_metrics": 73.91158889784157, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.7027027027027027, \"f1-score\": 0.7323943661971832, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.75, \"macro avg\": {\"precision\": 0.7880660597572362, \"recall\": 0.8148763683037876, \"f1-score\": 0.7951678756930036, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7600965098683151, \"recall\": 0.75, \"f1-score\": 0.7496476664087707, \"support\": 116.0}}", + "eval_f1_macro": 70.68158895048921, + "eval_f1_micro": 75.0, + "eval_f1_weighted": 74.96476664087707, + "eval_loss": 1.2283321619033813, + "eval_runtime": 2.3482, + "eval_samples_per_second": 49.4, + "step": 2736 + }, + { + "epoch": 172.0, + "learning_rate": 4.6666666666666665e-05, + "loss": 0.0118, + "step": 2752 + }, + { + "epoch": 172.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 73.06239436585838, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6756756756756757, \"f1-score\": 0.7142857142857142, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7769475214328155, \"recall\": 0.8114979899254092, \"f1-score\": 0.7869832415724911, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.75147428354832, \"recall\": 0.7413793103448276, \"f1-score\": 0.7401964947691324, \"support\": 116.0}}", + "eval_f1_macro": 69.95406591755476, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.01964947691323, + "eval_loss": 1.2305833101272583, + "eval_runtime": 2.3308, + "eval_samples_per_second": 49.768, + "step": 2752 + }, + { + "epoch": 173.0, + "learning_rate": 4.4999999999999996e-05, + "loss": 0.0001, + "step": 2768 + }, + { + "epoch": 173.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 73.06239436585838, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6756756756756757, \"f1-score\": 0.7142857142857142, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7769475214328155, \"recall\": 0.8114979899254092, \"f1-score\": 0.7869832415724911, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.75147428354832, \"recall\": 0.7413793103448276, \"f1-score\": 0.7401964947691324, \"support\": 116.0}}", + "eval_f1_macro": 69.95406591755476, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.01964947691323, + "eval_loss": 1.232358694076538, + "eval_runtime": 2.353, + "eval_samples_per_second": 49.298, + "step": 2768 + }, + { + "epoch": 174.0, + "learning_rate": 4.333333333333333e-05, + "loss": 0.0003, + "step": 2784 + }, + { + "epoch": 174.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 73.06239436585838, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6756756756756757, \"f1-score\": 0.7142857142857142, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7769475214328155, \"recall\": 0.8114979899254092, \"f1-score\": 0.7869832415724911, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.75147428354832, \"recall\": 0.7413793103448276, \"f1-score\": 0.7401964947691324, \"support\": 116.0}}", + "eval_f1_macro": 69.95406591755476, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.01964947691323, + "eval_loss": 1.2332383394241333, + "eval_runtime": 2.4493, + "eval_samples_per_second": 47.361, + "step": 2784 + }, + { + "epoch": 175.0, + "learning_rate": 4.1666666666666665e-05, + "loss": 0.0001, + "step": 2800 + }, + { + "epoch": 175.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 73.06239436585838, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6756756756756757, \"f1-score\": 0.7142857142857142, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7769475214328155, \"recall\": 0.8114979899254092, \"f1-score\": 0.7869832415724911, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.75147428354832, \"recall\": 0.7413793103448276, \"f1-score\": 0.7401964947691324, \"support\": 116.0}}", + "eval_f1_macro": 69.95406591755476, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.01964947691323, + "eval_loss": 1.2342748641967773, + "eval_runtime": 2.2579, + "eval_samples_per_second": 51.375, + "step": 2800 + }, + { + "epoch": 176.0, + "learning_rate": 3.9999999999999996e-05, + "loss": 0.0001, + "step": 2816 + }, + { + "epoch": 176.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 72.31492066595455, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6486486486486487, \"f1-score\": 0.6956521739130435, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7725283295136236, \"recall\": 0.8081196115470308, \"f1-score\": 0.7831198584940545, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7416345064671637, \"recall\": 0.7327586206896551, \"f1-score\": 0.7309730443752678, \"support\": 116.0}}", + "eval_f1_macro": 69.6106540883604, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.09730443752677, + "eval_loss": 1.210294485092163, + "eval_runtime": 2.3919, + "eval_samples_per_second": 48.497, + "step": 2816 + }, + { + "epoch": 177.0, + "learning_rate": 3.833333333333333e-05, + "loss": 0.0002, + "step": 2832 + }, + { + "epoch": 177.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 72.31269149223965, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.6774193548387096, \"f1-score\": 0.7, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7741935483870968, \"recall\": 0.6486486486486487, \"f1-score\": 0.7058823529411764, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7723197644413211, \"recall\": 0.8081196115470308, \"f1-score\": 0.7829155800251135, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7424399991601941, \"recall\": 0.7327586206896551, \"f1-score\": 0.7310654582879524, \"support\": 116.0}}", + "eval_f1_macro": 69.59249600223232, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.10654582879525, + "eval_loss": 1.207425832748413, + "eval_runtime": 2.1302, + "eval_samples_per_second": 54.455, + "step": 2832 + }, + { + "epoch": 178.0, + "learning_rate": 3.666666666666666e-05, + "loss": 0.0, + "step": 2848 + }, + { + "epoch": 178.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 72.31269149223965, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.6774193548387096, \"f1-score\": 0.7, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7741935483870968, \"recall\": 0.6486486486486487, \"f1-score\": 0.7058823529411764, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7723197644413211, \"recall\": 0.8081196115470308, \"f1-score\": 0.7829155800251135, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7424399991601941, \"recall\": 0.7327586206896551, \"f1-score\": 0.7310654582879524, \"support\": 116.0}}", + "eval_f1_macro": 69.59249600223232, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.10654582879525, + "eval_loss": 1.2198625802993774, + "eval_runtime": 2.2148, + "eval_samples_per_second": 52.376, + "step": 2848 + }, + { + "epoch": 179.0, + "learning_rate": 3.5e-05, + "loss": 0.0002, + "step": 2864 + }, + { + "epoch": 179.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 73.05760487940339, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6756756756756757, \"f1-score\": 0.7246376811594203, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7764345795136236, \"recall\": 0.8114979899254092, \"f1-score\": 0.7867430468998516, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7516021788809568, \"recall\": 0.7413793103448276, \"f1-score\": 0.7402184216866121, \"support\": 116.0}}", + "eval_f1_macro": 69.93271527998681, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.02184216866121, + "eval_loss": 1.2318569421768188, + "eval_runtime": 2.3945, + "eval_samples_per_second": 48.445, + "step": 2864 + }, + { + "epoch": 180.0, + "learning_rate": 3.333333333333333e-05, + "loss": 0.0001, + "step": 2880 + }, + { + "epoch": 180.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 72.9164008835973, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.7027027027027027, \"f1-score\": 0.7323943661971832, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 1.0, \"f1-score\": 0.8571428571428571, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7871732026143791, \"recall\": 0.7992513683037876, \"f1-score\": 0.780043526358128, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7583107955826008, \"recall\": 0.7413793103448276, \"f1-score\": 0.7405253912247894, \"support\": 116.0}}", + "eval_f1_macro": 69.33720234294472, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.05253912247895, + "eval_loss": 1.238344669342041, + "eval_runtime": 2.1836, + "eval_samples_per_second": 53.123, + "step": 2880 + }, + { + "epoch": 181.0, + "learning_rate": 3.1666666666666666e-05, + "loss": 0.0, + "step": 2896 + }, + { + "epoch": 181.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 71.54716438209535, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.6774193548387096, \"f1-score\": 0.7, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7931034482758621, \"recall\": 0.6216216216216216, \"f1-score\": 0.696969696969697, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7669435638469213, \"recall\": 0.8047412331686525, \"f1-score\": 0.776239762767277, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7420661391615456, \"recall\": 0.7241379310344828, \"f1-score\": 0.723619812977269, \"support\": 116.0}}", + "eval_f1_macro": 68.99909002375796, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.3619812977269, + "eval_loss": 1.264186978340149, + "eval_runtime": 2.4029, + "eval_samples_per_second": 48.275, + "step": 2896 + }, + { + "epoch": 182.0, + "learning_rate": 2.9999999999999997e-05, + "loss": 0.0015, + "step": 2912 + }, + { + "epoch": 182.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 70.66688997087125, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7931034482758621, \"recall\": 0.6216216216216216, \"f1-score\": 0.696969696969697, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 1.0, \"f1-score\": 0.8571428571428571, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7732607380520267, \"recall\": 0.7891162331686525, \"f1-score\": 0.7656020398344721, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7401777332749233, \"recall\": 0.7155172413793104, \"f1-score\": 0.7151059695566986, \"support\": 116.0}}", + "eval_f1_macro": 68.05351465195308, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.51059695566985, + "eval_loss": 1.265453815460205, + "eval_runtime": 2.3141, + "eval_samples_per_second": 50.127, + "step": 2912 + }, + { + "epoch": 183.0, + "learning_rate": 2.833333333333333e-05, + "loss": 0.0057, + "step": 2928 + }, + { + "epoch": 183.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 70.757218513014, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7666666666666667, \"recall\": 0.6216216216216216, \"f1-score\": 0.6865671641791045, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7784788676236045, \"recall\": 0.7891162331686525, \"f1-score\": 0.7696588660927909, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7370352802067865, \"recall\": 0.7155172413793104, \"f1-score\": 0.7151130434572363, \"support\": 116.0}}", + "eval_f1_macro": 68.4141214304703, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.51130434572363, + "eval_loss": 1.2580338716506958, + "eval_runtime": 2.4483, + "eval_samples_per_second": 47.38, + "step": 2928 + }, + { + "epoch": 184.0, + "learning_rate": 2.6666666666666667e-05, + "loss": 0.0017, + "step": 2944 + }, + { + "epoch": 184.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 70.757218513014, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7666666666666667, \"recall\": 0.6216216216216216, \"f1-score\": 0.6865671641791045, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7784788676236045, \"recall\": 0.7891162331686525, \"f1-score\": 0.7696588660927909, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7370352802067865, \"recall\": 0.7155172413793104, \"f1-score\": 0.7151130434572363, \"support\": 116.0}}", + "eval_f1_macro": 68.4141214304703, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.51130434572363, + "eval_loss": 1.2513511180877686, + "eval_runtime": 2.2687, + "eval_samples_per_second": 51.13, + "step": 2944 + }, + { + "epoch": 185.0, + "learning_rate": 2.4999999999999998e-05, + "loss": 0.0001, + "step": 2960 + }, + { + "epoch": 185.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 71.50327381251103, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5, \"recall\": 0.8333333333333334, \"f1-score\": 0.625, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6756756756756757, \"f1-score\": 0.7142857142857142, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7827068764568765, \"recall\": 0.7918407318608931, \"f1-score\": 0.7722572480409168, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7499145968973555, \"recall\": 0.7241379310344828, \"f1-score\": 0.7254042032839941, \"support\": 116.0}}", + "eval_f1_macro": 68.64508871474816, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.5404203283994, + "eval_loss": 1.290483832359314, + "eval_runtime": 2.2629, + "eval_samples_per_second": 51.262, + "step": 2960 + }, + { + "epoch": 186.0, + "learning_rate": 2.3333333333333332e-05, + "loss": 0.0001, + "step": 2976 + }, + { + "epoch": 186.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 71.50327381251103, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5, \"recall\": 0.8333333333333334, \"f1-score\": 0.625, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6756756756756757, \"f1-score\": 0.7142857142857142, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7827068764568765, \"recall\": 0.7918407318608931, \"f1-score\": 0.7722572480409168, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7499145968973555, \"recall\": 0.7241379310344828, \"f1-score\": 0.7254042032839941, \"support\": 116.0}}", + "eval_f1_macro": 68.64508871474816, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.5404203283994, + "eval_loss": 1.298017978668213, + "eval_runtime": 2.339, + "eval_samples_per_second": 49.594, + "step": 2976 + }, + { + "epoch": 187.0, + "learning_rate": 2.1666666666666664e-05, + "loss": 0.0066, + "step": 2992 + }, + { + "epoch": 187.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 72.35246834449421, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5, \"recall\": 0.8333333333333334, \"f1-score\": 0.625, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.7027027027027027, \"f1-score\": 0.7323943661971832, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7938254147812971, \"recall\": 0.7952191102392715, \"f1-score\": 0.7804418821614294, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7585368232173506, \"recall\": 0.7327586206896551, \"f1-score\": 0.7348553749236323, \"support\": 116.0}}", + "eval_f1_macro": 69.37261174768261, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.48553749236324, + "eval_loss": 1.2962392568588257, + "eval_runtime": 2.4107, + "eval_samples_per_second": 48.118, + "step": 2992 + }, + { + "epoch": 188.0, + "learning_rate": 1.9999999999999998e-05, + "loss": 0.0005, + "step": 3008 + }, + { + "epoch": 188.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 72.35124316220256, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.7027027027027027, \"f1-score\": 0.7222222222222223, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7943837960285329, \"recall\": 0.7952191102392715, \"f1-score\": 0.7816905254548818, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.754290151772003, \"recall\": 0.7327586206896551, \"f1-score\": 0.7336964624822306, \"support\": 116.0}}", + "eval_f1_macro": 69.48360226265616, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.36964624822306, + "eval_loss": 1.28173828125, + "eval_runtime": 2.3316, + "eval_samples_per_second": 49.751, + "step": 3008 + }, + { + "epoch": 189.0, + "learning_rate": 1.833333333333333e-05, + "loss": 0.0001, + "step": 3024 + }, + { + "epoch": 189.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 73.08583159191204, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.7297297297297297, \"f1-score\": 0.7397260273972601, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.798931623931624, \"recall\": 0.7985974886176499, \"f1-score\": 0.7865666731447722, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7595932802829355, \"recall\": 0.7413793103448276, \"f1-score\": 0.7415042668581396, \"support\": 116.0}}", + "eval_f1_macro": 69.91703761286865, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.15042668581397, + "eval_loss": 1.2783206701278687, + "eval_runtime": 2.2566, + "eval_samples_per_second": 51.406, + "step": 3024 + }, + { + "epoch": 190.0, + "learning_rate": 1.6666666666666664e-05, + "loss": 0.001, + "step": 3040 + }, + { + "epoch": 190.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 72.25528619564864, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7297297297297297, \"recall\": 0.7297297297297297, \"f1-score\": 0.7297297297297297, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7926099613599613, \"recall\": 0.7847085997287611, \"f1-score\": 0.7775790406982358, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7507766658628727, \"recall\": 0.7327586206896551, \"f1-score\": 0.7335128369370923, \"support\": 116.0}}", + "eval_f1_macro": 69.1181369509543, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.35128369370923, + "eval_loss": 1.278748631477356, + "eval_runtime": 2.3035, + "eval_samples_per_second": 50.357, + "step": 3040 + }, + { + "epoch": 191.0, + "learning_rate": 1.4999999999999999e-05, + "loss": 0.0001, + "step": 3056 + }, + { + "epoch": 191.0, + "eval_accuracy": 75.0, + "eval_average_metrics": 73.71814517325353, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.8333333333333334, \"recall\": 0.6451612903225806, \"f1-score\": 0.7272727272727272, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7435897435897436, \"recall\": 0.7837837837837838, \"f1-score\": 0.7631578947368421, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.75, \"macro avg\": {\"precision\": 0.8023552836052836, \"recall\": 0.7914653564855179, \"f1-score\": 0.7849473539876017, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7723283900008039, \"recall\": 0.75, \"f1-score\": 0.7509948256078287, \"support\": 116.0}}", + "eval_f1_macro": 69.77309813223127, + "eval_f1_micro": 75.0, + "eval_f1_weighted": 75.09948256078287, + "eval_loss": 1.2824381589889526, + "eval_runtime": 2.3428, + "eval_samples_per_second": 49.513, + "step": 3056 + }, + { + "epoch": 192.0, + "learning_rate": 1.3333333333333333e-05, + "loss": 0.0003, + "step": 3072 + }, + { + "epoch": 192.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 72.98695667868029, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.8, \"recall\": 0.6451612903225806, \"f1-score\": 0.7142857142857142, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7368421052631579, \"recall\": 0.7567567567567568, \"f1-score\": 0.7466666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7973451621477937, \"recall\": 0.7880869781071395, \"f1-score\": 0.7812625738554531, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7612680800747951, \"recall\": 0.7413793103448276, \"f1-score\": 0.742264025252709, \"support\": 116.0}}", + "eval_f1_macro": 69.44556212048472, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.22640252527088, + "eval_loss": 1.2781001329421997, + "eval_runtime": 2.2829, + "eval_samples_per_second": 50.812, + "step": 3072 + }, + { + "epoch": 193.0, + "learning_rate": 1.1666666666666666e-05, + "loss": 0.0006, + "step": 3088 + }, + { + "epoch": 193.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 72.25528619564864, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7297297297297297, \"recall\": 0.7297297297297297, \"f1-score\": 0.7297297297297297, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7926099613599613, \"recall\": 0.7847085997287611, \"f1-score\": 0.7775790406982358, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7507766658628727, \"recall\": 0.7327586206896551, \"f1-score\": 0.7335128369370923, \"support\": 116.0}}", + "eval_f1_macro": 69.1181369509543, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.35128369370923, + "eval_loss": 1.2707769870758057, + "eval_runtime": 2.3425, + "eval_samples_per_second": 49.519, + "step": 3088 + }, + { + "epoch": 194.0, + "learning_rate": 9.999999999999999e-06, + "loss": 0.0001, + "step": 3104 + }, + { + "epoch": 194.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 72.25528619564864, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7297297297297297, \"recall\": 0.7297297297297297, \"f1-score\": 0.7297297297297297, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7926099613599613, \"recall\": 0.7847085997287611, \"f1-score\": 0.7775790406982358, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7507766658628727, \"recall\": 0.7327586206896551, \"f1-score\": 0.7335128369370923, \"support\": 116.0}}", + "eval_f1_macro": 69.1181369509543, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.35128369370923, + "eval_loss": 1.267704725265503, + "eval_runtime": 2.4427, + "eval_samples_per_second": 47.488, + "step": 3104 + }, + { + "epoch": 195.0, + "learning_rate": 8.333333333333332e-06, + "loss": 0.0001, + "step": 3120 + }, + { + "epoch": 195.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 72.25528619564864, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7297297297297297, \"recall\": 0.7297297297297297, \"f1-score\": 0.7297297297297297, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7926099613599613, \"recall\": 0.7847085997287611, \"f1-score\": 0.7775790406982358, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7507766658628727, \"recall\": 0.7327586206896551, \"f1-score\": 0.7335128369370923, \"support\": 116.0}}", + "eval_f1_macro": 69.1181369509543, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.35128369370923, + "eval_loss": 1.2668423652648926, + "eval_runtime": 2.2674, + "eval_samples_per_second": 51.16, + "step": 3120 + }, + { + "epoch": 196.0, + "learning_rate": 6.666666666666667e-06, + "loss": 0.0001, + "step": 3136 + }, + { + "epoch": 196.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 71.52003226355563, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5, \"recall\": 0.8333333333333334, \"f1-score\": 0.625, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.7027027027027027, \"f1-score\": 0.7222222222222223, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7873064435564436, \"recall\": 0.7813302213503827, \"f1-score\": 0.7714322689264641, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7492167315443178, \"recall\": 0.7241379310344828, \"f1-score\": 0.7268078560941804, \"support\": 116.0}}", + "eval_f1_macro": 68.57175723790792, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.68078560941804, + "eval_loss": 1.2636795043945312, + "eval_runtime": 2.3362, + "eval_samples_per_second": 49.653, + "step": 3136 + }, + { + "epoch": 197.0, + "learning_rate": 4.9999999999999996e-06, + "loss": 0.0001, + "step": 3152 + }, + { + "epoch": 197.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 71.52003226355563, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5, \"recall\": 0.8333333333333334, \"f1-score\": 0.625, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.7027027027027027, \"f1-score\": 0.7222222222222223, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7873064435564436, \"recall\": 0.7813302213503827, \"f1-score\": 0.7714322689264641, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7492167315443178, \"recall\": 0.7241379310344828, \"f1-score\": 0.7268078560941804, \"support\": 116.0}}", + "eval_f1_macro": 68.57175723790792, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.68078560941804, + "eval_loss": 1.2627345323562622, + "eval_runtime": 2.4056, + "eval_samples_per_second": 48.222, + "step": 3152 + }, + { + "epoch": 198.0, + "learning_rate": 3.3333333333333333e-06, + "loss": 0.0001, + "step": 3168 + }, + { + "epoch": 198.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 70.70675091356331, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5, \"recall\": 0.8333333333333334, \"f1-score\": 0.625, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.7027027027027027, \"f1-score\": 0.7123287671232876, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7833381895881897, \"recall\": 0.7674413324614937, \"f1-score\": 0.7628856455186293, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7417728344452482, \"recall\": 0.7155172413793104, \"f1-score\": 0.7191149799895746, \"support\": 116.0}}", + "eval_f1_macro": 67.81205737943371, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.91149799895746, + "eval_loss": 1.2624974250793457, + "eval_runtime": 2.3963, + "eval_samples_per_second": 48.408, + "step": 3168 + }, + { + "epoch": 199.0, + "learning_rate": 1.6666666666666667e-06, + "loss": 0.0, + "step": 3184 + }, + { + "epoch": 199.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 70.70675091356331, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5, \"recall\": 0.8333333333333334, \"f1-score\": 0.625, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.7027027027027027, \"f1-score\": 0.7123287671232876, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7833381895881897, \"recall\": 0.7674413324614937, \"f1-score\": 0.7628856455186293, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7417728344452482, \"recall\": 0.7155172413793104, \"f1-score\": 0.7191149799895746, \"support\": 116.0}}", + "eval_f1_macro": 67.81205737943371, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.91149799895746, + "eval_loss": 1.262513279914856, + "eval_runtime": 2.3048, + "eval_samples_per_second": 50.331, + "step": 3184 + }, + { + "epoch": 200.0, + "learning_rate": 0.0, + "loss": 0.0009, + "step": 3200 + }, + { + "epoch": 200.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 70.70675091356331, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accusation\": {\"precision\": 0.5, \"recall\": 0.8333333333333334, \"f1-score\": 0.625, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.7027027027027027, \"f1-score\": 0.7123287671232876, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.625, \"f1-score\": 0.7692307692307693, \"support\": 8.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7833381895881897, \"recall\": 0.7674413324614937, \"f1-score\": 0.7628856455186293, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7417728344452482, \"recall\": 0.7155172413793104, \"f1-score\": 0.7191149799895746, \"support\": 116.0}}", + "eval_f1_macro": 67.81205737943371, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.91149799895746, + "eval_loss": 1.2624309062957764, + "eval_runtime": 2.4394, + "eval_samples_per_second": 47.554, + "step": 3200 + }, + { + "epoch": 200.0, + "step": 3200, + "total_flos": 1.565785344442368e+16, + "train_runtime": 2488.3254, + "train_samples_per_second": 1.286 + } + ], + "max_steps": 3200, + "num_train_epochs": 200, + "total_flos": 1.565785344442368e+16, + "trial_name": null, + "trial_params": null +}