diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,18708 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.9990118577075099, + "eval_steps": 106, + "global_step": 337, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0, + "step": 0, + "train_accuracy_first_token_arguments": 0.7777777777777778, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 0.8, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_order": 1.0, + "train_accuracy_recipient_token_order_total": 1, + "train_accuracy_recipient_token_set": 0.0, + "train_accuracy_recipient_token_set_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0, + "step": 0, + "train_accuracy_first_token_arguments": 0.6129032258064516, + "train_accuracy_first_token_arguments_total": 31, + "train_accuracy_recipient_token": 0.6111111111111112, + "train_accuracy_recipient_token_all": 0.6, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.3333333333333333, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_identify": 0.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0, + "step": 0, + "train_accuracy_first_token_arguments": 0.5555555555555556, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 0.6842105263157895, + "train_accuracy_recipient_token_all": 0.7, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.5, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_histor": 0.5, + "train_accuracy_recipient_token_histor_total": 2, + "train_accuracy_recipient_token_med": 1.0, + "train_accuracy_recipient_token_med_total": 2, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.0029644268774703555, + "grad_norm": 3.2711954063798103, + "learning_rate": 4.5454545454545457e-07, + "loss": 0.3754, + "step": 1 + }, + { + "epoch": 0.0029644268774703555, + "step": 1, + "train_accuracy_first_token_arguments": 0.5245901639344263, + "train_accuracy_first_token_arguments_total": 61, + "train_accuracy_recipient_token": 0.6, + "train_accuracy_recipient_token_all": 0.6, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_estimate": 0.5, + "train_accuracy_recipient_token_estimate_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_prior": 0.0, + "train_accuracy_recipient_token_prior_total": 1, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.0029644268774703555, + "step": 1, + "train_accuracy_first_token_arguments": 0.8076923076923077, + "train_accuracy_first_token_arguments_total": 26, + "train_accuracy_recipient_token": 0.6666666666666666, + "train_accuracy_recipient_token_all": 0.6, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_search": 0.5, + "train_accuracy_recipient_token_search_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.0029644268774703555, + "step": 1, + "train_accuracy_first_token_arguments": 0.8181818181818182, + "train_accuracy_first_token_arguments_total": 22, + "train_accuracy_recipient_token": 0.7894736842105263, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 0.75, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_search": 0.0, + "train_accuracy_recipient_token_search_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.005928853754940711, + "grad_norm": 2.9342358559874504, + "learning_rate": 9.090909090909091e-07, + "loss": 0.2983, + "step": 2 + }, + { + "epoch": 0.005928853754940711, + "step": 2, + "train_accuracy_first_token_arguments": 0.9166666666666666, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.7368421052631579, + "train_accuracy_recipient_token_all": 0.7, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_c": 0.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_get": 0.875, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.005928853754940711, + "step": 2, + "train_accuracy_first_token_arguments": 0.6666666666666666, + "train_accuracy_first_token_arguments_total": 24, + "train_accuracy_recipient_token": 0.8, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_art": 1.0, + "train_accuracy_recipient_token_art_total": 2, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_interpret": 0.6666666666666666, + "train_accuracy_recipient_token_interpret_total": 3, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.005928853754940711, + "step": 2, + "train_accuracy_first_token_arguments": 0.8387096774193549, + "train_accuracy_first_token_arguments_total": 31, + "train_accuracy_recipient_token": 0.7368421052631579, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_check": 0.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.008893280632411068, + "grad_norm": 3.7666597300364524, + "learning_rate": 1.3636363636363636e-06, + "loss": 0.3456, + "step": 3 + }, + { + "epoch": 0.008893280632411068, + "step": 3, + "train_accuracy_first_token_arguments": 0.7241379310344828, + "train_accuracy_first_token_arguments_total": 29, + "train_accuracy_recipient_token": 0.6666666666666666, + "train_accuracy_recipient_token_all": 0.6, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_find": 0.6666666666666666, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 0.5, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.008893280632411068, + "step": 3, + "train_accuracy_first_token_arguments": 0.7333333333333333, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.631578947368421, + "train_accuracy_recipient_token_all": 0.6, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_list": 0.0, + "train_accuracy_recipient_token_list_total": 1, + "train_accuracy_recipient_token_vendor": 1.0, + "train_accuracy_recipient_token_vendor_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.008893280632411068, + "step": 3, + "train_accuracy_first_token_arguments": 0.7058823529411765, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.55, + "train_accuracy_recipient_token_all": 0.7, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_count": 0.0, + "train_accuracy_recipient_token_count_total": 1, + "train_accuracy_recipient_token_get": 0.5714285714285714, + "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_identify": 0.0, + "train_accuracy_recipient_token_identify_total": 2, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.011857707509881422, + "grad_norm": 4.455480512503962, + "learning_rate": 1.8181818181818183e-06, + "loss": 0.3511, + "step": 4 + }, + { + "epoch": 0.011857707509881422, + "step": 4, + "train_accuracy_first_token_arguments": 0.975609756097561, + "train_accuracy_first_token_arguments_total": 41, + "train_accuracy_recipient_token": 0.6, + "train_accuracy_recipient_token_all": 0.5, + "train_accuracy_recipient_token_all_total": 8, + "train_accuracy_recipient_token_d": 0.0, + "train_accuracy_recipient_token_d_total": 1, + "train_accuracy_recipient_token_generate": 1.0, + "train_accuracy_recipient_token_generate_total": 1, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_order": 1.0, + "train_accuracy_recipient_token_order_total": 1, + "train_total_number_recipient_token": 15 + }, + { + "epoch": 0.011857707509881422, + "step": 4, + "train_accuracy_first_token_arguments": 0.8125, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 0.6, + "train_accuracy_recipient_token_all": 0.7, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 0.5, + "train_accuracy_recipient_token_analyze_total": 2, + "train_accuracy_recipient_token_get": 0.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_locate": 0.0, + "train_accuracy_recipient_token_locate_total": 1, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 15 + }, + { + "epoch": 0.011857707509881422, + "step": 4, + "train_accuracy_first_token_arguments": 0.75, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 0.5625, + "train_accuracy_recipient_token_all": 0.6, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 0.0, + "train_accuracy_recipient_token_analyze_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_rate": 1.0, + "train_accuracy_recipient_token_rate_total": 1, + "train_accuracy_recipient_token_recommend": 0.0, + "train_accuracy_recipient_token_recommend_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.014822134387351778, + "grad_norm": 4.251122308361701, + "learning_rate": 2.2727272727272728e-06, + "loss": 0.5231, + "step": 5 + }, + { + "epoch": 0.014822134387351778, + "step": 5, + "train_accuracy_first_token_arguments": 0.7391304347826086, + "train_accuracy_first_token_arguments_total": 23, + "train_accuracy_recipient_token": 0.6875, + "train_accuracy_recipient_token_all": 0.6, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_fruit": 1.0, + "train_accuracy_recipient_token_fruit_total": 3, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_select": 1.0, + "train_accuracy_recipient_token_select_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.014822134387351778, + "step": 5, + "train_accuracy_first_token_arguments": 0.8888888888888888, + "train_accuracy_first_token_arguments_total": 9, + "train_accuracy_recipient_token": 0.6111111111111112, + "train_accuracy_recipient_token_all": 0.6, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.7142857142857143, + "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_identify": 0.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.014822134387351778, + "step": 5, + "train_accuracy_first_token_arguments": 0.6216216216216216, + "train_accuracy_first_token_arguments_total": 37, + "train_accuracy_recipient_token": 0.7894736842105263, + "train_accuracy_recipient_token_all": 0.7, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compose": 0.0, + "train_accuracy_recipient_token_compose_total": 1, + "train_accuracy_recipient_token_create": 1.0, + "train_accuracy_recipient_token_create_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_g": 1.0, + "train_accuracy_recipient_token_g_total": 4, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.017786561264822136, + "grad_norm": 2.9047536525561006, + "learning_rate": 2.7272727272727272e-06, + "loss": 0.3063, + "step": 6 + }, + { + "epoch": 0.017786561264822136, + "step": 6, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.7222222222222222, + "train_accuracy_recipient_token_all": 0.6363636363636364, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 0.8333333333333334, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.017786561264822136, + "step": 6, + "train_accuracy_first_token_arguments": 0.8648648648648649, + "train_accuracy_first_token_arguments_total": 37, + "train_accuracy_recipient_token": 0.5294117647058824, + "train_accuracy_recipient_token_all": 0.6363636363636364, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_get": 0.3333333333333333, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_identify": 0.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_validate": 0.0, + "train_accuracy_recipient_token_validate_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.017786561264822136, + "step": 6, + "train_accuracy_first_token_arguments": 0.76, + "train_accuracy_first_token_arguments_total": 25, + "train_accuracy_recipient_token": 0.75, + "train_accuracy_recipient_token_all": 0.8333333333333334, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_evaluate": 0.0, + "train_accuracy_recipient_token_evaluate_total": 1, + "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.020750988142292492, + "grad_norm": 3.2977375008302783, + "learning_rate": 3.181818181818182e-06, + "loss": 0.4802, + "step": 7 + }, + { + "epoch": 0.020750988142292492, + "step": 7, + "train_accuracy_first_token_arguments": 0.9090909090909091, + "train_accuracy_first_token_arguments_total": 11, + "train_accuracy_recipient_token": 0.7777777777777778, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.5714285714285714, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.020750988142292492, + "step": 7, + "train_accuracy_first_token_arguments": 0.7692307692307693, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.75, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.5, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.020750988142292492, + "step": 7, + "train_accuracy_first_token_arguments": 0.8421052631578947, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.7368421052631579, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.5555555555555556, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.023715415019762844, + "grad_norm": 2.6597485903906652, + "learning_rate": 3.6363636363636366e-06, + "loss": 0.4234, + "step": 8 + }, + { + "epoch": 0.023715415019762844, + "step": 8, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 0.6666666666666666, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 0.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_count": 0.5, + "train_accuracy_recipient_token_count_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_identify": 0.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.023715415019762844, + "step": 8, + "train_accuracy_first_token_arguments": 0.9, + "train_accuracy_first_token_arguments_total": 10, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.023715415019762844, + "step": 8, + "train_accuracy_first_token_arguments": 0.8867924528301887, + "train_accuracy_first_token_arguments_total": 53, + "train_accuracy_recipient_token": 0.7368421052631579, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_describe": 1.0, + "train_accuracy_recipient_token_describe_total": 2, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.5, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_sum": 0.0, + "train_accuracy_recipient_token_sum_total": 2, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.0266798418972332, + "grad_norm": 2.552647042706623, + "learning_rate": 4.0909090909090915e-06, + "loss": 0.3278, + "step": 9 + }, + { + "epoch": 0.0266798418972332, + "step": 9, + "train_accuracy_first_token_arguments": 0.7894736842105263, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.7777777777777778, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_prior": 1.0, + "train_accuracy_recipient_token_prior_total": 1, + "train_accuracy_recipient_token_schedule": 0.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.0266798418972332, + "step": 9, + "train_accuracy_first_token_arguments": 0.9333333333333333, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.5, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_estimate": 0.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_load": 0.0, + "train_accuracy_recipient_token_load_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.0266798418972332, + "step": 9, + "train_accuracy_first_token_arguments": 0.7142857142857143, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_property": 1.0, + "train_accuracy_recipient_token_property_total": 3, + "train_accuracy_recipient_token_real": 1.0, + "train_accuracy_recipient_token_real_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.029644268774703556, + "grad_norm": 2.3528382096065767, + "learning_rate": 4.5454545454545455e-06, + "loss": 0.2897, + "step": 10 + }, + { + "epoch": 0.029644268774703556, + "step": 10, + "train_accuracy_first_token_arguments": 0.76, + "train_accuracy_first_token_arguments_total": 25, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_assemble": 1.0, + "train_accuracy_recipient_token_assemble_total": 1, + "train_accuracy_recipient_token_clean": 1.0, + "train_accuracy_recipient_token_clean_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_repair": 0.5, + "train_accuracy_recipient_token_repair_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.029644268774703556, + "step": 10, + "train_accuracy_first_token_arguments": 0.8260869565217391, + "train_accuracy_first_token_arguments_total": 23, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 4, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 1, + "train_accuracy_recipient_token_search": 0.0, + "train_accuracy_recipient_token_search_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.029644268774703556, + "step": 10, + "train_accuracy_first_token_arguments": 0.6333333333333333, + "train_accuracy_first_token_arguments_total": 30, + "train_accuracy_recipient_token": 0.6666666666666666, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.5, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_suggest": 0.5, + "train_accuracy_recipient_token_suggest_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.03260869565217391, + "grad_norm": 2.351257384059913, + "learning_rate": 5e-06, + "loss": 0.2987, + "step": 11 + }, + { + "epoch": 0.03260869565217391, + "step": 11, + "train_accuracy_first_token_arguments": 0.5106382978723404, + "train_accuracy_first_token_arguments_total": 47, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_add": 1.0, + "train_accuracy_recipient_token_add_total": 5, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.03260869565217391, + "step": 11, + "train_accuracy_first_token_arguments": 0.9130434782608695, + "train_accuracy_first_token_arguments_total": 23, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_h": 1.0, + "train_accuracy_recipient_token_h_total": 3, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.03260869565217391, + "step": 11, + "train_accuracy_first_token_arguments": 0.7142857142857143, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_fabric": 1.0, + "train_accuracy_recipient_token_fabric_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.03557312252964427, + "grad_norm": 2.266360756598947, + "learning_rate": 4.999883916312e-06, + "loss": 0.3119, + "step": 12 + }, + { + "epoch": 0.03557312252964427, + "step": 12, + "train_accuracy_first_token_arguments": 0.696969696969697, + "train_accuracy_first_token_arguments_total": 33, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 5, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_total_number_recipient_token": 10 + }, + { + "epoch": 0.03557312252964427, + "step": 12, + "train_accuracy_first_token_arguments": 0.8048780487804879, + "train_accuracy_first_token_arguments_total": 41, + "train_accuracy_recipient_token": 0.7058823529411765, + "train_accuracy_recipient_token_all": 0.7, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_create": 1.0, + "train_accuracy_recipient_token_create_total": 2, + "train_accuracy_recipient_token_explain": 0.0, + "train_accuracy_recipient_token_explain_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_translate": 0.0, + "train_accuracy_recipient_token_translate_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.03557312252964427, + "step": 12, + "train_accuracy_first_token_arguments": 0.8333333333333334, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.038537549407114624, + "grad_norm": 10.867702658903724, + "learning_rate": 4.999535676028338e-06, + "loss": 0.4711, + "step": 13 + }, + { + "epoch": 0.038537549407114624, + "step": 13, + "train_accuracy_first_token_arguments": 0.9333333333333333, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.7058823529411765, + "train_accuracy_recipient_token_all": 0.7, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.8333333333333334, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_suggest": 0.0, + "train_accuracy_recipient_token_suggest_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.038537549407114624, + "step": 13, + "train_accuracy_first_token_arguments": 0.875, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_verify": 1.0, + "train_accuracy_recipient_token_verify_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.038537549407114624, + "step": 13, + "train_accuracy_first_token_arguments": 0.9285714285714286, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_generate": 1.0, + "train_accuracy_recipient_token_generate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_track": 1.0, + "train_accuracy_recipient_token_track_total": 1, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.041501976284584984, + "grad_norm": 2.8220096523893234, + "learning_rate": 4.998955311489025e-06, + "loss": 0.3082, + "step": 14 + }, + { + "epoch": 0.041501976284584984, + "step": 14, + "train_accuracy_first_token_arguments": 0.8636363636363636, + "train_accuracy_first_token_arguments_total": 22, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.7, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.041501976284584984, + "step": 14, + "train_accuracy_first_token_arguments": 0.7777777777777778, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_buy": 1.0, + "train_accuracy_recipient_token_buy_total": 1, + "train_accuracy_recipient_token_exp": 1.0, + "train_accuracy_recipient_token_exp_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.041501976284584984, + "step": 14, + "train_accuracy_first_token_arguments": 0.8285714285714286, + "train_accuracy_first_token_arguments_total": 35, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_filter": 1.0, + "train_accuracy_recipient_token_filter_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_locate": 1.0, + "train_accuracy_recipient_token_locate_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.044466403162055336, + "grad_norm": 2.258107289871962, + "learning_rate": 4.998142876590749e-06, + "loss": 0.3097, + "step": 15 + }, + { + "epoch": 0.044466403162055336, + "step": 15, + "train_accuracy_first_token_arguments": 0.75, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_clean": 1.0, + "train_accuracy_recipient_token_clean_total": 1, + "train_accuracy_recipient_token_generate": 1.0, + "train_accuracy_recipient_token_generate_total": 5, + "train_accuracy_recipient_token_remove": 1.0, + "train_accuracy_recipient_token_remove_total": 1, + "train_accuracy_recipient_token_replace": 1.0, + "train_accuracy_recipient_token_replace_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.044466403162055336, + "step": 15, + "train_accuracy_first_token_arguments": 0.75, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 0.7058823529411765, + "train_accuracy_recipient_token_all": 0.6, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_identify": 0.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_register": 1.0, + "train_accuracy_recipient_token_register_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.044466403162055336, + "step": 15, + "train_accuracy_first_token_arguments": 0.9545454545454546, + "train_accuracy_first_token_arguments_total": 22, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_schedule": 0.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_accuracy_recipient_token_suggest": 1.0, + "train_accuracy_recipient_token_suggest_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.04743083003952569, + "grad_norm": 3.2749980475323635, + "learning_rate": 4.997098446781861e-06, + "loss": 0.2419, + "step": 16 + }, + { + "epoch": 0.04743083003952569, + "step": 16, + "train_accuracy_first_token_arguments": 0.85, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 5, + "train_accuracy_recipient_token_get": 0.5, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.04743083003952569, + "step": 16, + "train_accuracy_first_token_arguments": 0.7096774193548387, + "train_accuracy_first_token_arguments_total": 62, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 3, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.04743083003952569, + "step": 16, + "train_accuracy_first_token_arguments": 0.9285714285714286, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_histor": 1.0, + "train_accuracy_recipient_token_histor_total": 1, + "train_accuracy_recipient_token_math": 1.0, + "train_accuracy_recipient_token_math_total": 1, + "train_accuracy_recipient_token_symbol": 1.0, + "train_accuracy_recipient_token_symbol_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.05039525691699605, + "grad_norm": 2.3223843516425937, + "learning_rate": 4.9958221190553705e-06, + "loss": 0.2414, + "step": 17 + }, + { + "epoch": 0.05039525691699605, + "step": 17, + "train_accuracy_first_token_arguments": 0.9583333333333334, + "train_accuracy_first_token_arguments_total": 24, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.7272727272727273, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.05039525691699605, + "step": 17, + "train_accuracy_first_token_arguments": 0.92, + "train_accuracy_first_token_arguments_total": 25, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.05039525691699605, + "step": 17, + "train_accuracy_first_token_arguments": 0.8571428571428571, + "train_accuracy_first_token_arguments_total": 21, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_nutrition": 1.0, + "train_accuracy_recipient_token_nutrition_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.0533596837944664, + "grad_norm": 2.7061439489759453, + "learning_rate": 4.994314011939941e-06, + "loss": 0.2925, + "step": 18 + }, + { + "epoch": 0.0533596837944664, + "step": 18, + "train_accuracy_first_token_arguments": 0.9666666666666667, + "train_accuracy_first_token_arguments_total": 30, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_building": 1.0, + "train_accuracy_recipient_token_building_total": 1, + "train_accuracy_recipient_token_find": 0.75, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_property": 1.0, + "train_accuracy_recipient_token_property_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.0533596837944664, + "step": 18, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.0533596837944664, + "step": 18, + "train_accuracy_first_token_arguments": 0.6428571428571429, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_decode": 0.0, + "train_accuracy_recipient_token_decode_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.05632411067193676, + "grad_norm": 2.8315352410694907, + "learning_rate": 4.992574265488883e-06, + "loss": 0.3402, + "step": 19 + }, + { + "epoch": 0.05632411067193676, + "step": 19, + "train_accuracy_first_token_arguments": 0.7916666666666666, + "train_accuracy_first_token_arguments_total": 24, + "train_accuracy_recipient_token": 0.7894736842105263, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_plan": 0.0, + "train_accuracy_recipient_token_plan_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.05632411067193676, + "step": 19, + "train_accuracy_first_token_arguments": 0.8, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.05632411067193676, + "step": 19, + "train_accuracy_first_token_arguments": 0.8571428571428571, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.85, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_traffic": 1.0, + "train_accuracy_recipient_token_traffic_total": 2, + "train_accuracy_recipient_token_verify": 1.0, + "train_accuracy_recipient_token_verify_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.05928853754940711, + "grad_norm": 3.725703665798811, + "learning_rate": 4.990603041267144e-06, + "loss": 0.2884, + "step": 20 + }, + { + "epoch": 0.05928853754940711, + "step": 20, + "train_accuracy_first_token_arguments": 0.7241379310344828, + "train_accuracy_first_token_arguments_total": 29, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.8, + "train_accuracy_recipient_token_find_total": 5, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_order": 1.0, + "train_accuracy_recipient_token_order_total": 1, + "train_accuracy_recipient_token_suggest": 1.0, + "train_accuracy_recipient_token_suggest_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.05928853754940711, + "step": 20, + "train_accuracy_first_token_arguments": 0.8, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.8888888888888888, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.05928853754940711, + "step": 20, + "train_accuracy_first_token_arguments": 0.8, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.5625, + "train_accuracy_recipient_token_all": 0.6363636363636364, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_deploy": 1.0, + "train_accuracy_recipient_token_deploy_total": 1, + "train_accuracy_recipient_token_identify": 0.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_install": 1.0, + "train_accuracy_recipient_token_install_total": 1, + "train_accuracy_recipient_token_schedule": 0.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.06225296442687747, + "grad_norm": 2.4295313275927413, + "learning_rate": 4.988400522336304e-06, + "loss": 0.2966, + "step": 21 + }, + { + "epoch": 0.06225296442687747, + "step": 21, + "train_accuracy_first_token_arguments": 0.9473684210526315, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.7647058823529411, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_order": 0.0, + "train_accuracy_recipient_token_order_total": 1, + "train_accuracy_recipient_token_prior": 1.0, + "train_accuracy_recipient_token_prior_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.06225296442687747, + "step": 21, + "train_accuracy_first_token_arguments": 0.8461538461538461, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.7647058823529411, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.7142857142857143, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.06225296442687747, + "step": 21, + "train_accuracy_first_token_arguments": 0.9, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_translate": 0.0, + "train_accuracy_recipient_token_translate_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.06521739130434782, + "grad_norm": 3.0399009498719938, + "learning_rate": 4.985966913237581e-06, + "loss": 0.3759, + "step": 22 + }, + { + "epoch": 0.06521739130434782, + "step": 22, + "train_accuracy_first_token_arguments": 0.9230769230769231, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.7777777777777778, + "train_accuracy_recipient_token_all": 0.7272727272727273, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_estimate": 0.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_style": 1.0, + "train_accuracy_recipient_token_style_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.06521739130434782, + "step": 22, + "train_accuracy_first_token_arguments": 0.5, + "train_accuracy_first_token_arguments_total": 6, + "train_accuracy_recipient_token": 0.7333333333333333, + "train_accuracy_recipient_token_all": 0.7, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_c": 0.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 15 + }, + { + "epoch": 0.06521739130434782, + "step": 22, + "train_accuracy_first_token_arguments": 0.7804878048780488, + "train_accuracy_first_token_arguments_total": 41, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_count": 1.0, + "train_accuracy_recipient_token_count_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_list": 0.0, + "train_accuracy_recipient_token_list_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.06818181818181818, + "grad_norm": 2.939285198394914, + "learning_rate": 4.9833024399728295e-06, + "loss": 0.4389, + "step": 23 + }, + { + "epoch": 0.06818181818181818, + "step": 23, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.06818181818181818, + "step": 23, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.5, + "train_accuracy_recipient_token_all": 0.6363636363636364, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 0.25, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_track": 0.0, + "train_accuracy_recipient_token_track_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.06818181818181818, + "step": 23, + "train_accuracy_first_token_arguments": 0.9375, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 0.8125, + "train_accuracy_recipient_token_all": 0.7, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.07114624505928854, + "grad_norm": 3.2811864983396806, + "learning_rate": 4.980407349983556e-06, + "loss": 0.2645, + "step": 24 + }, + { + "epoch": 0.07114624505928854, + "step": 24, + "train_accuracy_first_token_arguments": 0.9166666666666666, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 5, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.07114624505928854, + "step": 24, + "train_accuracy_first_token_arguments": 0.9016393442622951, + "train_accuracy_first_token_arguments_total": 61, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.07114624505928854, + "step": 24, + "train_accuracy_first_token_arguments": 0.76, + "train_accuracy_first_token_arguments_total": 25, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_list": 1.0, + "train_accuracy_recipient_token_list_total": 1, + "train_accuracy_recipient_token_order": 1.0, + "train_accuracy_recipient_token_order_total": 1, + "train_accuracy_recipient_token_rate": 1.0, + "train_accuracy_recipient_token_rate_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.0741106719367589, + "grad_norm": 1.8226124894214228, + "learning_rate": 4.9772819121279395e-06, + "loss": 0.2212, + "step": 25 + }, + { + "epoch": 0.0741106719367589, + "step": 25, + "train_accuracy_first_token_arguments": 0.9193548387096774, + "train_accuracy_first_token_arguments_total": 62, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_count": 1.0, + "train_accuracy_recipient_token_count_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.0741106719367589, + "step": 25, + "train_accuracy_first_token_arguments": 0.76, + "train_accuracy_first_token_arguments_total": 25, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.0741106719367589, + "step": 25, + "train_accuracy_first_token_arguments": 0.6666666666666666, + "train_accuracy_first_token_arguments_total": 24, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_plan": 0.0, + "train_accuracy_recipient_token_plan_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.07707509881422925, + "grad_norm": 1.524883365948892, + "learning_rate": 4.973926416655863e-06, + "loss": 0.1674, + "step": 26 + }, + { + "epoch": 0.07707509881422925, + "step": 26, + "train_accuracy_first_token_arguments": 0.8333333333333334, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.7777777777777778, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_animal": 0.6666666666666666, + "train_accuracy_recipient_token_animal_total": 3, + "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_im": 1.0, + "train_accuracy_recipient_token_im_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.07707509881422925, + "step": 26, + "train_accuracy_first_token_arguments": 0.8260869565217391, + "train_accuracy_first_token_arguments_total": 23, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_plan": 0.0, + "train_accuracy_recipient_token_plan_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.07707509881422925, + "step": 26, + "train_accuracy_first_token_arguments": 0.75, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 0.9047619047619048, + "train_accuracy_recipient_token_all": 0.9166666666666666, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_get": 0.875, + "train_accuracy_recipient_token_get_total": 8, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 21 + }, + { + "epoch": 0.0800395256916996, + "grad_norm": 2.0177323408527643, + "learning_rate": 4.970341175181957e-06, + "loss": 0.2729, + "step": 27 + }, + { + "epoch": 0.0800395256916996, + "step": 27, + "train_accuracy_first_token_arguments": 0.8888888888888888, + "train_accuracy_first_token_arguments_total": 9, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_list": 1.0, + "train_accuracy_recipient_token_list_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.0800395256916996, + "step": 27, + "train_accuracy_first_token_arguments": 0.8787878787878788, + "train_accuracy_first_token_arguments_total": 33, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_draw": 1.0, + "train_accuracy_recipient_token_draw_total": 5, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.0800395256916996, + "step": 27, + "train_accuracy_first_token_arguments": 0.7692307692307693, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.85, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_get": 0.875, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.08300395256916997, + "grad_norm": 2.0988150229289273, + "learning_rate": 4.966526520656663e-06, + "loss": 0.3187, + "step": 28 + }, + { + "epoch": 0.08300395256916997, + "step": 28, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.7692307692307693, + "train_accuracy_recipient_token_all_total": 13, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.08300395256916997, + "step": 28, + "train_accuracy_first_token_arguments": 0.7857142857142857, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.08300395256916997, + "step": 28, + "train_accuracy_first_token_arguments": 0.7241379310344828, + "train_accuracy_first_token_arguments_total": 29, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.7272727272727273, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_count": 1.0, + "train_accuracy_recipient_token_count_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_rec": 1.0, + "train_accuracy_recipient_token_rec_total": 1, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.08596837944664032, + "grad_norm": 2.2452954917456447, + "learning_rate": 4.9624828073353144e-06, + "loss": 0.2444, + "step": 29 + }, + { + "epoch": 0.08596837944664032, + "step": 29, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 24, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.08596837944664032, + "step": 29, + "train_accuracy_first_token_arguments": 0.9, + "train_accuracy_first_token_arguments_total": 10, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.08596837944664032, + "step": 29, + "train_accuracy_first_token_arguments": 0.75, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.8, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_im": 1.0, + "train_accuracy_recipient_token_im_total": 1, + "train_accuracy_recipient_token_install": 0.0, + "train_accuracy_recipient_token_install_total": 1, + "train_total_number_recipient_token": 15 + }, + { + "epoch": 0.08893280632411067, + "grad_norm": 2.1813680120346604, + "learning_rate": 4.958210410745237e-06, + "loss": 0.2787, + "step": 30 + }, + { + "epoch": 0.08893280632411067, + "step": 30, + "train_accuracy_first_token_arguments": 0.9259259259259259, + "train_accuracy_first_token_arguments_total": 54, + "train_accuracy_recipient_token": 0.8125, + "train_accuracy_recipient_token_all": 0.7, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_event": 1.0, + "train_accuracy_recipient_token_event_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.08893280632411067, + "step": 30, + "train_accuracy_first_token_arguments": 0.8235294117647058, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.7777777777777778, + "train_accuracy_recipient_token_all": 0.6666666666666666, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.08893280632411067, + "step": 30, + "train_accuracy_first_token_arguments": 0.88, + "train_accuracy_first_token_arguments_total": 25, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 6, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_hire": 1.0, + "train_accuracy_recipient_token_hire_total": 1, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.09189723320158102, + "grad_norm": 1.9535807669288685, + "learning_rate": 4.95370972765087e-06, + "loss": 0.384, + "step": 31 + }, + { + "epoch": 0.09189723320158102, + "step": 31, + "train_accuracy_first_token_arguments": 0.6666666666666666, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_convert": 1.0, + "train_accuracy_recipient_token_convert_total": 5, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.09189723320158102, + "step": 31, + "train_accuracy_first_token_arguments": 0.75, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_evaluate": 1.0, + "train_accuracy_recipient_token_evaluate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_stat": 1.0, + "train_accuracy_recipient_token_stat_total": 2, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.09189723320158102, + "step": 31, + "train_accuracy_first_token_arguments": 0.7368421052631579, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_filter": 0.8, + "train_accuracy_recipient_token_filter_total": 5, + "train_accuracy_recipient_token_food": 1.0, + "train_accuracy_recipient_token_food_total": 4, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.09486166007905138, + "grad_norm": 2.5184069445357213, + "learning_rate": 4.94898117601693e-06, + "loss": 0.255, + "step": 32 + }, + { + "epoch": 0.09486166007905138, + "step": 32, + "train_accuracy_first_token_arguments": 0.7272727272727273, + "train_accuracy_first_token_arguments_total": 22, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.09486166007905138, + "step": 32, + "train_accuracy_first_token_arguments": 0.8461538461538461, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_schedule": 0.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.09486166007905138, + "step": 32, + "train_accuracy_first_token_arguments": 0.9444444444444444, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_suggest": 1.0, + "train_accuracy_recipient_token_suggest_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.09782608695652174, + "grad_norm": 1.7872388531938566, + "learning_rate": 4.944025194969586e-06, + "loss": 0.241, + "step": 33 + }, + { + "epoch": 0.09782608695652174, + "step": 33, + "train_accuracy_first_token_arguments": 0.6428571428571429, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.09782608695652174, + "step": 33, + "train_accuracy_first_token_arguments": 0.9528301886792453, + "train_accuracy_first_token_arguments_total": 106, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_assert": 1.0, + "train_accuracy_recipient_token_assert_total": 5, + "train_accuracy_recipient_token_describe": 1.0, + "train_accuracy_recipient_token_describe_total": 1, + "train_accuracy_recipient_token_mount": 1.0, + "train_accuracy_recipient_token_mount_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.09782608695652174, + "step": 33, + "train_accuracy_first_token_arguments": 0.9811320754716981, + "train_accuracy_first_token_arguments_total": 53, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9166666666666666, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_get": 0.8571428571428571, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.1007905138339921, + "grad_norm": 1.826928371292299, + "learning_rate": 4.938842244755683e-06, + "loss": 0.3343, + "step": 34 + }, + { + "epoch": 0.1007905138339921, + "step": 34, + "train_accuracy_first_token_arguments": 0.75, + "train_accuracy_first_token_arguments_total": 8, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.8571428571428571, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.1007905138339921, + "step": 34, + "train_accuracy_first_token_arguments": 0.8888888888888888, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.8888888888888888, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.1007905138339921, + "step": 34, + "train_accuracy_first_token_arguments": 0.9333333333333333, + "train_accuracy_first_token_arguments_total": 30, + "train_accuracy_recipient_token": 0.7894736842105263, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_compare": 0.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.10375494071146245, + "grad_norm": 2.586283846482814, + "learning_rate": 4.933432806700004e-06, + "loss": 0.2806, + "step": 35 + }, + { + "epoch": 0.10375494071146245, + "step": 35, + "train_accuracy_first_token_arguments": 0.6956521739130435, + "train_accuracy_first_token_arguments_total": 23, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_validate": 0.0, + "train_accuracy_recipient_token_validate_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.10375494071146245, + "step": 35, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 22, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 10, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.10375494071146245, + "step": 35, + "train_accuracy_first_token_arguments": 0.6666666666666666, + "train_accuracy_first_token_arguments_total": 33, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_art": 0.6666666666666666, + "train_accuracy_recipient_token_art_total": 3, + "train_accuracy_recipient_token_inspect": 1.0, + "train_accuracy_recipient_token_inspect_total": 1, + "train_accuracy_recipient_token_repair": 1.0, + "train_accuracy_recipient_token_repair_total": 1, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.1067193675889328, + "grad_norm": 2.28772413603275, + "learning_rate": 4.927797383160561e-06, + "loss": 0.3219, + "step": 36 + }, + { + "epoch": 0.1067193675889328, + "step": 36, + "train_accuracy_first_token_arguments": 0.9090909090909091, + "train_accuracy_first_token_arguments_total": 11, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.1067193675889328, + "step": 36, + "train_accuracy_first_token_arguments": 0.8, + "train_accuracy_first_token_arguments_total": 25, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.1067193675889328, + "step": 36, + "train_accuracy_first_token_arguments": 0.6333333333333333, + "train_accuracy_first_token_arguments_total": 30, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 0.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.10968379446640317, + "grad_norm": 1.7442627302155183, + "learning_rate": 4.921936497481956e-06, + "loss": 0.3006, + "step": 37 + }, + { + "epoch": 0.10968379446640317, + "step": 37, + "train_accuracy_first_token_arguments": 0.8, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.10968379446640317, + "step": 37, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_track": 1.0, + "train_accuracy_recipient_token_track_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.10968379446640317, + "step": 37, + "train_accuracy_first_token_arguments": 0.9166666666666666, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 21 + }, + { + "epoch": 0.11264822134387352, + "grad_norm": 2.0180662009823, + "learning_rate": 4.915850693946766e-06, + "loss": 0.266, + "step": 38 + }, + { + "epoch": 0.11264822134387352, + "step": 38, + "train_accuracy_first_token_arguments": 0.8387096774193549, + "train_accuracy_first_token_arguments_total": 31, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_log": 1.0, + "train_accuracy_recipient_token_log_total": 1, + "train_accuracy_recipient_token_plant": 1.0, + "train_accuracy_recipient_token_plant_total": 3, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.11264822134387352, + "step": 38, + "train_accuracy_first_token_arguments": 0.6274509803921569, + "train_accuracy_first_token_arguments_total": 51, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_plan": 1.0, + "train_accuracy_recipient_token_plan_total": 1, + "train_accuracy_recipient_token_urban": 1.0, + "train_accuracy_recipient_token_urban_total": 3, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.11264822134387352, + "step": 38, + "train_accuracy_first_token_arguments": 0.6862745098039216, + "train_accuracy_first_token_arguments_total": 51, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_upgrade": 1.0, + "train_accuracy_recipient_token_upgrade_total": 2, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.11561264822134387, + "grad_norm": 2.0013783198828907, + "learning_rate": 4.909540537725007e-06, + "loss": 0.3129, + "step": 39 + }, + { + "epoch": 0.11561264822134387, + "step": 39, + "train_accuracy_first_token_arguments": 0.8888888888888888, + "train_accuracy_first_token_arguments_total": 9, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.11561264822134387, + "step": 39, + "train_accuracy_first_token_arguments": 0.9375, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 2, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.11561264822134387, + "step": 39, + "train_accuracy_first_token_arguments": 0.7560975609756098, + "train_accuracy_first_token_arguments_total": 41, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_track": 1.0, + "train_accuracy_recipient_token_track_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.11857707509881422, + "grad_norm": 2.383877337768723, + "learning_rate": 4.903006614821645e-06, + "loss": 0.2901, + "step": 40 + }, + { + "epoch": 0.11857707509881422, + "step": 40, + "train_accuracy_first_token_arguments": 0.7666666666666667, + "train_accuracy_first_token_arguments_total": 30, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 2, + "train_accuracy_recipient_token_translate": 0.5, + "train_accuracy_recipient_token_translate_total": 2, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.11857707509881422, + "step": 40, + "train_accuracy_first_token_arguments": 0.8571428571428571, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 10, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.11857707509881422, + "step": 40, + "train_accuracy_first_token_arguments": 0.9183673469387755, + "train_accuracy_first_token_arguments_total": 49, + "train_accuracy_recipient_token": 0.7, + "train_accuracy_recipient_token_all": 0.6666666666666666, + "train_accuracy_recipient_token_all_total": 15, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_setup": 0.0, + "train_accuracy_recipient_token_setup_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.12154150197628459, + "grad_norm": 2.3845477671070725, + "learning_rate": 4.8962495320221714e-06, + "loss": 0.2275, + "step": 41 + }, + { + "epoch": 0.12154150197628459, + "step": 41, + "train_accuracy_first_token_arguments": 0.65, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.12154150197628459, + "step": 41, + "train_accuracy_first_token_arguments": 0.8888888888888888, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_decode": 1.0, + "train_accuracy_recipient_token_decode_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.12154150197628459, + "step": 41, + "train_accuracy_first_token_arguments": 0.5769230769230769, + "train_accuracy_first_token_arguments_total": 52, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_list": 1.0, + "train_accuracy_recipient_token_list_total": 1, + "train_accuracy_recipient_token_merge": 1.0, + "train_accuracy_recipient_token_merge_total": 1, + "train_accuracy_recipient_token_recommend": 0.75, + "train_accuracy_recipient_token_recommend_total": 4, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.12450592885375494, + "grad_norm": 2.2451847636665194, + "learning_rate": 4.8892699168362626e-06, + "loss": 0.2985, + "step": 42 + }, + { + "epoch": 0.12450592885375494, + "step": 42, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compile": 1.0, + "train_accuracy_recipient_token_compile_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.12450592885375494, + "step": 42, + "train_accuracy_first_token_arguments": 0.8, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.9375, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 2, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.12450592885375494, + "step": 42, + "train_accuracy_first_token_arguments": 0.8823529411764706, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.1274703557312253, + "grad_norm": 2.1597197573051443, + "learning_rate": 4.8820684174394935e-06, + "loss": 0.3257, + "step": 43 + }, + { + "epoch": 0.1274703557312253, + "step": 43, + "train_accuracy_first_token_arguments": 0.6428571428571429, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.9375, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.1274703557312253, + "step": 43, + "train_accuracy_first_token_arguments": 0.75, + "train_accuracy_first_token_arguments_total": 8, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.8461538461538461, + "train_accuracy_recipient_token_all_total": 13, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.1274703557312253, + "step": 43, + "train_accuracy_first_token_arguments": 0.8333333333333334, + "train_accuracy_first_token_arguments_total": 6, + "train_accuracy_recipient_token": 0.8666666666666667, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 15 + }, + { + "epoch": 0.13043478260869565, + "grad_norm": 2.4009249785875033, + "learning_rate": 4.874645702613152e-06, + "loss": 0.4131, + "step": 44 + }, + { + "epoch": 0.13043478260869565, + "step": 44, + "train_accuracy_first_token_arguments": 0.7727272727272727, + "train_accuracy_first_token_arguments_total": 22, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.13043478260869565, + "step": 44, + "train_accuracy_first_token_arguments": 0.8095238095238095, + "train_accuracy_first_token_arguments_total": 21, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_artifact": 1.0, + "train_accuracy_recipient_token_artifact_total": 3, + "train_accuracy_recipient_token_get": 0.8571428571428571, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.13043478260869565, + "step": 44, + "train_accuracy_first_token_arguments": 0.6, + "train_accuracy_first_token_arguments_total": 35, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_build": 1.0, + "train_accuracy_recipient_token_build_total": 1, + "train_accuracy_recipient_token_calculate": 0.75, + "train_accuracy_recipient_token_calculate_total": 4, + "train_accuracy_recipient_token_fire": 1.0, + "train_accuracy_recipient_token_fire_total": 3, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.13339920948616601, + "grad_norm": 2.2178791639658186, + "learning_rate": 4.867002461682129e-06, + "loss": 0.401, + "step": 45 + }, + { + "epoch": 0.13339920948616601, + "step": 45, + "train_accuracy_first_token_arguments": 0.8235294117647058, + "train_accuracy_first_token_arguments_total": 34, + "train_accuracy_recipient_token": 0.7368421052631579, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 0.5, + "train_accuracy_recipient_token_compare_total": 2, + "train_accuracy_recipient_token_find": 0.4, + "train_accuracy_recipient_token_find_total": 5, + "train_accuracy_recipient_token_get": 0.5, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.13339920948616601, + "step": 45, + "train_accuracy_first_token_arguments": 0.9333333333333333, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_ex": 0.3333333333333333, + "train_accuracy_recipient_token_ex_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.13339920948616601, + "step": 45, + "train_accuracy_first_token_arguments": 0.8709677419354839, + "train_accuracy_first_token_arguments_total": 31, + "train_accuracy_recipient_token": 0.85, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.6666666666666666, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_log": 1.0, + "train_accuracy_recipient_token_log_total": 1, + "train_accuracy_recipient_token_review": 0.5, + "train_accuracy_recipient_token_review_total": 2, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.13636363636363635, + "grad_norm": 1.7428907811945011, + "learning_rate": 4.8591394044508985e-06, + "loss": 0.2065, + "step": 46 + }, + { + "epoch": 0.13636363636363635, + "step": 46, + "train_accuracy_first_token_arguments": 0.9, + "train_accuracy_first_token_arguments_total": 200, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_filter": 0.75, + "train_accuracy_recipient_token_filter_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.13636363636363635, + "step": 46, + "train_accuracy_first_token_arguments": 0.8235294117647058, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.85, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_order": 1.0, + "train_accuracy_recipient_token_order_total": 1, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.13636363636363635, + "step": 46, + "train_accuracy_first_token_arguments": 0.9259259259259259, + "train_accuracy_first_token_arguments_total": 27, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_city": 0.75, + "train_accuracy_recipient_token_city_total": 4, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.13932806324110672, + "grad_norm": 2.3699963492078684, + "learning_rate": 4.851057261137608e-06, + "loss": 0.2718, + "step": 47 + }, + { + "epoch": 0.13932806324110672, + "step": 47, + "train_accuracy_first_token_arguments": 0.8, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.8571428571428571, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.13932806324110672, + "step": 47, + "train_accuracy_first_token_arguments": 0.59375, + "train_accuracy_first_token_arguments_total": 32, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 2, + "train_accuracy_recipient_token_detect": 1.0, + "train_accuracy_recipient_token_detect_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 2, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.13932806324110672, + "step": 47, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 22, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.8333333333333334, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_calculate": 0.5, + "train_accuracy_recipient_token_calculate_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.1422924901185771, + "grad_norm": 2.553448245564224, + "learning_rate": 4.842756782306261e-06, + "loss": 0.3375, + "step": 48 + }, + { + "epoch": 0.1422924901185771, + "step": 48, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.6470588235294118, + "train_accuracy_recipient_token_all": 0.7272727272727273, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_organ": 0.0, + "train_accuracy_recipient_token_organ_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.1422924901185771, + "step": 48, + "train_accuracy_first_token_arguments": 0.5384615384615384, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_art": 1.0, + "train_accuracy_recipient_token_art_total": 2, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_d": 1.0, + "train_accuracy_recipient_token_d_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.1422924901185771, + "step": 48, + "train_accuracy_first_token_arguments": 0.7894736842105263, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_c": 0.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 2, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.14525691699604742, + "grad_norm": 2.047932501451832, + "learning_rate": 4.8342387387970105e-06, + "loss": 0.2978, + "step": 49 + }, + { + "epoch": 0.14525691699604742, + "step": 49, + "train_accuracy_first_token_arguments": 0.9090909090909091, + "train_accuracy_first_token_arguments_total": 33, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.875, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.14525691699604742, + "step": 49, + "train_accuracy_first_token_arguments": 0.85, + "train_accuracy_first_token_arguments_total": 40, + "train_accuracy_recipient_token": 0.875, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 8, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_setup": 1.0, + "train_accuracy_recipient_token_setup_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.14525691699604742, + "step": 49, + "train_accuracy_first_token_arguments": 0.8181818181818182, + "train_accuracy_first_token_arguments_total": 55, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_count": 1.0, + "train_accuracy_recipient_token_count_total": 1, + "train_accuracy_recipient_token_g": 0.6666666666666666, + "train_accuracy_recipient_token_g_total": 3, + "train_accuracy_recipient_token_track": 1.0, + "train_accuracy_recipient_token_track_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.1482213438735178, + "grad_norm": 2.251584762112145, + "learning_rate": 4.825503921654582e-06, + "loss": 0.1679, + "step": 50 + }, + { + "epoch": 0.1482213438735178, + "step": 50, + "train_accuracy_first_token_arguments": 0.9090909090909091, + "train_accuracy_first_token_arguments_total": 11, + "train_accuracy_recipient_token": 0.8125, + "train_accuracy_recipient_token_all": 0.7, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_report": 1.0, + "train_accuracy_recipient_token_report_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.1482213438735178, + "step": 50, + "train_accuracy_first_token_arguments": 0.6923076923076923, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.875, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.1482213438735178, + "step": 50, + "train_accuracy_first_token_arguments": 0.8, + "train_accuracy_first_token_arguments_total": 10, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 0.5, + "train_accuracy_recipient_token_analyze_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_predict": 1.0, + "train_accuracy_recipient_token_predict_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.15118577075098813, + "grad_norm": 2.326983001481526, + "learning_rate": 4.816553142054806e-06, + "loss": 0.2279, + "step": 51 + }, + { + "epoch": 0.15118577075098813, + "step": 51, + "train_accuracy_first_token_arguments": 0.7692307692307693, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_plan": 1.0, + "train_accuracy_recipient_token_plan_total": 1, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 3, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.15118577075098813, + "step": 51, + "train_accuracy_first_token_arguments": 0.875, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 0.875, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_max": 1.0, + "train_accuracy_recipient_token_max_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.15118577075098813, + "step": 51, + "train_accuracy_first_token_arguments": 0.8421052631578947, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 4, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.1541501976284585, + "grad_norm": 2.104029483073964, + "learning_rate": 4.807387231229287e-06, + "loss": 0.3356, + "step": 52 + }, + { + "epoch": 0.1541501976284585, + "step": 52, + "train_accuracy_first_token_arguments": 0.7, + "train_accuracy_first_token_arguments_total": 10, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_list": 1.0, + "train_accuracy_recipient_token_list_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.1541501976284585, + "step": 52, + "train_accuracy_first_token_arguments": 0.8620689655172413, + "train_accuracy_first_token_arguments_total": 29, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_evaluate": 1.0, + "train_accuracy_recipient_token_evaluate_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_im": 1.0, + "train_accuracy_recipient_token_im_total": 1, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.1541501976284585, + "step": 52, + "train_accuracy_first_token_arguments": 0.782608695652174, + "train_accuracy_first_token_arguments_total": 23, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_display": 1.0, + "train_accuracy_recipient_token_display_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_h": 1.0, + "train_accuracy_recipient_token_h_total": 3, + "train_accuracy_recipient_token_suggest": 1.0, + "train_accuracy_recipient_token_suggest_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.15711462450592886, + "grad_norm": 2.1920664784328907, + "learning_rate": 4.798007040388212e-06, + "loss": 0.2247, + "step": 53 + }, + { + "epoch": 0.15711462450592886, + "step": 53, + "train_accuracy_first_token_arguments": 0.9428571428571428, + "train_accuracy_first_token_arguments_total": 35, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_calculate": 0.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.15711462450592886, + "step": 53, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 11, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.15711462450592886, + "step": 53, + "train_accuracy_first_token_arguments": 0.6857142857142857, + "train_accuracy_first_token_arguments_total": 35, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_generate": 1.0, + "train_accuracy_recipient_token_generate_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.1600790513833992, + "grad_norm": 1.6828766971614149, + "learning_rate": 4.788413440641297e-06, + "loss": 0.2468, + "step": 54 + }, + { + "epoch": 0.1600790513833992, + "step": 54, + "train_accuracy_first_token_arguments": 0.9130434782608695, + "train_accuracy_first_token_arguments_total": 23, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_flag": 0.6666666666666666, + "train_accuracy_recipient_token_flag_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.1600790513833992, + "step": 54, + "train_accuracy_first_token_arguments": 0.8461538461538461, + "train_accuracy_first_token_arguments_total": 26, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.8571428571428571, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.1600790513833992, + "step": 54, + "train_accuracy_first_token_arguments": 0.5476190476190477, + "train_accuracy_first_token_arguments_total": 42, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_log": 1.0, + "train_accuracy_recipient_token_log_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.16304347826086957, + "grad_norm": 2.3713370104618146, + "learning_rate": 4.778607322916896e-06, + "loss": 0.4323, + "step": 55 + }, + { + "epoch": 0.16304347826086957, + "step": 55, + "train_accuracy_first_token_arguments": 0.8627450980392157, + "train_accuracy_first_token_arguments_total": 51, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_create": 1.0, + "train_accuracy_recipient_token_create_total": 1, + "train_accuracy_recipient_token_draw": 1.0, + "train_accuracy_recipient_token_draw_total": 1, + "train_accuracy_recipient_token_generate": 1.0, + "train_accuracy_recipient_token_generate_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.16304347826086957, + "step": 55, + "train_accuracy_first_token_arguments": 0.9473684210526315, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_build": 0.5, + "train_accuracy_recipient_token_build_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_paint": 1.0, + "train_accuracy_recipient_token_paint_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.16304347826086957, + "step": 55, + "train_accuracy_first_token_arguments": 0.7368421052631579, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_fetch": 1.0, + "train_accuracy_recipient_token_fetch_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.16600790513833993, + "grad_norm": 1.892415917915294, + "learning_rate": 4.7685895978792564e-06, + "loss": 0.2455, + "step": 56 + }, + { + "epoch": 0.16600790513833993, + "step": 56, + "train_accuracy_first_token_arguments": 0.68, + "train_accuracy_first_token_arguments_total": 25, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_monitor": 0.0, + "train_accuracy_recipient_token_monitor_total": 1, + "train_accuracy_recipient_token_remove": 1.0, + "train_accuracy_recipient_token_remove_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.16600790513833993, + "step": 56, + "train_accuracy_first_token_arguments": 0.9761904761904762, + "train_accuracy_first_token_arguments_total": 42, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_decode": 1.0, + "train_accuracy_recipient_token_decode_total": 1, + "train_accuracy_recipient_token_explain": 1.0, + "train_accuracy_recipient_token_explain_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_number": 1.0, + "train_accuracy_recipient_token_number_total": 4, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.16600790513833993, + "step": 56, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.8571428571428571, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.16897233201581027, + "grad_norm": 1.7624413054210974, + "learning_rate": 4.7583611958439514e-06, + "loss": 0.2711, + "step": 57 + }, + { + "epoch": 0.16897233201581027, + "step": 57, + "train_accuracy_first_token_arguments": 0.8205128205128205, + "train_accuracy_first_token_arguments_total": 39, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_hire": 1.0, + "train_accuracy_recipient_token_hire_total": 1, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.16897233201581027, + "step": 57, + "train_accuracy_first_token_arguments": 0.8717948717948718, + "train_accuracy_first_token_arguments_total": 39, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.8571428571428571, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.16897233201581027, + "step": 57, + "train_accuracy_first_token_arguments": 0.7735849056603774, + "train_accuracy_first_token_arguments_total": 106, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_ex": 1.0, + "train_accuracy_recipient_token_ex_total": 3, + "train_accuracy_recipient_token_format": 0.0, + "train_accuracy_recipient_token_format_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.17193675889328064, + "grad_norm": 2.096786172099922, + "learning_rate": 4.747923066691487e-06, + "loss": 0.2883, + "step": 58 + }, + { + "epoch": 0.17193675889328064, + "step": 58, + "train_accuracy_first_token_arguments": 0.9090909090909091, + "train_accuracy_first_token_arguments_total": 11, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 10, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.17193675889328064, + "step": 58, + "train_accuracy_first_token_arguments": 0.58, + "train_accuracy_first_token_arguments_total": 50, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_f": 1.0, + "train_accuracy_recipient_token_f_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_generate": 1.0, + "train_accuracy_recipient_token_generate_total": 1, + "train_accuracy_recipient_token_wall": 1.0, + "train_accuracy_recipient_token_wall_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.17193675889328064, + "step": 58, + "train_accuracy_first_token_arguments": 0.5, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_arch": 1.0, + "train_accuracy_recipient_token_arch_total": 1, + "train_accuracy_recipient_token_building": 1.0, + "train_accuracy_recipient_token_building_total": 1, + "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.17490118577075098, + "grad_norm": 1.6122336767210732, + "learning_rate": 4.7372761797790836e-06, + "loss": 0.2393, + "step": 59 + }, + { + "epoch": 0.17490118577075098, + "step": 59, + "train_accuracy_first_token_arguments": 0.6153846153846154, + "train_accuracy_first_token_arguments_total": 26, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_detect": 1.0, + "train_accuracy_recipient_token_detect_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_truck": 1.0, + "train_accuracy_recipient_token_truck_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.17490118577075098, + "step": 59, + "train_accuracy_first_token_arguments": 0.6428571428571429, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.17490118577075098, + "step": 59, + "train_accuracy_first_token_arguments": 0.75, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.17786561264822134, + "grad_norm": 1.7027092503573524, + "learning_rate": 4.726421523850662e-06, + "loss": 0.1559, + "step": 60 + }, + { + "epoch": 0.17786561264822134, + "step": 60, + "train_accuracy_first_token_arguments": 0.7647058823529411, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_door": 1.0, + "train_accuracy_recipient_token_door_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_invest": 1.0, + "train_accuracy_recipient_token_invest_total": 1, + "train_accuracy_recipient_token_trans": 1.0, + "train_accuracy_recipient_token_trans_total": 3, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.17786561264822134, + "step": 60, + "train_accuracy_first_token_arguments": 0.8947368421052632, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.8125, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_generate": 0.0, + "train_accuracy_recipient_token_generate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.17786561264822134, + "step": 60, + "train_accuracy_first_token_arguments": 0.7692307692307693, + "train_accuracy_first_token_arguments_total": 26, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 2, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.1808300395256917, + "grad_norm": 1.640939742247205, + "learning_rate": 4.715360106945015e-06, + "loss": 0.2414, + "step": 61 + }, + { + "epoch": 0.1808300395256917, + "step": 61, + "train_accuracy_first_token_arguments": 0.875, + "train_accuracy_first_token_arguments_total": 24, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_animal": 1.0, + "train_accuracy_recipient_token_animal_total": 4, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_store": 1.0, + "train_accuracy_recipient_token_store_total": 3, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.1808300395256917, + "step": 61, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_plant": 0.75, + "train_accuracy_recipient_token_plant_total": 4, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.1808300395256917, + "step": 61, + "train_accuracy_first_token_arguments": 0.8947368421052632, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.18379446640316205, + "grad_norm": 2.3050792009276817, + "learning_rate": 4.7040929563021975e-06, + "loss": 0.2552, + "step": 62 + }, + { + "epoch": 0.18379446640316205, + "step": 62, + "train_accuracy_first_token_arguments": 0.9411764705882353, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.18379446640316205, + "step": 62, + "train_accuracy_first_token_arguments": 0.8, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_par": 1.0, + "train_accuracy_recipient_token_par_total": 2, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.18379446640316205, + "step": 62, + "train_accuracy_first_token_arguments": 0.8695652173913043, + "train_accuracy_first_token_arguments_total": 23, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_flower": 1.0, + "train_accuracy_recipient_token_flower_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.18675889328063242, + "grad_norm": 2.559937679792714, + "learning_rate": 4.6926211182681295e-06, + "loss": 0.4525, + "step": 63 + }, + { + "epoch": 0.18675889328063242, + "step": 63, + "train_accuracy_first_token_arguments": 0.9565217391304348, + "train_accuracy_first_token_arguments_total": 46, + "train_accuracy_recipient_token": 0.7058823529411765, + "train_accuracy_recipient_token_all": 0.6666666666666666, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_compare": 0.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 2, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.18675889328063242, + "step": 63, + "train_accuracy_first_token_arguments": 0.6190476190476191, + "train_accuracy_first_token_arguments_total": 21, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_emergency": 1.0, + "train_accuracy_recipient_token_emergency_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_first": 0.0, + "train_accuracy_recipient_token_first_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.18675889328063242, + "step": 63, + "train_accuracy_first_token_arguments": 0.6129032258064516, + "train_accuracy_first_token_arguments_total": 31, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_explain": 1.0, + "train_accuracy_recipient_token_explain_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.18972332015810275, + "grad_norm": 2.2063521775652393, + "learning_rate": 4.680945658197425e-06, + "loss": 0.2713, + "step": 64 + }, + { + "epoch": 0.18972332015810275, + "step": 64, + "train_accuracy_first_token_arguments": 0.9230769230769231, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.18972332015810275, + "step": 64, + "train_accuracy_first_token_arguments": 0.75, + "train_accuracy_first_token_arguments_total": 8, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.18972332015810275, + "step": 64, + "train_accuracy_first_token_arguments": 0.95, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.19268774703557312, + "grad_norm": 2.1111555785819514, + "learning_rate": 4.669067660354456e-06, + "loss": 0.2806, + "step": 65 + }, + { + "epoch": 0.19268774703557312, + "step": 65, + "train_accuracy_first_token_arguments": 0.7272727272727273, + "train_accuracy_first_token_arguments_total": 11, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_gr": 0.8, + "train_accuracy_recipient_token_gr_total": 5, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.19268774703557312, + "step": 65, + "train_accuracy_first_token_arguments": 0.7692307692307693, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.8666666666666667, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_license": 1.0, + "train_accuracy_recipient_token_license_total": 1, + "train_total_number_recipient_token": 15 + }, + { + "epoch": 0.19268774703557312, + "step": 65, + "train_accuracy_first_token_arguments": 0.9230769230769231, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.7777777777777778, + "train_accuracy_recipient_token_all": 0.6666666666666666, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_compose": 1.0, + "train_accuracy_recipient_token_compose_total": 4, + "train_accuracy_recipient_token_crop": 1.0, + "train_accuracy_recipient_token_crop_total": 1, + "train_accuracy_recipient_token_generate": 1.0, + "train_accuracy_recipient_token_generate_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.1956521739130435, + "grad_norm": 2.438873695871041, + "learning_rate": 4.656988227812658e-06, + "loss": 0.3234, + "step": 66 + }, + { + "epoch": 0.1956521739130435, + "step": 66, + "train_accuracy_first_token_arguments": 0.9534883720930233, + "train_accuracy_first_token_arguments_total": 43, + "train_accuracy_recipient_token": 0.7058823529411765, + "train_accuracy_recipient_token_all": 0.6923076923076923, + "train_accuracy_recipient_token_all_total": 13, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_schedule": 0.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.1956521739130435, + "step": 66, + "train_accuracy_first_token_arguments": 0.55, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 0.7, + "train_accuracy_recipient_token_add": 0.0, + "train_accuracy_recipient_token_add_total": 2, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 0.3333333333333333, + "train_accuracy_recipient_token_calculate_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_multiply": 0.0, + "train_accuracy_recipient_token_multiply_total": 2, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.1956521739130435, + "step": 66, + "train_accuracy_first_token_arguments": 0.9090909090909091, + "train_accuracy_first_token_arguments_total": 11, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.8571428571428571, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.19861660079051383, + "grad_norm": 2.7417774764849683, + "learning_rate": 4.644708482352093e-06, + "loss": 0.2789, + "step": 67 + }, + { + "epoch": 0.19861660079051383, + "step": 67, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_get": 0.875, + "train_accuracy_recipient_token_get_total": 8, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.19861660079051383, + "step": 67, + "train_accuracy_first_token_arguments": 0.9047619047619048, + "train_accuracy_first_token_arguments_total": 21, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_count": 1.0, + "train_accuracy_recipient_token_count_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.19861660079051383, + "step": 67, + "train_accuracy_first_token_arguments": 0.8846153846153846, + "train_accuracy_first_token_arguments_total": 26, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_build": 1.0, + "train_accuracy_recipient_token_build_total": 1, + "train_accuracy_recipient_token_conduct": 0.0, + "train_accuracy_recipient_token_conduct_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_max": 1.0, + "train_accuracy_recipient_token_max_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.2015810276679842, + "grad_norm": 2.3471565473735976, + "learning_rate": 4.632229564355275e-06, + "loss": 0.1919, + "step": 68 + }, + { + "epoch": 0.2015810276679842, + "step": 68, + "train_accuracy_first_token_arguments": 0.8888888888888888, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 0.75, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_be": 0.0, + "train_accuracy_recipient_token_be_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.2015810276679842, + "step": 68, + "train_accuracy_first_token_arguments": 0.6551724137931034, + "train_accuracy_first_token_arguments_total": 29, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_identify": 0.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.2015810276679842, + "step": 68, + "train_accuracy_first_token_arguments": 0.6, + "train_accuracy_first_token_arguments_total": 25, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_choose": 0.0, + "train_accuracy_recipient_token_choose_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_order": 0.0, + "train_accuracy_recipient_token_order_total": 1, + "train_accuracy_recipient_token_plan": 0.0, + "train_accuracy_recipient_token_plan_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.20454545454545456, + "grad_norm": 2.845262992508057, + "learning_rate": 4.619552632701263e-06, + "loss": 0.2993, + "step": 69 + }, + { + "epoch": 0.20454545454545456, + "step": 69, + "train_accuracy_first_token_arguments": 0.9166666666666666, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_armor": 1.0, + "train_accuracy_recipient_token_armor_total": 4, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.20454545454545456, + "step": 69, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 36, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.8461538461538461, + "train_accuracy_recipient_token_all_total": 13, + "train_accuracy_recipient_token_customize": 1.0, + "train_accuracy_recipient_token_customize_total": 2, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.20454545454545456, + "step": 69, + "train_accuracy_first_token_arguments": 0.9166666666666666, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 0.7142857142857143, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.2075098814229249, + "grad_norm": 2.2300979588264487, + "learning_rate": 4.606678864658039e-06, + "loss": 0.2752, + "step": 70 + }, + { + "epoch": 0.2075098814229249, + "step": 70, + "train_accuracy_first_token_arguments": 0.6666666666666666, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 2, + "train_accuracy_recipient_token_convert": 1.0, + "train_accuracy_recipient_token_convert_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_render": 0.0, + "train_accuracy_recipient_token_render_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.2075098814229249, + "step": 70, + "train_accuracy_first_token_arguments": 0.9827586206896551, + "train_accuracy_first_token_arguments_total": 58, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 0.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_mon": 0.75, + "train_accuracy_recipient_token_mon_total": 4, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.2075098814229249, + "step": 70, + "train_accuracy_first_token_arguments": 0.9047619047619048, + "train_accuracy_first_token_arguments_total": 21, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.21047430830039526, + "grad_norm": 2.1107565687083176, + "learning_rate": 4.5936094557731815e-06, + "loss": 0.228, + "step": 71 + }, + { + "epoch": 0.21047430830039526, + "step": 71, + "train_accuracy_first_token_arguments": 0.7307692307692307, + "train_accuracy_first_token_arguments_total": 26, + "train_accuracy_recipient_token": 0.7894736842105263, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.42857142857142855, + "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 2, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.21047430830039526, + "step": 71, + "train_accuracy_first_token_arguments": 0.7777777777777778, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_flower": 0.75, + "train_accuracy_recipient_token_flower_total": 4, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 2, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.21047430830039526, + "step": 71, + "train_accuracy_first_token_arguments": 0.7333333333333333, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.8571428571428571, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.2134387351778656, + "grad_norm": 2.1086876066685165, + "learning_rate": 4.5803456197628374e-06, + "loss": 0.2305, + "step": 72 + }, + { + "epoch": 0.2134387351778656, + "step": 72, + "train_accuracy_first_token_arguments": 0.8823529411764706, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 0.8888888888888888, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.2134387351778656, + "step": 72, + "train_accuracy_first_token_arguments": 0.5333333333333333, + "train_accuracy_first_token_arguments_total": 30, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_draw": 1.0, + "train_accuracy_recipient_token_draw_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 0.6, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.2134387351778656, + "step": 72, + "train_accuracy_first_token_arguments": 0.9, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 0.8, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_prior": 0.0, + "train_accuracy_recipient_token_prior_total": 1, + "train_accuracy_recipient_token_report": 0.0, + "train_accuracy_recipient_token_report_total": 1, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_accuracy_recipient_token_store": 0.5, + "train_accuracy_recipient_token_store_total": 2, + "train_total_number_recipient_token": 15 + }, + { + "epoch": 0.21640316205533597, + "grad_norm": 2.406206005805186, + "learning_rate": 4.566888588399007e-06, + "loss": 0.3295, + "step": 73 + }, + { + "epoch": 0.21640316205533597, + "step": 73, + "train_accuracy_first_token_arguments": 0.8846153846153846, + "train_accuracy_first_token_arguments_total": 26, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.8888888888888888, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.21640316205533597, + "step": 73, + "train_accuracy_first_token_arguments": 0.875, + "train_accuracy_first_token_arguments_total": 24, + "train_accuracy_recipient_token": 0.875, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_log": 0.5, + "train_accuracy_recipient_token_log_total": 2, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.21640316205533597, + "step": 73, + "train_accuracy_first_token_arguments": 0.9285714285714286, + "train_accuracy_first_token_arguments_total": 28, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_hire": 1.0, + "train_accuracy_recipient_token_hire_total": 1, + "train_accuracy_recipient_token_track": 1.0, + "train_accuracy_recipient_token_track_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.21936758893280633, + "grad_norm": 2.1304599088968823, + "learning_rate": 4.553239611395156e-06, + "loss": 0.2566, + "step": 74 + }, + { + "epoch": 0.21936758893280633, + "step": 74, + "train_accuracy_first_token_arguments": 0.8947368421052632, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_project": 1.0, + "train_accuracy_recipient_token_project_total": 1, + "train_accuracy_recipient_token_stat": 1.0, + "train_accuracy_recipient_token_stat_total": 4, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.21936758893280633, + "step": 74, + "train_accuracy_first_token_arguments": 0.5517241379310345, + "train_accuracy_first_token_arguments_total": 29, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 2, + "train_accuracy_recipient_token_suggest": 1.0, + "train_accuracy_recipient_token_suggest_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.21936758893280633, + "step": 74, + "train_accuracy_first_token_arguments": 0.75, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.7058823529411765, + "train_accuracy_recipient_token_all": 0.6363636363636364, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_church": 1.0, + "train_accuracy_recipient_token_church_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.22233201581027667, + "grad_norm": 2.03869187550964, + "learning_rate": 4.539399956290152e-06, + "loss": 0.3017, + "step": 75 + }, + { + "epoch": 0.22233201581027667, + "step": 75, + "train_accuracy_first_token_arguments": 0.7391304347826086, + "train_accuracy_first_token_arguments_total": 46, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 9, + "train_accuracy_recipient_token_f": 0.0, + "train_accuracy_recipient_token_f_total": 1, + "train_accuracy_recipient_token_plant": 1.0, + "train_accuracy_recipient_token_plant_total": 1, + "train_accuracy_recipient_token_se": 0.0, + "train_accuracy_recipient_token_se_total": 1, + "train_accuracy_recipient_token_trim": 1.0, + "train_accuracy_recipient_token_trim_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.22233201581027667, + "step": 75, + "train_accuracy_first_token_arguments": 0.9285714285714286, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.7857142857142857, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_f": 1.0, + "train_accuracy_recipient_token_f_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_restore": 0.0, + "train_accuracy_recipient_token_restore_total": 1, + "train_total_number_recipient_token": 14 + }, + { + "epoch": 0.22233201581027667, + "step": 75, + "train_accuracy_first_token_arguments": 0.9411764705882353, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_gr": 0.6666666666666666, + "train_accuracy_recipient_token_gr_total": 3, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.22529644268774704, + "grad_norm": 2.1715977940434494, + "learning_rate": 4.525370908330564e-06, + "loss": 0.2781, + "step": 76 + }, + { + "epoch": 0.22529644268774704, + "step": 76, + "train_accuracy_first_token_arguments": 0.8421052631578947, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.7647058823529411, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 0.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.5, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.22529644268774704, + "step": 76, + "train_accuracy_first_token_arguments": 0.71875, + "train_accuracy_first_token_arguments_total": 64, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.7272727272727273, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.22529644268774704, + "step": 76, + "train_accuracy_first_token_arguments": 0.9523809523809523, + "train_accuracy_first_token_arguments_total": 21, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.22826086956521738, + "grad_norm": 2.6029973414283223, + "learning_rate": 4.511153770351288e-06, + "loss": 0.3502, + "step": 77 + }, + { + "epoch": 0.22826086956521738, + "step": 77, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.22826086956521738, + "step": 77, + "train_accuracy_first_token_arguments": 0.6153846153846154, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_flower": 1.0, + "train_accuracy_recipient_token_flower_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.22826086956521738, + "step": 77, + "train_accuracy_first_token_arguments": 0.7049180327868853, + "train_accuracy_first_token_arguments_total": 61, + "train_accuracy_recipient_token": 0.9333333333333333, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 15 + }, + { + "epoch": 0.23122529644268774, + "grad_norm": 2.67404592050148, + "learning_rate": 4.496749862654574e-06, + "loss": 0.3309, + "step": 78 + }, + { + "epoch": 0.23122529644268774, + "step": 78, + "train_accuracy_first_token_arguments": 0.9, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_add": 1.0, + "train_accuracy_recipient_token_add_total": 1, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.23122529644268774, + "step": 78, + "train_accuracy_first_token_arguments": 0.95, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.23122529644268774, + "step": 78, + "train_accuracy_first_token_arguments": 0.84, + "train_accuracy_first_token_arguments_total": 25, + "train_accuracy_recipient_token": 0.9523809523809523, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_extract": 1.0, + "train_accuracy_recipient_token_extract_total": 1, + "train_accuracy_recipient_token_generate": 1.0, + "train_accuracy_recipient_token_generate_total": 2, + "train_accuracy_recipient_token_sum": 1.0, + "train_accuracy_recipient_token_sum_total": 1, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 2, + "train_total_number_recipient_token": 21 + }, + { + "epoch": 0.2341897233201581, + "grad_norm": 1.8475653464555015, + "learning_rate": 4.482160522887404e-06, + "loss": 0.228, + "step": 79 + }, + { + "epoch": 0.2341897233201581, + "step": 79, + "train_accuracy_first_token_arguments": 0.6153846153846154, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.2341897233201581, + "step": 79, + "train_accuracy_first_token_arguments": 0.75, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_arc": 1.0, + "train_accuracy_recipient_token_arc_total": 4, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.2341897233201581, + "step": 79, + "train_accuracy_first_token_arguments": 0.4444444444444444, + "train_accuracy_first_token_arguments_total": 45, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.23715415019762845, + "grad_norm": 2.189256435693527, + "learning_rate": 4.467387105917269e-06, + "loss": 0.3682, + "step": 80 + }, + { + "epoch": 0.23715415019762845, + "step": 80, + "train_accuracy_first_token_arguments": 0.8275862068965517, + "train_accuracy_first_token_arguments_total": 29, + "train_accuracy_recipient_token": 0.8125, + "train_accuracy_recipient_token_all": 0.7, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_property": 1.0, + "train_accuracy_recipient_token_property_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.23715415019762845, + "step": 80, + "train_accuracy_first_token_arguments": 0.7333333333333333, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.23715415019762845, + "step": 80, + "train_accuracy_first_token_arguments": 0.8571428571428571, + "train_accuracy_first_token_arguments_total": 21, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 0.6666666666666666, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.24011857707509882, + "grad_norm": 1.6646080394168112, + "learning_rate": 4.452430983706351e-06, + "loss": 0.1579, + "step": 81 + }, + { + "epoch": 0.24011857707509882, + "step": 81, + "train_accuracy_first_token_arguments": 0.875, + "train_accuracy_first_token_arguments_total": 8, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_art": 1.0, + "train_accuracy_recipient_token_art_total": 2, + "train_accuracy_recipient_token_explain": 1.0, + "train_accuracy_recipient_token_explain_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_sc": 1.0, + "train_accuracy_recipient_token_sc_total": 2, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.24011857707509882, + "step": 81, + "train_accuracy_first_token_arguments": 0.8, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.8125, + "train_accuracy_recipient_token_all": 0.7, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.24011857707509882, + "step": 81, + "train_accuracy_first_token_arguments": 0.75, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_box": 1.0, + "train_accuracy_recipient_token_box_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.24308300395256918, + "grad_norm": 1.9011227117633136, + "learning_rate": 4.437293545184111e-06, + "loss": 0.2544, + "step": 82 + }, + { + "epoch": 0.24308300395256918, + "step": 82, + "train_accuracy_first_token_arguments": 0.9090909090909091, + "train_accuracy_first_token_arguments_total": 11, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.24308300395256918, + "step": 82, + "train_accuracy_first_token_arguments": 0.631578947368421, + "train_accuracy_first_token_arguments_total": 38, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_get": 0.8333333333333334, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_inspect": 1.0, + "train_accuracy_recipient_token_inspect_total": 1, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.24308300395256918, + "step": 82, + "train_accuracy_first_token_arguments": 0.8846153846153846, + "train_accuracy_first_token_arguments_total": 26, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 9, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.24604743083003952, + "grad_norm": 1.9670272993707267, + "learning_rate": 4.421976196118297e-06, + "loss": 0.2366, + "step": 83 + }, + { + "epoch": 0.24604743083003952, + "step": 83, + "train_accuracy_first_token_arguments": 0.9090909090909091, + "train_accuracy_first_token_arguments_total": 22, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.24604743083003952, + "step": 83, + "train_accuracy_first_token_arguments": 0.7368421052631579, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.7777777777777778, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 0.5, + "train_accuracy_recipient_token_compare_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.5, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.24604743083003952, + "step": 83, + "train_accuracy_first_token_arguments": 0.7, + "train_accuracy_first_token_arguments_total": 10, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compile": 1.0, + "train_accuracy_recipient_token_compile_total": 2, + "train_accuracy_recipient_token_compress": 1.0, + "train_accuracy_recipient_token_compress_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_order": 1.0, + "train_accuracy_recipient_token_order_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.2490118577075099, + "grad_norm": 2.214361868557108, + "learning_rate": 4.40648035898441e-06, + "loss": 0.3157, + "step": 84 + }, + { + "epoch": 0.2490118577075099, + "step": 84, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.8571428571428571, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.2490118577075099, + "step": 84, + "train_accuracy_first_token_arguments": 0.9285714285714286, + "train_accuracy_first_token_arguments_total": 28, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 9, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_research": 1.0, + "train_accuracy_recipient_token_research_total": 2, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.2490118577075099, + "step": 84, + "train_accuracy_first_token_arguments": 0.8888888888888888, + "train_accuracy_first_token_arguments_total": 9, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compose": 1.0, + "train_accuracy_recipient_token_compose_total": 4, + "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.2519762845849802, + "grad_norm": 1.79425216853278, + "learning_rate": 4.390807472833585e-06, + "loss": 0.2629, + "step": 85 + }, + { + "epoch": 0.2519762845849802, + "step": 85, + "train_accuracy_first_token_arguments": 0.9375, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_write": 1.0, + "train_accuracy_recipient_token_write_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.2519762845849802, + "step": 85, + "train_accuracy_first_token_arguments": 0.8461538461538461, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.7777777777777778, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.2519762845849802, + "step": 85, + "train_accuracy_first_token_arguments": 0.7272727272727273, + "train_accuracy_first_token_arguments_total": 11, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_explain": 1.0, + "train_accuracy_recipient_token_explain_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.2549407114624506, + "grad_norm": 2.256598889106997, + "learning_rate": 4.374958993158965e-06, + "loss": 0.2383, + "step": 86 + }, + { + "epoch": 0.2549407114624506, + "step": 86, + "train_accuracy_first_token_arguments": 0.7368421052631579, + "train_accuracy_first_token_arguments_total": 38, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_flag": 1.0, + "train_accuracy_recipient_token_flag_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.2549407114624506, + "step": 86, + "train_accuracy_first_token_arguments": 0.8695652173913043, + "train_accuracy_first_token_arguments_total": 23, + "train_accuracy_recipient_token": 0.7894736842105263, + "train_accuracy_recipient_token_all": 0.8461538461538461, + "train_accuracy_recipient_token_all_total": 13, + "train_accuracy_recipient_token_compare": 0.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_describe": 1.0, + "train_accuracy_recipient_token_describe_total": 4, + "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.2549407114624506, + "step": 86, + "train_accuracy_first_token_arguments": 0.8378378378378378, + "train_accuracy_first_token_arguments_total": 37, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_ass": 0.5, + "train_accuracy_recipient_token_ass_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_plan": 1.0, + "train_accuracy_recipient_token_plan_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.25790513833992096, + "grad_norm": 2.49197628899743, + "learning_rate": 4.358936391760524e-06, + "loss": 0.4685, + "step": 87 + }, + { + "epoch": 0.25790513833992096, + "step": 87, + "train_accuracy_first_token_arguments": 0.9285714285714286, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_accuracy_recipient_token_identify": 0.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.25790513833992096, + "step": 87, + "train_accuracy_first_token_arguments": 0.9583333333333334, + "train_accuracy_first_token_arguments_total": 24, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9166666666666666, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_cross": 1.0, + "train_accuracy_recipient_token_cross_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.25790513833992096, + "step": 87, + "train_accuracy_first_token_arguments": 0.9047619047619048, + "train_accuracy_first_token_arguments_total": 21, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.2608695652173913, + "grad_norm": 2.280644257269206, + "learning_rate": 4.342741156608392e-06, + "loss": 0.2713, + "step": 88 + }, + { + "epoch": 0.2608695652173913, + "step": 88, + "train_accuracy_first_token_arguments": 0.9444444444444444, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 0.875, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.8333333333333334, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.2608695652173913, + "step": 88, + "train_accuracy_first_token_arguments": 0.7407407407407407, + "train_accuracy_first_token_arguments_total": 27, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_track": 1.0, + "train_accuracy_recipient_token_track_total": 1, + "train_accuracy_recipient_token_write": 0.0, + "train_accuracy_recipient_token_write_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.2608695652173913, + "step": 88, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_list": 1.0, + "train_accuracy_recipient_token_list_total": 1, + "train_accuracy_recipient_token_rent": 1.0, + "train_accuracy_recipient_token_rent_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.26383399209486164, + "grad_norm": 2.337688603854322, + "learning_rate": 4.32637479170467e-06, + "loss": 0.3134, + "step": 89 + }, + { + "epoch": 0.26383399209486164, + "step": 89, + "train_accuracy_first_token_arguments": 0.6666666666666666, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_explain": 0.0, + "train_accuracy_recipient_token_explain_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_track": 1.0, + "train_accuracy_recipient_token_track_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.26383399209486164, + "step": 89, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_evaluate": 1.0, + "train_accuracy_recipient_token_evaluate_total": 1, + "train_accuracy_recipient_token_fetch": 1.0, + "train_accuracy_recipient_token_fetch_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_sport": 1.0, + "train_accuracy_recipient_token_sport_total": 3, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.26383399209486164, + "step": 89, + "train_accuracy_first_token_arguments": 0.9047619047619048, + "train_accuracy_first_token_arguments_total": 21, + "train_accuracy_recipient_token": 0.8125, + "train_accuracy_recipient_token_add": 1.0, + "train_accuracy_recipient_token_add_total": 1, + "train_accuracy_recipient_token_all": 0.7692307692307693, + "train_accuracy_recipient_token_all_total": 13, + "train_accuracy_recipient_token_apply": 1.0, + "train_accuracy_recipient_token_apply_total": 1, + "train_accuracy_recipient_token_crop": 1.0, + "train_accuracy_recipient_token_crop_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.26679841897233203, + "grad_norm": 2.596535079636875, + "learning_rate": 4.309838816943755e-06, + "loss": 0.2432, + "step": 90 + }, + { + "epoch": 0.26679841897233203, + "step": 90, + "train_accuracy_first_token_arguments": 0.76, + "train_accuracy_first_token_arguments_total": 25, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.26679841897233203, + "step": 90, + "train_accuracy_first_token_arguments": 0.9333333333333333, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_m": 1.0, + "train_accuracy_recipient_token_m_total": 3, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.26679841897233203, + "step": 90, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.75, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.26976284584980237, + "grad_norm": 2.2037749517640015, + "learning_rate": 4.293134767971193e-06, + "loss": 0.2113, + "step": 91 + }, + { + "epoch": 0.26976284584980237, + "step": 91, + "train_accuracy_first_token_arguments": 0.7, + "train_accuracy_first_token_arguments_total": 10, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_ele": 1.0, + "train_accuracy_recipient_token_ele_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.26976284584980237, + "step": 91, + "train_accuracy_first_token_arguments": 0.6363636363636364, + "train_accuracy_first_token_arguments_total": 11, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_decode": 1.0, + "train_accuracy_recipient_token_decode_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.26976284584980237, + "step": 91, + "train_accuracy_first_token_arguments": 0.8103448275862069, + "train_accuracy_first_token_arguments_total": 58, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 2, + "train_accuracy_recipient_token_object": 1.0, + "train_accuracy_recipient_token_object_total": 2, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.2727272727272727, + "grad_norm": 2.1267599638494628, + "learning_rate": 4.276264196041074e-06, + "loss": 0.3654, + "step": 92 + }, + { + "epoch": 0.2727272727272727, + "step": 92, + "train_accuracy_first_token_arguments": 0.8292682926829268, + "train_accuracy_first_token_arguments_total": 41, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_deploy": 1.0, + "train_accuracy_recipient_token_deploy_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_plan": 1.0, + "train_accuracy_recipient_token_plan_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.2727272727272727, + "step": 92, + "train_accuracy_first_token_arguments": 0.7105263157894737, + "train_accuracy_first_token_arguments_total": 38, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_convert": 1.0, + "train_accuracy_recipient_token_convert_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.2727272727272727, + "step": 92, + "train_accuracy_first_token_arguments": 0.8421052631578947, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_detect": 0.0, + "train_accuracy_recipient_token_detect_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 2, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.2756916996047431, + "grad_norm": 2.2199589796264414, + "learning_rate": 4.259228667871963e-06, + "loss": 0.3508, + "step": 93 + }, + { + "epoch": 0.2756916996047431, + "step": 93, + "train_accuracy_first_token_arguments": 0.9444444444444444, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 0.7894736842105263, + "train_accuracy_recipient_token_add": 1.0, + "train_accuracy_recipient_token_add_total": 1, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_calculate": 0.5, + "train_accuracy_recipient_token_calculate_total": 2, + "train_accuracy_recipient_token_decode": 1.0, + "train_accuracy_recipient_token_decode_total": 4, + "train_accuracy_recipient_token_subtract": 0.0, + "train_accuracy_recipient_token_subtract_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.2756916996047431, + "step": 93, + "train_accuracy_first_token_arguments": 0.7407407407407407, + "train_accuracy_first_token_arguments_total": 27, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.2756916996047431, + "step": 93, + "train_accuracy_first_token_arguments": 0.875, + "train_accuracy_first_token_arguments_total": 24, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_get": 0.8333333333333334, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_study": 1.0, + "train_accuracy_recipient_token_study_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.27865612648221344, + "grad_norm": 2.6263899594688755, + "learning_rate": 4.242029765501411e-06, + "loss": 0.2873, + "step": 94 + }, + { + "epoch": 0.27865612648221344, + "step": 94, + "train_accuracy_first_token_arguments": 0.9473684210526315, + "train_accuracy_first_token_arguments_total": 76, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_d": 0.5, + "train_accuracy_recipient_token_d_total": 2, + "train_accuracy_recipient_token_fly": 1.0, + "train_accuracy_recipient_token_fly_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.27865612648221344, + "step": 94, + "train_accuracy_first_token_arguments": 0.64, + "train_accuracy_first_token_arguments_total": 25, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 0.6666666666666666, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.27865612648221344, + "step": 94, + "train_accuracy_first_token_arguments": 0.8125, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_horse": 1.0, + "train_accuracy_recipient_token_horse_total": 2, + "train_accuracy_recipient_token_invest": 1.0, + "train_accuracy_recipient_token_invest_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.2816205533596838, + "grad_norm": 1.8705277085481686, + "learning_rate": 4.22466908613903e-06, + "loss": 0.26, + "step": 95 + }, + { + "epoch": 0.2816205533596838, + "step": 95, + "train_accuracy_first_token_arguments": 0.8076923076923077, + "train_accuracy_first_token_arguments_total": 26, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_arr": 1.0, + "train_accuracy_recipient_token_arr_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_order": 1.0, + "train_accuracy_recipient_token_order_total": 1, + "train_accuracy_recipient_token_suggest": 1.0, + "train_accuracy_recipient_token_suggest_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.2816205533596838, + "step": 95, + "train_accuracy_first_token_arguments": 0.5510204081632653, + "train_accuracy_first_token_arguments_total": 49, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_f": 1.0, + "train_accuracy_recipient_token_f_total": 4, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_im": 1.0, + "train_accuracy_recipient_token_im_total": 1, + "train_accuracy_recipient_token_schedule": 0.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.2816205533596838, + "step": 95, + "train_accuracy_first_token_arguments": 0.8947368421052632, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.7272727272727273, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_sh": 1.0, + "train_accuracy_recipient_token_sh_total": 3, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.2845849802371542, + "grad_norm": 2.0123868841853065, + "learning_rate": 4.20714824201817e-06, + "loss": 0.3846, + "step": 96 + }, + { + "epoch": 0.2845849802371542, + "step": 96, + "train_accuracy_first_token_arguments": 0.8928571428571429, + "train_accuracy_first_token_arguments_total": 28, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.7272727272727273, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.2845849802371542, + "step": 96, + "train_accuracy_first_token_arguments": 0.88, + "train_accuracy_first_token_arguments_total": 25, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_accuracy_recipient_token_verify": 1.0, + "train_accuracy_recipient_token_verify_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.2845849802371542, + "step": 96, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 26, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.8888888888888888, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.2875494071146245, + "grad_norm": 1.6695567253827632, + "learning_rate": 4.189468860246192e-06, + "loss": 0.1603, + "step": 97 + }, + { + "epoch": 0.2875494071146245, + "step": 97, + "train_accuracy_first_token_arguments": 0.75, + "train_accuracy_first_token_arguments_total": 36, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_enh": 1.0, + "train_accuracy_recipient_token_enh_total": 1, + "train_accuracy_recipient_token_explain": 1.0, + "train_accuracy_recipient_token_explain_total": 1, + "train_accuracy_recipient_token_public": 1.0, + "train_accuracy_recipient_token_public_total": 1, + "train_accuracy_recipient_token_street": 1.0, + "train_accuracy_recipient_token_street_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.2875494071146245, + "step": 97, + "train_accuracy_first_token_arguments": 0.9090909090909091, + "train_accuracy_first_token_arguments_total": 44, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.75, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.2875494071146245, + "step": 97, + "train_accuracy_first_token_arguments": 0.8947368421052632, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.29051383399209485, + "grad_norm": 2.3366857423927834, + "learning_rate": 4.171632582653368e-06, + "loss": 0.3712, + "step": 98 + }, + { + "epoch": 0.29051383399209485, + "step": 98, + "train_accuracy_first_token_arguments": 0.8636363636363636, + "train_accuracy_first_token_arguments_total": 22, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 0.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_c": 0.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.29051383399209485, + "step": 98, + "train_accuracy_first_token_arguments": 0.8260869565217391, + "train_accuracy_first_token_arguments_total": 23, + "train_accuracy_recipient_token": 0.8125, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 0.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_assign": 1.0, + "train_accuracy_recipient_token_assign_total": 1, + "train_accuracy_recipient_token_configure": 1.0, + "train_accuracy_recipient_token_configure_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.29051383399209485, + "step": 98, + "train_accuracy_first_token_arguments": 0.6153846153846154, + "train_accuracy_first_token_arguments_total": 52, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 9, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.29347826086956524, + "grad_norm": 2.2368586683604175, + "learning_rate": 4.153641065640402e-06, + "loss": 0.238, + "step": 99 + }, + { + "epoch": 0.29347826086956524, + "step": 99, + "train_accuracy_first_token_arguments": 0.9375, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.8, + "train_accuracy_recipient_token_find_total": 5, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.29347826086956524, + "step": 99, + "train_accuracy_first_token_arguments": 0.8, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 0.875, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_clean": 1.0, + "train_accuracy_recipient_token_clean_total": 1, + "train_accuracy_recipient_token_install": 1.0, + "train_accuracy_recipient_token_install_total": 1, + "train_accuracy_recipient_token_remove": 0.0, + "train_accuracy_recipient_token_remove_total": 1, + "train_accuracy_recipient_token_replace": 0.0, + "train_accuracy_recipient_token_replace_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.29347826086956524, + "step": 99, + "train_accuracy_first_token_arguments": 0.9411764705882353, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 4, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 2, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.2964426877470356, + "grad_norm": 2.6698723558857944, + "learning_rate": 4.1354959800246155e-06, + "loss": 0.2791, + "step": 100 + }, + { + "epoch": 0.2964426877470356, + "step": 100, + "train_accuracy_first_token_arguments": 0.6666666666666666, + "train_accuracy_first_token_arguments_total": 21, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_facility": 1.0, + "train_accuracy_recipient_token_facility_total": 3, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.2964426877470356, + "step": 100, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 0.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_mount": 1.0, + "train_accuracy_recipient_token_mount_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.2964426877470356, + "step": 100, + "train_accuracy_first_token_arguments": 0.71875, + "train_accuracy_first_token_arguments_total": 32, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 0.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_park": 0.6666666666666666, + "train_accuracy_recipient_token_park_total": 3, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.2994071146245059, + "grad_norm": 1.5724579692892722, + "learning_rate": 4.1171990108847705e-06, + "loss": 0.1851, + "step": 101 + }, + { + "epoch": 0.2994071146245059, + "step": 101, + "train_accuracy_first_token_arguments": 0.9069767441860465, + "train_accuracy_first_token_arguments_total": 43, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_order": 1.0, + "train_accuracy_recipient_token_order_total": 1, + "train_accuracy_recipient_token_schedule": 0.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.2994071146245059, + "step": 101, + "train_accuracy_first_token_arguments": 0.5909090909090909, + "train_accuracy_first_token_arguments_total": 22, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_detect": 1.0, + "train_accuracy_recipient_token_detect_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.2994071146245059, + "step": 101, + "train_accuracy_first_token_arguments": 0.7692307692307693, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.8333333333333334, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.30237154150197626, + "grad_norm": 1.3808933034142918, + "learning_rate": 4.098751857404595e-06, + "loss": 0.1429, + "step": 102 + }, + { + "epoch": 0.30237154150197626, + "step": 102, + "train_accuracy_first_token_arguments": 0.9166666666666666, + "train_accuracy_first_token_arguments_total": 24, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.7272727272727273, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_detect": 1.0, + "train_accuracy_recipient_token_detect_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_sum": 1.0, + "train_accuracy_recipient_token_sum_total": 1, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.30237154150197626, + "step": 102, + "train_accuracy_first_token_arguments": 0.5833333333333334, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_city": 1.0, + "train_accuracy_recipient_token_city_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.30237154150197626, + "step": 102, + "train_accuracy_first_token_arguments": 0.8421052631578947, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_event": 1.0, + "train_accuracy_recipient_token_event_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.30533596837944665, + "grad_norm": 2.4623439109068372, + "learning_rate": 4.080156232714976e-06, + "loss": 0.4204, + "step": 103 + }, + { + "epoch": 0.30533596837944665, + "step": 103, + "train_accuracy_first_token_arguments": 0.95, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_bike": 1.0, + "train_accuracy_recipient_token_bike_total": 2, + "train_accuracy_recipient_token_calculate": 0.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.30533596837944665, + "step": 103, + "train_accuracy_first_token_arguments": 0.9, + "train_accuracy_first_token_arguments_total": 10, + "train_accuracy_recipient_token": 0.9375, + "train_accuracy_recipient_token_all": 0.9166666666666666, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_disc": 1.0, + "train_accuracy_recipient_token_disc_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.30533596837944665, + "step": 103, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_convert": 1.0, + "train_accuracy_recipient_token_convert_total": 2, + "train_accuracy_recipient_token_count": 1.0, + "train_accuracy_recipient_token_count_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 21 + }, + { + "epoch": 0.308300395256917, + "grad_norm": 1.9389616807058327, + "learning_rate": 4.061413863734869e-06, + "loss": 0.169, + "step": 104 + }, + { + "epoch": 0.308300395256917, + "step": 104, + "train_accuracy_first_token_arguments": 0.7, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 2, + "train_accuracy_recipient_token_interpret": 1.0, + "train_accuracy_recipient_token_interpret_total": 2, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.308300395256917, + "step": 104, + "train_accuracy_first_token_arguments": 0.9523809523809523, + "train_accuracy_first_token_arguments_total": 21, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 5, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.308300395256917, + "step": 104, + "train_accuracy_first_token_arguments": 0.8421052631578947, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_list": 1.0, + "train_accuracy_recipient_token_list_total": 1, + "train_accuracy_recipient_token_villa": 1.0, + "train_accuracy_recipient_token_villa_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.31126482213438733, + "grad_norm": 1.819003861346643, + "learning_rate": 4.0425264910109245e-06, + "loss": 0.1843, + "step": 105 + }, + { + "epoch": 0.31126482213438733, + "step": 105, + "train_accuracy_first_token_arguments": 0.8333333333333334, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_m": 1.0, + "train_accuracy_recipient_token_m_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.31126482213438733, + "step": 105, + "train_accuracy_first_token_arguments": 0.875, + "train_accuracy_first_token_arguments_total": 8, + "train_accuracy_recipient_token": 0.9375, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_list": 1.0, + "train_accuracy_recipient_token_list_total": 1, + "train_accuracy_recipient_token_sum": 1.0, + "train_accuracy_recipient_token_sum_total": 1, + "train_accuracy_recipient_token_verify": 1.0, + "train_accuracy_recipient_token_verify_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.31126482213438733, + "step": 105, + "train_accuracy_first_token_arguments": 0.8888888888888888, + "train_accuracy_first_token_arguments_total": 9, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.3142292490118577, + "grad_norm": 2.4539744307883553, + "learning_rate": 4.023495868555848e-06, + "loss": 0.4031, + "step": 106 + }, + { + "epoch": 0.3142292490118577, + "eval_accuracy_first_token_arguments": 0.9266727772685609, + "eval_accuracy_first_token_arguments_total": 1091, + "eval_accuracy_recipient_token": 0.9235033259423503, + "eval_accuracy_recipient_token_all": 0.9044834307992202, + "eval_accuracy_recipient_token_all_total": 513, + "eval_accuracy_recipient_token_calculate": 1.0, + "eval_accuracy_recipient_token_calculate_total": 13, + "eval_accuracy_recipient_token_find": 0.9649122807017544, + "eval_accuracy_recipient_token_find_total": 57, + "eval_accuracy_recipient_token_get": 0.9744897959183674, + "eval_accuracy_recipient_token_get_total": 196, + "eval_accuracy_recipient_token_identify": 0.8333333333333334, + "eval_accuracy_recipient_token_identify_total": 6, + "eval_loss": 0.25004321336746216, + "eval_perplexity": 1.0396230889094866, + "eval_runtime": 119.1349, + "eval_samples_per_second": 0.848, + "eval_steps_per_second": 0.428, + "eval_total_number_recipient_token": 902, + "step": 106 + }, + { + "epoch": 0.3142292490118577, + "step": 106, + "train_accuracy_first_token_arguments": 0.8, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_order": 1.0, + "train_accuracy_recipient_token_order_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.3142292490118577, + "step": 106, + "train_accuracy_first_token_arguments": 0.88, + "train_accuracy_first_token_arguments_total": 25, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 5, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_report": 1.0, + "train_accuracy_recipient_token_report_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.3142292490118577, + "step": 106, + "train_accuracy_first_token_arguments": 0.4166666666666667, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_interpret": 1.0, + "train_accuracy_recipient_token_interpret_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.31719367588932806, + "grad_norm": 2.6219196171389454, + "learning_rate": 4.004323763685511e-06, + "loss": 0.2683, + "step": 107 + }, + { + "epoch": 0.31719367588932806, + "step": 107, + "train_accuracy_first_token_arguments": 0.9130434782608695, + "train_accuracy_first_token_arguments_total": 23, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_d": 1.0, + "train_accuracy_recipient_token_d_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.31719367588932806, + "step": 107, + "train_accuracy_first_token_arguments": 0.918918918918919, + "train_accuracy_first_token_arguments_total": 37, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_audit": 1.0, + "train_accuracy_recipient_token_audit_total": 1, + "train_accuracy_recipient_token_environment": 1.0, + "train_accuracy_recipient_token_environment_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_public": 1.0, + "train_accuracy_recipient_token_public_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.31719367588932806, + "step": 107, + "train_accuracy_first_token_arguments": 0.8333333333333334, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_f": 1.0, + "train_accuracy_recipient_token_f_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_search": 0.6666666666666666, + "train_accuracy_recipient_token_search_total": 3, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.3201581027667984, + "grad_norm": 2.3979253771216524, + "learning_rate": 3.985011956854826e-06, + "loss": 0.3371, + "step": 108 + }, + { + "epoch": 0.3201581027667984, + "step": 108, + "train_accuracy_first_token_arguments": 0.9285714285714286, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.3201581027667984, + "step": 108, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 0.875, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_histor": 1.0, + "train_accuracy_recipient_token_histor_total": 2, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.3201581027667984, + "step": 108, + "train_accuracy_first_token_arguments": 0.8571428571428571, + "train_accuracy_first_token_arguments_total": 28, + "train_accuracy_recipient_token": 0.5, + "train_accuracy_recipient_token_all": 0.5454545454545454, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_detect": 0.0, + "train_accuracy_recipient_token_detect_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_translate": 0.5, + "train_accuracy_recipient_token_translate_total": 2, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.3231225296442688, + "grad_norm": 2.9561525888835276, + "learning_rate": 3.965562241492401e-06, + "loss": 0.3818, + "step": 109 + }, + { + "epoch": 0.3231225296442688, + "step": 109, + "train_accuracy_first_token_arguments": 0.7142857142857143, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_plan": 1.0, + "train_accuracy_recipient_token_plan_total": 1, + "train_accuracy_recipient_token_suggest": 1.0, + "train_accuracy_recipient_token_suggest_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.3231225296442688, + "step": 109, + "train_accuracy_first_token_arguments": 0.9230769230769231, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.3231225296442688, + "step": 109, + "train_accuracy_first_token_arguments": 0.96875, + "train_accuracy_first_token_arguments_total": 32, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.32608695652173914, + "grad_norm": 1.9549696262976763, + "learning_rate": 3.945976423833987e-06, + "loss": 0.1918, + "step": 110 + }, + { + "epoch": 0.32608695652173914, + "step": 110, + "train_accuracy_first_token_arguments": 0.9333333333333333, + "train_accuracy_first_token_arguments_total": 30, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.7692307692307693, + "train_accuracy_recipient_token_all_total": 13, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_detect": 1.0, + "train_accuracy_recipient_token_detect_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.32608695652173914, + "step": 110, + "train_accuracy_first_token_arguments": 0.7894736842105263, + "train_accuracy_first_token_arguments_total": 38, + "train_accuracy_recipient_token": 0.9375, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_race": 1.0, + "train_accuracy_recipient_token_race_total": 2, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.32608695652173914, + "step": 110, + "train_accuracy_first_token_arguments": 0.9333333333333333, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 0.9166666666666666, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_evaluate": 1.0, + "train_accuracy_recipient_token_evaluate_total": 1, + "train_accuracy_recipient_token_get": 0.8571428571428571, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.3290513833992095, + "grad_norm": 2.295435621606061, + "learning_rate": 3.92625632275474e-06, + "loss": 0.378, + "step": 111 + }, + { + "epoch": 0.3290513833992095, + "step": 111, + "train_accuracy_first_token_arguments": 0.8888888888888888, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_add": 1.0, + "train_accuracy_recipient_token_add_total": 1, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_change": 1.0, + "train_accuracy_recipient_token_change_total": 1, + "train_accuracy_recipient_token_highlight": 0.5, + "train_accuracy_recipient_token_highlight_total": 2, + "train_accuracy_recipient_token_remove": 1.0, + "train_accuracy_recipient_token_remove_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.3290513833992095, + "step": 111, + "train_accuracy_first_token_arguments": 0.896551724137931, + "train_accuracy_first_token_arguments_total": 29, + "train_accuracy_recipient_token": 0.7647058823529411, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_configure": 1.0, + "train_accuracy_recipient_token_configure_total": 1, + "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.5, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.3290513833992095, + "step": 111, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.8571428571428571, + "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 2, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.33201581027667987, + "grad_norm": 2.4185983751815114, + "learning_rate": 3.906403769600311e-06, + "loss": 0.2139, + "step": 112 + }, + { + "epoch": 0.33201581027667987, + "step": 112, + "train_accuracy_first_token_arguments": 0.9090909090909091, + "train_accuracy_first_token_arguments_total": 22, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_dispose": 1.0, + "train_accuracy_recipient_token_dispose_total": 1, + "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_up": 1.0, + "train_accuracy_recipient_token_up_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.33201581027667987, + "step": 112, + "train_accuracy_first_token_arguments": 0.9230769230769231, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.33201581027667987, + "step": 112, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_accuracy_recipient_token_verify": 1.0, + "train_accuracy_recipient_token_verify_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.3349802371541502, + "grad_norm": 2.321966400855487, + "learning_rate": 3.886420608016767e-06, + "loss": 0.2771, + "step": 113 + }, + { + "epoch": 0.3349802371541502, + "step": 113, + "train_accuracy_first_token_arguments": 0.9666666666666667, + "train_accuracy_first_token_arguments_total": 30, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_add": 1.0, + "train_accuracy_recipient_token_add_total": 1, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_change": 1.0, + "train_accuracy_recipient_token_change_total": 1, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_replace": 0.5, + "train_accuracy_recipient_token_replace_total": 2, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.3349802371541502, + "step": 113, + "train_accuracy_first_token_arguments": 0.6, + "train_accuracy_first_token_arguments_total": 10, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 2, + "train_accuracy_recipient_token_bird": 1.0, + "train_accuracy_recipient_token_bird_total": 4, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.3349802371541502, + "step": 113, + "train_accuracy_first_token_arguments": 0.5833333333333334, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.7777777777777778, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.33794466403162055, + "grad_norm": 2.028221423810436, + "learning_rate": 3.8663086937793845e-06, + "loss": 0.2435, + "step": 114 + }, + { + "epoch": 0.33794466403162055, + "step": 114, + "train_accuracy_first_token_arguments": 0.9473684210526315, + "train_accuracy_first_token_arguments_total": 114, + "train_accuracy_recipient_token": 0.7368421052631579, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_arch": 0.0, + "train_accuracy_recipient_token_arch_total": 1, + "train_accuracy_recipient_token_building": 1.0, + "train_accuracy_recipient_token_building_total": 1, + "train_accuracy_recipient_token_get": 0.5, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_histor": 0.0, + "train_accuracy_recipient_token_histor_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.33794466403162055, + "step": 114, + "train_accuracy_first_token_arguments": 0.5833333333333334, + "train_accuracy_first_token_arguments_total": 36, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_im": 1.0, + "train_accuracy_recipient_token_im_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.33794466403162055, + "step": 114, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 11, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.3409090909090909, + "grad_norm": 1.8641084259637979, + "learning_rate": 3.846069894620306e-06, + "loss": 0.2402, + "step": 115 + }, + { + "epoch": 0.3409090909090909, + "step": 115, + "train_accuracy_first_token_arguments": 0.75, + "train_accuracy_first_token_arguments_total": 24, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 0.6, + "train_accuracy_recipient_token_find_total": 5, + "train_accuracy_recipient_token_organ": 1.0, + "train_accuracy_recipient_token_organ_total": 1, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.3409090909090909, + "step": 115, + "train_accuracy_first_token_arguments": 0.8181818181818182, + "train_accuracy_first_token_arguments_total": 22, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 0.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.3409090909090909, + "step": 115, + "train_accuracy_first_token_arguments": 0.84, + "train_accuracy_first_token_arguments_total": 25, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_bookmark": 1.0, + "train_accuracy_recipient_token_bookmark_total": 1, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_refresh": 1.0, + "train_accuracy_recipient_token_refresh_total": 1, + "train_accuracy_recipient_token_switch": 0.5, + "train_accuracy_recipient_token_switch_total": 2, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.3438735177865613, + "grad_norm": 2.0312557408429788, + "learning_rate": 3.825706090055088e-06, + "loss": 0.2456, + "step": 116 + }, + { + "epoch": 0.3438735177865613, + "step": 116, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_remove": 1.0, + "train_accuracy_recipient_token_remove_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.3438735177865613, + "step": 116, + "train_accuracy_first_token_arguments": 0.7391304347826086, + "train_accuracy_first_token_arguments_total": 23, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_suggest": 1.0, + "train_accuracy_recipient_token_suggest_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.3438735177865613, + "step": 116, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_create": 1.0, + "train_accuracy_recipient_token_create_total": 2, + "train_accuracy_recipient_token_design": 1.0, + "train_accuracy_recipient_token_design_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.3468379446640316, + "grad_norm": 2.9747145067980383, + "learning_rate": 3.80521917120816e-06, + "loss": 0.3586, + "step": 117 + }, + { + "epoch": 0.3468379446640316, + "step": 117, + "train_accuracy_first_token_arguments": 0.8043478260869565, + "train_accuracy_first_token_arguments_total": 46, + "train_accuracy_recipient_token": 0.9375, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.3468379446640316, + "step": 117, + "train_accuracy_first_token_arguments": 0.8888888888888888, + "train_accuracy_first_token_arguments_total": 27, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_count": 1.0, + "train_accuracy_recipient_token_count_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.3468379446640316, + "step": 117, + "train_accuracy_first_token_arguments": 0.7894736842105263, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.7272727272727273, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_setup": 1.0, + "train_accuracy_recipient_token_setup_total": 2, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.34980237154150196, + "grad_norm": 2.619218782176094, + "learning_rate": 3.784611040637198e-06, + "loss": 0.386, + "step": 118 + }, + { + "epoch": 0.34980237154150196, + "step": 118, + "train_accuracy_first_token_arguments": 0.9285714285714286, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.34980237154150196, + "step": 118, + "train_accuracy_first_token_arguments": 0.9, + "train_accuracy_first_token_arguments_total": 10, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_project": 1.0, + "train_accuracy_recipient_token_project_total": 1, + "train_accuracy_recipient_token_sc": 1.0, + "train_accuracy_recipient_token_sc_total": 2, + "train_accuracy_recipient_token_verify": 1.0, + "train_accuracy_recipient_token_verify_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.34980237154150196, + "step": 118, + "train_accuracy_first_token_arguments": 0.76, + "train_accuracy_first_token_arguments_total": 25, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 6, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.35276679841897235, + "grad_norm": 1.783725464528477, + "learning_rate": 3.7638836121564414e-06, + "loss": 0.3088, + "step": 119 + }, + { + "epoch": 0.35276679841897235, + "step": 119, + "train_accuracy_first_token_arguments": 0.8695652173913043, + "train_accuracy_first_token_arguments_total": 23, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.9, + "train_accuracy_recipient_token_get_total": 10, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.35276679841897235, + "step": 119, + "train_accuracy_first_token_arguments": 0.9090909090909091, + "train_accuracy_first_token_arguments_total": 11, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_financial": 1.0, + "train_accuracy_recipient_token_financial_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_invest": 1.0, + "train_accuracy_recipient_token_invest_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.35276679841897235, + "step": 119, + "train_accuracy_first_token_arguments": 0.967741935483871, + "train_accuracy_first_token_arguments_total": 31, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.3557312252964427, + "grad_norm": 1.880678138487034, + "learning_rate": 3.7430388106589632e-06, + "loss": 0.2326, + "step": 120 + }, + { + "epoch": 0.3557312252964427, + "step": 120, + "train_accuracy_first_token_arguments": 0.7894736842105263, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_detect": 1.0, + "train_accuracy_recipient_token_detect_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.3557312252964427, + "step": 120, + "train_accuracy_first_token_arguments": 0.7916666666666666, + "train_accuracy_first_token_arguments_total": 24, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 2, + "train_accuracy_recipient_token_buy": 1.0, + "train_accuracy_recipient_token_buy_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_suggest": 1.0, + "train_accuracy_recipient_token_suggest_total": 2, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.3557312252964427, + "step": 120, + "train_accuracy_first_token_arguments": 0.8947368421052632, + "train_accuracy_first_token_arguments_total": 57, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 3, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.358695652173913, + "grad_norm": 2.3660230978435117, + "learning_rate": 3.72207857193791e-06, + "loss": 0.2103, + "step": 121 + }, + { + "epoch": 0.358695652173913, + "step": 121, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 9, + "train_accuracy_recipient_token": 0.7647058823529411, + "train_accuracy_recipient_token_all": 0.6, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.358695652173913, + "step": 121, + "train_accuracy_first_token_arguments": 0.7575757575757576, + "train_accuracy_first_token_arguments_total": 33, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_cl": 1.0, + "train_accuracy_recipient_token_cl_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_im": 1.0, + "train_accuracy_recipient_token_im_total": 1, + "train_accuracy_recipient_token_search": 0.5, + "train_accuracy_recipient_token_search_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.358695652173913, + "step": 121, + "train_accuracy_first_token_arguments": 0.8260869565217391, + "train_accuracy_first_token_arguments_total": 23, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_list": 1.0, + "train_accuracy_recipient_token_list_total": 1, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.3616600790513834, + "grad_norm": 1.9095006331871722, + "learning_rate": 3.7010048425067317e-06, + "loss": 0.2754, + "step": 122 + }, + { + "epoch": 0.3616600790513834, + "step": 122, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.3616600790513834, + "step": 122, + "train_accuracy_first_token_arguments": 0.8571428571428571, + "train_accuracy_first_token_arguments_total": 28, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_art": 1.0, + "train_accuracy_recipient_token_art_total": 3, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.3616600790513834, + "step": 122, + "train_accuracy_first_token_arguments": 0.8666666666666667, + "train_accuracy_first_token_arguments_total": 45, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_decode": 1.0, + "train_accuracy_recipient_token_decode_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.36462450592885376, + "grad_norm": 1.6047057072988866, + "learning_rate": 3.679819579418414e-06, + "loss": 0.1743, + "step": 123 + }, + { + "epoch": 0.36462450592885376, + "step": 123, + "train_accuracy_first_token_arguments": 0.71875, + "train_accuracy_first_token_arguments_total": 32, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_arch": 1.0, + "train_accuracy_recipient_token_arch_total": 2, + "train_accuracy_recipient_token_build": 1.0, + "train_accuracy_recipient_token_build_total": 3, + "train_accuracy_recipient_token_histor": 0.0, + "train_accuracy_recipient_token_histor_total": 1, + "train_accuracy_recipient_token_suggest": 1.0, + "train_accuracy_recipient_token_suggest_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.36462450592885376, + "step": 123, + "train_accuracy_first_token_arguments": 0.8, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9166666666666666, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 5, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.36462450592885376, + "step": 123, + "train_accuracy_first_token_arguments": 0.9375, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_land": 1.0, + "train_accuracy_recipient_token_land_total": 4, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.3675889328063241, + "grad_norm": 1.6871352500543018, + "learning_rate": 3.658524750083733e-06, + "loss": 0.3106, + "step": 124 + }, + { + "epoch": 0.3675889328063241, + "step": 124, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 7, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_generate": 1.0, + "train_accuracy_recipient_token_generate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.3675889328063241, + "step": 124, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.3675889328063241, + "step": 124, + "train_accuracy_first_token_arguments": 0.9090909090909091, + "train_accuracy_first_token_arguments_total": 22, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_evaluate": 1.0, + "train_accuracy_recipient_token_evaluate_total": 1, + "train_accuracy_recipient_token_phrase": 1.0, + "train_accuracy_recipient_token_phrase_total": 4, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 1, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.3705533596837945, + "grad_norm": 2.016083076720157, + "learning_rate": 3.6371223320885492e-06, + "loss": 0.3026, + "step": 125 + }, + { + "epoch": 0.3705533596837945, + "step": 125, + "train_accuracy_first_token_arguments": 0.6764705882352942, + "train_accuracy_first_token_arguments_total": 34, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_create": 1.0, + "train_accuracy_recipient_token_create_total": 1, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_plan": 0.0, + "train_accuracy_recipient_token_plan_total": 1, + "train_accuracy_recipient_token_suggest": 1.0, + "train_accuracy_recipient_token_suggest_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.3705533596837945, + "step": 125, + "train_accuracy_first_token_arguments": 0.9090909090909091, + "train_accuracy_first_token_arguments_total": 11, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.3705533596837945, + "step": 125, + "train_accuracy_first_token_arguments": 0.9523809523809523, + "train_accuracy_first_token_arguments_total": 21, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_suggest": 1.0, + "train_accuracy_recipient_token_suggest_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.37351778656126483, + "grad_norm": 2.1413931479434973, + "learning_rate": 3.6156143130101516e-06, + "loss": 0.3208, + "step": 126 + }, + { + "epoch": 0.37351778656126483, + "step": 126, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.8666666666666667, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_organ": 1.0, + "train_accuracy_recipient_token_organ_total": 1, + "train_total_number_recipient_token": 15 + }, + { + "epoch": 0.37351778656126483, + "step": 126, + "train_accuracy_first_token_arguments": 0.9705882352941176, + "train_accuracy_first_token_arguments_total": 34, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.37351778656126483, + "step": 126, + "train_accuracy_first_token_arguments": 0.8, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_count": 1.0, + "train_accuracy_recipient_token_count_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.37648221343873517, + "grad_norm": 2.3478599036765084, + "learning_rate": 3.5940026902326825e-06, + "loss": 0.2732, + "step": 127 + }, + { + "epoch": 0.37648221343873517, + "step": 127, + "train_accuracy_first_token_arguments": 0.782608695652174, + "train_accuracy_first_token_arguments_total": 69, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.37648221343873517, + "step": 127, + "train_accuracy_first_token_arguments": 0.7631578947368421, + "train_accuracy_first_token_arguments_total": 38, + "train_accuracy_recipient_token": 0.7222222222222222, + "train_accuracy_recipient_token_all": 0.7, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 0.5, + "train_accuracy_recipient_token_compare_total": 2, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 0.6666666666666666, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.37648221343873517, + "step": 127, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.8888888888888888, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.3794466403162055, + "grad_norm": 1.8233289275067917, + "learning_rate": 3.5722894707616417e-06, + "loss": 0.2708, + "step": 128 + }, + { + "epoch": 0.3794466403162055, + "step": 128, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.3794466403162055, + "step": 128, + "train_accuracy_first_token_arguments": 0.8095238095238095, + "train_accuracy_first_token_arguments_total": 21, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_clean": 1.0, + "train_accuracy_recipient_token_clean_total": 1, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 3, + "train_accuracy_recipient_token_repair": 1.0, + "train_accuracy_recipient_token_repair_total": 2, + "train_accuracy_recipient_token_restore": 1.0, + "train_accuracy_recipient_token_restore_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.3794466403162055, + "step": 128, + "train_accuracy_first_token_arguments": 0.9722222222222222, + "train_accuracy_first_token_arguments_total": 36, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_book": 0.5, + "train_accuracy_recipient_token_book_total": 2, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_send": 1.0, + "train_accuracy_recipient_token_send_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.3824110671936759, + "grad_norm": 2.7033960410551843, + "learning_rate": 3.550476671037505e-06, + "loss": 0.3598, + "step": 129 + }, + { + "epoch": 0.3824110671936759, + "step": 129, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 22, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.3824110671936759, + "step": 129, + "train_accuracy_first_token_arguments": 0.803921568627451, + "train_accuracy_first_token_arguments_total": 51, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.3824110671936759, + "step": 129, + "train_accuracy_first_token_arguments": 0.7142857142857143, + "train_accuracy_first_token_arguments_total": 7, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_arch": 1.0, + "train_accuracy_recipient_token_arch_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.38537549407114624, + "grad_norm": 1.817103444485219, + "learning_rate": 3.528566316748462e-06, + "loss": 0.2691, + "step": 130 + }, + { + "epoch": 0.38537549407114624, + "step": 130, + "train_accuracy_first_token_arguments": 0.875, + "train_accuracy_first_token_arguments_total": 8, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.38537549407114624, + "step": 130, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.7, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_log": 1.0, + "train_accuracy_recipient_token_log_total": 4, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.38537549407114624, + "step": 130, + "train_accuracy_first_token_arguments": 0.9411764705882353, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_book": 0.5, + "train_accuracy_recipient_token_book_total": 2, + "train_accuracy_recipient_token_find": 0.75, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.3883399209486166, + "grad_norm": 2.3178394841243724, + "learning_rate": 3.5065604426422995e-06, + "loss": 0.2196, + "step": 131 + }, + { + "epoch": 0.3883399209486166, + "step": 131, + "train_accuracy_first_token_arguments": 0.9166666666666666, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.3883399209486166, + "step": 131, + "train_accuracy_first_token_arguments": 0.7580645161290323, + "train_accuracy_first_token_arguments_total": 62, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 0.3333333333333333, + "train_accuracy_recipient_token_analyze_total": 3, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_log": 1.0, + "train_accuracy_recipient_token_log_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.3883399209486166, + "step": 131, + "train_accuracy_first_token_arguments": 0.7368421052631579, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_list": 1.0, + "train_accuracy_recipient_token_list_total": 1, + "train_accuracy_recipient_token_recommend": 0.5, + "train_accuracy_recipient_token_recommend_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.391304347826087, + "grad_norm": 1.7888078327053374, + "learning_rate": 3.484461092337434e-06, + "loss": 0.2452, + "step": 132 + }, + { + "epoch": 0.391304347826087, + "step": 132, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_trans": 1.0, + "train_accuracy_recipient_token_trans_total": 3, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.391304347826087, + "step": 132, + "train_accuracy_first_token_arguments": 0.5897435897435898, + "train_accuracy_first_token_arguments_total": 39, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 0.6666666666666666, + "train_accuracy_recipient_token_analyze_total": 3, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_histor": 1.0, + "train_accuracy_recipient_token_histor_total": 3, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.391304347826087, + "step": 132, + "train_accuracy_first_token_arguments": 0.8947368421052632, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.8571428571428571, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.7777777777777778, + "train_accuracy_recipient_token_get_total": 9, + "train_accuracy_recipient_token_order": 0.0, + "train_accuracy_recipient_token_order_total": 1, + "train_total_number_recipient_token": 21 + }, + { + "epoch": 0.3942687747035573, + "grad_norm": 2.3453355040982307, + "learning_rate": 3.462270318133136e-06, + "loss": 0.273, + "step": 133 + }, + { + "epoch": 0.3942687747035573, + "step": 133, + "train_accuracy_first_token_arguments": 0.75, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_identify": 0.5, + "train_accuracy_recipient_token_identify_total": 2, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.3942687747035573, + "step": 133, + "train_accuracy_first_token_arguments": 0.7894736842105263, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_add": 1.0, + "train_accuracy_recipient_token_add_total": 4, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_construct": 1.0, + "train_accuracy_recipient_token_construct_total": 1, + "train_accuracy_recipient_token_remove": 1.0, + "train_accuracy_recipient_token_remove_total": 1, + "train_accuracy_recipient_token_select": 1.0, + "train_accuracy_recipient_token_select_total": 2, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.3942687747035573, + "step": 133, + "train_accuracy_first_token_arguments": 0.8333333333333334, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 0.5, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.39723320158102765, + "grad_norm": 2.500149372884034, + "learning_rate": 3.4399901808189327e-06, + "loss": 0.2709, + "step": 134 + }, + { + "epoch": 0.39723320158102765, + "step": 134, + "train_accuracy_first_token_arguments": 0.5555555555555556, + "train_accuracy_first_token_arguments_total": 9, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.39723320158102765, + "step": 134, + "train_accuracy_first_token_arguments": 0.8181818181818182, + "train_accuracy_first_token_arguments_total": 11, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compose": 1.0, + "train_accuracy_recipient_token_compose_total": 1, + "train_accuracy_recipient_token_create": 1.0, + "train_accuracy_recipient_token_create_total": 1, + "train_accuracy_recipient_token_generate": 1.0, + "train_accuracy_recipient_token_generate_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.39723320158102765, + "step": 134, + "train_accuracy_first_token_arguments": 0.75, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.8571428571428571, + "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.40019762845849804, + "grad_norm": 2.056275874523913, + "learning_rate": 3.4176227494832305e-06, + "loss": 0.3526, + "step": 135 + }, + { + "epoch": 0.40019762845849804, + "step": 135, + "train_accuracy_first_token_arguments": 0.9230769230769231, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.85, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 0.7777777777777778, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.40019762845849804, + "step": 135, + "train_accuracy_first_token_arguments": 0.8888888888888888, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.40019762845849804, + "step": 135, + "train_accuracy_first_token_arguments": 0.9090909090909091, + "train_accuracy_first_token_arguments_total": 11, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_convert": 1.0, + "train_accuracy_recipient_token_convert_total": 1, + "train_accuracy_recipient_token_get": 0.8333333333333334, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.4031620553359684, + "grad_norm": 1.9432341730217852, + "learning_rate": 3.3951701013211665e-06, + "loss": 0.214, + "step": 136 + }, + { + "epoch": 0.4031620553359684, + "step": 136, + "train_accuracy_first_token_arguments": 0.8, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.6666666666666666, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_locate": 1.0, + "train_accuracy_recipient_token_locate_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.4031620553359684, + "step": 136, + "train_accuracy_first_token_arguments": 0.9130434782608695, + "train_accuracy_first_token_arguments_total": 23, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_invest": 1.0, + "train_accuracy_recipient_token_invest_total": 1, + "train_accuracy_recipient_token_tractor": 1.0, + "train_accuracy_recipient_token_tractor_total": 4, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.4031620553359684, + "step": 136, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 28, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 0.9166666666666666, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_detect": 1.0, + "train_accuracy_recipient_token_detect_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.4061264822134387, + "grad_norm": 1.6348659932120928, + "learning_rate": 3.3726343214417023e-06, + "loss": 0.1342, + "step": 137 + }, + { + "epoch": 0.4061264822134387, + "step": 137, + "train_accuracy_first_token_arguments": 0.8333333333333334, + "train_accuracy_first_token_arguments_total": 24, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_pay": 0.5, + "train_accuracy_recipient_token_pay_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.4061264822134387, + "step": 137, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.7, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_make": 1.0, + "train_accuracy_recipient_token_make_total": 4, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.4061264822134387, + "step": 137, + "train_accuracy_first_token_arguments": 0.9230769230769231, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_count": 1.0, + "train_accuracy_recipient_token_count_total": 1, + "train_accuracy_recipient_token_decode": 1.0, + "train_accuracy_recipient_token_decode_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.4090909090909091, + "grad_norm": 2.0536484514852344, + "learning_rate": 3.3500175026739916e-06, + "loss": 0.279, + "step": 138 + }, + { + "epoch": 0.4090909090909091, + "step": 138, + "train_accuracy_first_token_arguments": 0.7142857142857143, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_bird": 1.0, + "train_accuracy_recipient_token_bird_total": 4, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_in": 1.0, + "train_accuracy_recipient_token_in_total": 3, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.4090909090909091, + "step": 138, + "train_accuracy_first_token_arguments": 0.8048780487804879, + "train_accuracy_first_token_arguments_total": 41, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.4090909090909091, + "step": 138, + "train_accuracy_first_token_arguments": 0.9411764705882353, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.41205533596837945, + "grad_norm": 2.2289267853293144, + "learning_rate": 3.327321745373021e-06, + "loss": 0.3159, + "step": 139 + }, + { + "epoch": 0.41205533596837945, + "step": 139, + "train_accuracy_first_token_arguments": 0.9333333333333333, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.7272727272727273, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_pest": 1.0, + "train_accuracy_recipient_token_pest_total": 1, + "train_accuracy_recipient_token_sun": 1.0, + "train_accuracy_recipient_token_sun_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.41205533596837945, + "step": 139, + "train_accuracy_first_token_arguments": 0.8888888888888888, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_evaluate": 1.0, + "train_accuracy_recipient_token_evaluate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_provide": 1.0, + "train_accuracy_recipient_token_provide_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.41205533596837945, + "step": 139, + "train_accuracy_first_token_arguments": 0.8461538461538461, + "train_accuracy_first_token_arguments_total": 26, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 0.6666666666666666, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.4150197628458498, + "grad_norm": 2.1319285672179005, + "learning_rate": 3.304549157224558e-06, + "loss": 0.3213, + "step": 140 + }, + { + "epoch": 0.4150197628458498, + "step": 140, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 9, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.4150197628458498, + "step": 140, + "train_accuracy_first_token_arguments": 0.8333333333333334, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.4150197628458498, + "step": 140, + "train_accuracy_first_token_arguments": 0.8, + "train_accuracy_first_token_arguments_total": 60, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_explain": 0.5, + "train_accuracy_recipient_token_explain_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_invest": 1.0, + "train_accuracy_recipient_token_invest_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.41798418972332013, + "grad_norm": 2.0722211722052575, + "learning_rate": 3.2817018530494164e-06, + "loss": 0.2898, + "step": 141 + }, + { + "epoch": 0.41798418972332013, + "step": 141, + "train_accuracy_first_token_arguments": 0.7368421052631579, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_locate": 1.0, + "train_accuracy_recipient_token_locate_total": 1, + "train_accuracy_recipient_token_p": 1.0, + "train_accuracy_recipient_token_p_total": 3, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.41798418972332013, + "step": 141, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.9523809523809523, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 2, + "train_accuracy_recipient_token_count": 1.0, + "train_accuracy_recipient_token_count_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_get": 0.8333333333333334, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 21 + }, + { + "epoch": 0.41798418972332013, + "step": 141, + "train_accuracy_first_token_arguments": 0.8656716417910447, + "train_accuracy_first_token_arguments_total": 67, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_count": 1.0, + "train_accuracy_recipient_token_count_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 3, + "train_accuracy_recipient_token_verify": 1.0, + "train_accuracy_recipient_token_verify_total": 2, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.4209486166007905, + "grad_norm": 2.1191079314907766, + "learning_rate": 3.2587819546070596e-06, + "loss": 0.255, + "step": 142 + }, + { + "epoch": 0.4209486166007905, + "step": 142, + "train_accuracy_first_token_arguments": 0.7692307692307693, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.4209486166007905, + "step": 142, + "train_accuracy_first_token_arguments": 0.8333333333333334, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_l": 1.0, + "train_accuracy_recipient_token_l_total": 3, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.4209486166007905, + "step": 142, + "train_accuracy_first_token_arguments": 0.9, + "train_accuracy_first_token_arguments_total": 10, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_event": 0.6666666666666666, + "train_accuracy_recipient_token_event_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.42391304347826086, + "grad_norm": 2.5497967401591204, + "learning_rate": 3.2357915903985605e-06, + "loss": 0.3137, + "step": 143 + }, + { + "epoch": 0.42391304347826086, + "step": 143, + "train_accuracy_first_token_arguments": 0.95, + "train_accuracy_first_token_arguments_total": 40, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_clean": 1.0, + "train_accuracy_recipient_token_clean_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_paint": 1.0, + "train_accuracy_recipient_token_paint_total": 2, + "train_accuracy_recipient_token_remove": 1.0, + "train_accuracy_recipient_token_remove_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.42391304347826086, + "step": 143, + "train_accuracy_first_token_arguments": 0.7619047619047619, + "train_accuracy_first_token_arguments_total": 21, + "train_accuracy_recipient_token": 0.8, + "train_accuracy_recipient_token_all": 0.7, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_decode": 1.0, + "train_accuracy_recipient_token_decode_total": 2, + "train_accuracy_recipient_token_device": 1.0, + "train_accuracy_recipient_token_device_total": 3, + "train_accuracy_recipient_token_extract": 0.0, + "train_accuracy_recipient_token_extract_total": 1, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 2, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.42391304347826086, + "step": 143, + "train_accuracy_first_token_arguments": 0.7924528301886793, + "train_accuracy_first_token_arguments_total": 53, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_order": 1.0, + "train_accuracy_recipient_token_order_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.4268774703557312, + "grad_norm": 1.9145256420639876, + "learning_rate": 3.2127328954689307e-06, + "loss": 0.192, + "step": 144 + }, + { + "epoch": 0.4268774703557312, + "step": 144, + "train_accuracy_first_token_arguments": 0.7391304347826086, + "train_accuracy_first_token_arguments_total": 46, + "train_accuracy_recipient_token": 0.7894736842105263, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_accuracy_recipient_token_translate": 0.75, + "train_accuracy_recipient_token_translate_total": 4, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.4268774703557312, + "step": 144, + "train_accuracy_first_token_arguments": 0.875, + "train_accuracy_first_token_arguments_total": 104, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_optimize": 1.0, + "train_accuracy_recipient_token_optimize_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.4268774703557312, + "step": 144, + "train_accuracy_first_token_arguments": 0.9090909090909091, + "train_accuracy_first_token_arguments_total": 22, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 8, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.4298418972332016, + "grad_norm": 1.6421584423874946, + "learning_rate": 3.1896080112088477e-06, + "loss": 0.2441, + "step": 145 + }, + { + "epoch": 0.4298418972332016, + "step": 145, + "train_accuracy_first_token_arguments": 0.8846153846153846, + "train_accuracy_first_token_arguments_total": 26, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_deploy": 1.0, + "train_accuracy_recipient_token_deploy_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.4298418972332016, + "step": 145, + "train_accuracy_first_token_arguments": 0.8947368421052632, + "train_accuracy_first_token_arguments_total": 38, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_car": 1.0, + "train_accuracy_recipient_token_car_total": 1, + "train_accuracy_recipient_token_evaluate": 1.0, + "train_accuracy_recipient_token_evaluate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.5, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.4298418972332016, + "step": 145, + "train_accuracy_first_token_arguments": 0.8235294117647058, + "train_accuracy_first_token_arguments_total": 34, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_order": 1.0, + "train_accuracy_recipient_token_order_total": 1, + "train_accuracy_recipient_token_schedule": 0.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.43280632411067194, + "grad_norm": 1.9573866581692538, + "learning_rate": 3.166419085155793e-06, + "loss": 0.2593, + "step": 146 + }, + { + "epoch": 0.43280632411067194, + "step": 146, + "train_accuracy_first_token_arguments": 0.7857142857142857, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_list": 1.0, + "train_accuracy_recipient_token_list_total": 1, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 1, + "train_accuracy_recipient_token_stat": 1.0, + "train_accuracy_recipient_token_stat_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.43280632411067194, + "step": 146, + "train_accuracy_first_token_arguments": 0.84, + "train_accuracy_first_token_arguments_total": 25, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.9, + "train_accuracy_recipient_token_get_total": 10, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.43280632411067194, + "step": 146, + "train_accuracy_first_token_arguments": 0.7142857142857143, + "train_accuracy_first_token_arguments_total": 7, + "train_accuracy_recipient_token": 0.9375, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.4357707509881423, + "grad_norm": 2.277891831921691, + "learning_rate": 3.143168270794612e-06, + "loss": 0.4509, + "step": 147 + }, + { + "epoch": 0.4357707509881423, + "step": 147, + "train_accuracy_first_token_arguments": 0.8888888888888888, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_explain": 1.0, + "train_accuracy_recipient_token_explain_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_traffic": 1.0, + "train_accuracy_recipient_token_traffic_total": 2, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.4357707509881423, + "step": 147, + "train_accuracy_first_token_arguments": 0.85, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 0.9047619047619048, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 0.6666666666666666, + "train_accuracy_recipient_token_calculate_total": 3, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 21 + }, + { + "epoch": 0.4357707509881423, + "step": 147, + "train_accuracy_first_token_arguments": 0.9125, + "train_accuracy_first_token_arguments_total": 80, + "train_accuracy_recipient_token": 0.7777777777777778, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_arr": 1.0, + "train_accuracy_recipient_token_arr_total": 1, + "train_accuracy_recipient_token_book": 0.5, + "train_accuracy_recipient_token_book_total": 2, + "train_accuracy_recipient_token_hire": 1.0, + "train_accuracy_recipient_token_hire_total": 1, + "train_accuracy_recipient_token_plan": 1.0, + "train_accuracy_recipient_token_plan_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.43873517786561267, + "grad_norm": 2.085474389214059, + "learning_rate": 3.119857727357527e-06, + "loss": 0.2157, + "step": 148 + }, + { + "epoch": 0.43873517786561267, + "step": 148, + "train_accuracy_first_token_arguments": 0.782608695652174, + "train_accuracy_first_token_arguments_total": 23, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 0.5, + "train_accuracy_recipient_token_calculate_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.43873517786561267, + "step": 148, + "train_accuracy_first_token_arguments": 0.75, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.8333333333333334, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_stamp": 1.0, + "train_accuracy_recipient_token_stamp_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.43873517786561267, + "step": 148, + "train_accuracy_first_token_arguments": 0.9090909090909091, + "train_accuracy_first_token_arguments_total": 11, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_display": 1.0, + "train_accuracy_recipient_token_display_total": 1, + "train_accuracy_recipient_token_find": 0.75, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.441699604743083, + "grad_norm": 1.8828482742789217, + "learning_rate": 3.0964896196236217e-06, + "loss": 0.2592, + "step": 149 + }, + { + "epoch": 0.441699604743083, + "step": 149, + "train_accuracy_first_token_arguments": 0.8095238095238095, + "train_accuracy_first_token_arguments_total": 21, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 3, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.441699604743083, + "step": 149, + "train_accuracy_first_token_arguments": 0.6666666666666666, + "train_accuracy_first_token_arguments_total": 21, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_suggest": 1.0, + "train_accuracy_recipient_token_suggest_total": 2, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.441699604743083, + "step": 149, + "train_accuracy_first_token_arguments": 0.9285714285714286, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.44466403162055335, + "grad_norm": 1.6792776522615755, + "learning_rate": 3.0730661177177957e-06, + "loss": 0.2082, + "step": 150 + }, + { + "epoch": 0.44466403162055335, + "step": 150, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 10, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_discover": 1.0, + "train_accuracy_recipient_token_discover_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.44466403162055335, + "step": 150, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.44466403162055335, + "step": 150, + "train_accuracy_first_token_arguments": 0.9473684210526315, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_stamp": 1.0, + "train_accuracy_recipient_token_stamp_total": 3, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.44762845849802374, + "grad_norm": 2.0438916010331187, + "learning_rate": 3.0495893969092395e-06, + "loss": 0.341, + "step": 151 + }, + { + "epoch": 0.44762845849802374, + "step": 151, + "train_accuracy_first_token_arguments": 0.8888888888888888, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.44762845849802374, + "step": 151, + "train_accuracy_first_token_arguments": 0.9166666666666666, + "train_accuracy_first_token_arguments_total": 24, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_locate": 1.0, + "train_accuracy_recipient_token_locate_total": 1, + "train_accuracy_recipient_token_provide": 1.0, + "train_accuracy_recipient_token_provide_total": 1, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.44762845849802374, + "step": 151, + "train_accuracy_first_token_arguments": 0.8333333333333334, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.8125, + "train_accuracy_recipient_token_all": 0.7272727272727273, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_order": 1.0, + "train_accuracy_recipient_token_order_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.4505928853754941, + "grad_norm": 1.870929622393286, + "learning_rate": 3.0260616374094208e-06, + "loss": 0.2028, + "step": 152 + }, + { + "epoch": 0.4505928853754941, + "step": 152, + "train_accuracy_first_token_arguments": 0.7894736842105263, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_configure": 0.0, + "train_accuracy_recipient_token_configure_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_track": 0.0, + "train_accuracy_recipient_token_track_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.4505928853754941, + "step": 152, + "train_accuracy_first_token_arguments": 0.8636363636363636, + "train_accuracy_first_token_arguments_total": 44, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.4505928853754941, + "step": 152, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_download": 1.0, + "train_accuracy_recipient_token_download_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_rec": 1.0, + "train_accuracy_recipient_token_rec_total": 4, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.4535573122529644, + "grad_norm": 2.1374043090533816, + "learning_rate": 3.0024850241696128e-06, + "loss": 0.1932, + "step": 153 + }, + { + "epoch": 0.4535573122529644, + "step": 153, + "train_accuracy_first_token_arguments": 0.7916666666666666, + "train_accuracy_first_token_arguments_total": 24, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 2, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.4535573122529644, + "step": 153, + "train_accuracy_first_token_arguments": 0.8, + "train_accuracy_first_token_arguments_total": 35, + "train_accuracy_recipient_token": 0.7894736842105263, + "train_accuracy_recipient_token_all": 0.8333333333333334, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_describe": 0.0, + "train_accuracy_recipient_token_describe_total": 1, + "train_accuracy_recipient_token_find": 0.75, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_organ": 1.0, + "train_accuracy_recipient_token_organ_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.4535573122529644, + "step": 153, + "train_accuracy_first_token_arguments": 0.9555555555555556, + "train_accuracy_first_token_arguments_total": 90, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_generate": 1.0, + "train_accuracy_recipient_token_generate_total": 5, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.45652173913043476, + "grad_norm": 1.934815353317823, + "learning_rate": 2.9788617466779884e-06, + "loss": 0.2336, + "step": 154 + }, + { + "epoch": 0.45652173913043476, + "step": 154, + "train_accuracy_first_token_arguments": 0.8205128205128205, + "train_accuracy_first_token_arguments_total": 39, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_f": 1.0, + "train_accuracy_recipient_token_f_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.45652173913043476, + "step": 154, + "train_accuracy_first_token_arguments": 0.8095238095238095, + "train_accuracy_first_token_arguments_total": 21, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_multip": 0.5, + "train_accuracy_recipient_token_multip_total": 4, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.45652173913043476, + "step": 154, + "train_accuracy_first_token_arguments": 0.9130434782608695, + "train_accuracy_first_token_arguments_total": 23, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.875, + "train_accuracy_recipient_token_all_total": 8, + "train_accuracy_recipient_token_check": 0.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_histor": 1.0, + "train_accuracy_recipient_token_histor_total": 2, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.45948616600790515, + "grad_norm": 2.3253564073005224, + "learning_rate": 2.9551939987562866e-06, + "loss": 0.3042, + "step": 155 + }, + { + "epoch": 0.45948616600790515, + "step": 155, + "train_accuracy_first_token_arguments": 0.95, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 0.9166666666666666, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_decode": 1.0, + "train_accuracy_recipient_token_decode_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.45948616600790515, + "step": 155, + "train_accuracy_first_token_arguments": 0.84, + "train_accuracy_first_token_arguments_total": 75, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_estimate": 0.5, + "train_accuracy_recipient_token_estimate_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.45948616600790515, + "step": 155, + "train_accuracy_first_token_arguments": 0.8, + "train_accuracy_first_token_arguments_total": 10, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_translate": 0.0, + "train_accuracy_recipient_token_translate_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.4624505928853755, + "grad_norm": 2.2711295887254184, + "learning_rate": 2.93148397835608e-06, + "loss": 0.2997, + "step": 156 + }, + { + "epoch": 0.4624505928853755, + "step": 156, + "train_accuracy_first_token_arguments": 0.5, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.9375, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.4624505928853755, + "step": 156, + "train_accuracy_first_token_arguments": 0.8333333333333334, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.4624505928853755, + "step": 156, + "train_accuracy_first_token_arguments": 0.8333333333333334, + "train_accuracy_first_token_arguments_total": 30, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 5, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.4654150197628458, + "grad_norm": 2.0895540217432282, + "learning_rate": 2.907733887354657e-06, + "loss": 0.2419, + "step": 157 + }, + { + "epoch": 0.4654150197628458, + "step": 157, + "train_accuracy_first_token_arguments": 0.7837837837837838, + "train_accuracy_first_token_arguments_total": 37, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_sum": 0.0, + "train_accuracy_recipient_token_sum_total": 1, + "train_accuracy_recipient_token_translate": 0.5, + "train_accuracy_recipient_token_translate_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.4654150197628458, + "step": 157, + "train_accuracy_first_token_arguments": 0.7804878048780488, + "train_accuracy_first_token_arguments_total": 41, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 5, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_plan": 1.0, + "train_accuracy_recipient_token_plan_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.4654150197628458, + "step": 157, + "train_accuracy_first_token_arguments": 0.6896551724137931, + "train_accuracy_first_token_arguments_total": 58, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.3333333333333333, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_prior": 1.0, + "train_accuracy_recipient_token_prior_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.4683794466403162, + "grad_norm": 2.203763069427427, + "learning_rate": 2.8839459313505407e-06, + "loss": 0.2426, + "step": 158 + }, + { + "epoch": 0.4683794466403162, + "step": 158, + "train_accuracy_first_token_arguments": 0.9285714285714286, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_recommend": 0.6666666666666666, + "train_accuracy_recipient_token_recommend_total": 3, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.4683794466403162, + "step": 158, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_shop": 1.0, + "train_accuracy_recipient_token_shop_total": 2, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.4683794466403162, + "step": 158, + "train_accuracy_first_token_arguments": 0.9411764705882353, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.47134387351778656, + "grad_norm": 1.9347965319589504, + "learning_rate": 2.8601223194586613e-06, + "loss": 0.2107, + "step": 159 + }, + { + "epoch": 0.47134387351778656, + "step": 159, + "train_accuracy_first_token_arguments": 0.7777777777777778, + "train_accuracy_first_token_arguments_total": 27, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.47134387351778656, + "step": 159, + "train_accuracy_first_token_arguments": 0.9375, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_invest": 1.0, + "train_accuracy_recipient_token_invest_total": 1, + "train_accuracy_recipient_token_vehicle": 1.0, + "train_accuracy_recipient_token_vehicle_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.47134387351778656, + "step": 159, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.4743083003952569, + "grad_norm": 1.9897698233644936, + "learning_rate": 2.8362652641052024e-06, + "loss": 0.1806, + "step": 160 + }, + { + "epoch": 0.4743083003952569, + "step": 160, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.4743083003952569, + "step": 160, + "train_accuracy_first_token_arguments": 0.875, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_art": 1.0, + "train_accuracy_recipient_token_art_total": 3, + "train_accuracy_recipient_token_artist": 0.0, + "train_accuracy_recipient_token_artist_total": 1, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 1, + "train_accuracy_recipient_token_water": 1.0, + "train_accuracy_recipient_token_water_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.4743083003952569, + "step": 160, + "train_accuracy_first_token_arguments": 0.9130434782608695, + "train_accuracy_first_token_arguments_total": 23, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_schedule": 0.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.4772727272727273, + "grad_norm": 1.869456628968754, + "learning_rate": 2.8123769808221407e-06, + "loss": 0.1735, + "step": 161 + }, + { + "epoch": 0.4772727272727273, + "step": 161, + "train_accuracy_first_token_arguments": 0.6521739130434783, + "train_accuracy_first_token_arguments_total": 23, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.4772727272727273, + "step": 161, + "train_accuracy_first_token_arguments": 0.9166666666666666, + "train_accuracy_first_token_arguments_total": 36, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.4772727272727273, + "step": 161, + "train_accuracy_first_token_arguments": 0.8333333333333334, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.9375, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 2, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.48023715415019763, + "grad_norm": 2.502634515773549, + "learning_rate": 2.788459688041495e-06, + "loss": 0.2973, + "step": 162 + }, + { + "epoch": 0.48023715415019763, + "step": 162, + "train_accuracy_first_token_arguments": 0.9375, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 0.8571428571428571, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_detect": 1.0, + "train_accuracy_recipient_token_detect_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_translate": 0.5, + "train_accuracy_recipient_token_translate_total": 2, + "train_total_number_recipient_token": 21 + }, + { + "epoch": 0.48023715415019763, + "step": 162, + "train_accuracy_first_token_arguments": 0.9473684210526315, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 0.5, + "train_accuracy_recipient_token_calculate_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.48023715415019763, + "step": 162, + "train_accuracy_first_token_arguments": 0.9285714285714286, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_list": 1.0, + "train_accuracy_recipient_token_list_total": 5, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.48320158102766797, + "grad_norm": 2.341420807222338, + "learning_rate": 2.7645156068893075e-06, + "loss": 0.2014, + "step": 163 + }, + { + "epoch": 0.48320158102766797, + "step": 163, + "train_accuracy_first_token_arguments": 0.8333333333333334, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.8333333333333334, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 3, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.48320158102766797, + "step": 163, + "train_accuracy_first_token_arguments": 0.4782608695652174, + "train_accuracy_first_token_arguments_total": 46, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_color": 1.0, + "train_accuracy_recipient_token_color_total": 1, + "train_accuracy_recipient_token_draw": 1.0, + "train_accuracy_recipient_token_draw_total": 1, + "train_accuracy_recipient_token_object": 1.0, + "train_accuracy_recipient_token_object_total": 1, + "train_accuracy_recipient_token_scene": 1.0, + "train_accuracy_recipient_token_scene_total": 2, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.48320158102766797, + "step": 163, + "train_accuracy_first_token_arguments": 0.9090909090909091, + "train_accuracy_first_token_arguments_total": 11, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.48616600790513836, + "grad_norm": 1.6883872990934619, + "learning_rate": 2.7405469609793746e-06, + "loss": 0.2554, + "step": 164 + }, + { + "epoch": 0.48616600790513836, + "step": 164, + "train_accuracy_first_token_arguments": 0.7407407407407407, + "train_accuracy_first_token_arguments_total": 81, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_inspect": 1.0, + "train_accuracy_recipient_token_inspect_total": 4, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.48616600790513836, + "step": 164, + "train_accuracy_first_token_arguments": 0.8947368421052632, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_get": 0.8333333333333334, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_review": 1.0, + "train_accuracy_recipient_token_review_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.48616600790513836, + "step": 164, + "train_accuracy_first_token_arguments": 0.7142857142857143, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_learn": 1.0, + "train_accuracy_recipient_token_learn_total": 1, + "train_accuracy_recipient_token_research": 1.0, + "train_accuracy_recipient_token_research_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.4891304347826087, + "grad_norm": 1.7224600166430244, + "learning_rate": 2.716555976206748e-06, + "loss": 0.2725, + "step": 165 + }, + { + "epoch": 0.4891304347826087, + "step": 165, + "train_accuracy_first_token_arguments": 0.6, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 0.9333333333333333, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_stat": 1.0, + "train_accuracy_recipient_token_stat_total": 1, + "train_accuracy_recipient_token_suggest": 1.0, + "train_accuracy_recipient_token_suggest_total": 1, + "train_total_number_recipient_token": 15 + }, + { + "epoch": 0.4891304347826087, + "step": 165, + "train_accuracy_first_token_arguments": 0.9166666666666666, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_st": 1.0, + "train_accuracy_recipient_token_st_total": 3, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.4891304347826087, + "step": 165, + "train_accuracy_first_token_arguments": 0.6029411764705882, + "train_accuracy_first_token_arguments_total": 68, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_count": 1.0, + "train_accuracy_recipient_token_count_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 2, + "train_accuracy_recipient_token_flag": 1.0, + "train_accuracy_recipient_token_flag_total": 1, + "train_accuracy_recipient_token_pl": 1.0, + "train_accuracy_recipient_token_pl_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.49209486166007904, + "grad_norm": 2.488668771406232, + "learning_rate": 2.6925448805410197e-06, + "loss": 0.4772, + "step": 166 + }, + { + "epoch": 0.49209486166007904, + "step": 166, + "train_accuracy_first_token_arguments": 0.6153846153846154, + "train_accuracy_first_token_arguments_total": 26, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.49209486166007904, + "step": 166, + "train_accuracy_first_token_arguments": 0.96875, + "train_accuracy_first_token_arguments_total": 32, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_create": 1.0, + "train_accuracy_recipient_token_create_total": 2, + "train_accuracy_recipient_token_design": 1.0, + "train_accuracy_recipient_token_design_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.49209486166007904, + "step": 166, + "train_accuracy_first_token_arguments": 0.8333333333333334, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.4950592885375494, + "grad_norm": 2.6791883381737107, + "learning_rate": 2.6685159038194202e-06, + "loss": 0.2803, + "step": 167 + }, + { + "epoch": 0.4950592885375494, + "step": 167, + "train_accuracy_first_token_arguments": 0.8333333333333334, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.4950592885375494, + "step": 167, + "train_accuracy_first_token_arguments": 0.7083333333333334, + "train_accuracy_first_token_arguments_total": 24, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 2, + "train_accuracy_recipient_token_compile": 1.0, + "train_accuracy_recipient_token_compile_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.4950592885375494, + "step": 167, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 9, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.4980237154150198, + "grad_norm": 2.068138976829196, + "learning_rate": 2.6444712775397397e-06, + "loss": 0.3013, + "step": 168 + }, + { + "epoch": 0.4980237154150198, + "step": 168, + "train_accuracy_first_token_arguments": 0.6559139784946236, + "train_accuracy_first_token_arguments_total": 93, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.4980237154150198, + "step": 168, + "train_accuracy_first_token_arguments": 0.9782608695652174, + "train_accuracy_first_token_arguments_total": 92, + "train_accuracy_recipient_token": 0.75, + "train_accuracy_recipient_token_all": 0.6666666666666666, + "train_accuracy_recipient_token_all_total": 9, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 12 + }, + { + "epoch": 0.4980237154150198, + "step": 168, + "train_accuracy_first_token_arguments": 0.75, + "train_accuracy_first_token_arguments_total": 8, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5009881422924901, + "grad_norm": 2.097242382871384, + "learning_rate": 2.6204132346530936e-06, + "loss": 0.3079, + "step": 169 + }, + { + "epoch": 0.5009881422924901, + "step": 169, + "train_accuracy_first_token_arguments": 0.875, + "train_accuracy_first_token_arguments_total": 32, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.5009881422924901, + "step": 169, + "train_accuracy_first_token_arguments": 0.6764705882352942, + "train_accuracy_first_token_arguments_total": 34, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_camera": 1.0, + "train_accuracy_recipient_token_camera_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_invest": 1.0, + "train_accuracy_recipient_token_invest_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5009881422924901, + "step": 169, + "train_accuracy_first_token_arguments": 0.7692307692307693, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.5039525691699605, + "grad_norm": 1.961726304062012, + "learning_rate": 2.5963440093565567e-06, + "loss": 0.3133, + "step": 170 + }, + { + "epoch": 0.5039525691699605, + "step": 170, + "train_accuracy_first_token_arguments": 0.8461538461538461, + "train_accuracy_first_token_arguments_total": 26, + "train_accuracy_recipient_token": 0.6666666666666666, + "train_accuracy_recipient_token_all": 0.6, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 0.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_multiply": 0.75, + "train_accuracy_recipient_token_multiply_total": 4, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5039525691699605, + "step": 170, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5039525691699605, + "step": 170, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 11, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_extract": 1.0, + "train_accuracy_recipient_token_extract_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_recogn": 1.0, + "train_accuracy_recipient_token_recogn_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5069169960474308, + "grad_norm": 2.0506442694884104, + "learning_rate": 2.572265836885682e-06, + "loss": 0.2265, + "step": 171 + }, + { + "epoch": 0.5069169960474308, + "step": 171, + "train_accuracy_first_token_arguments": 0.9, + "train_accuracy_first_token_arguments_total": 30, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 10, + "train_total_number_recipient_token": 21 + }, + { + "epoch": 0.5069169960474308, + "step": 171, + "train_accuracy_first_token_arguments": 0.8125, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5069169960474308, + "step": 171, + "train_accuracy_first_token_arguments": 0.8888888888888888, + "train_accuracy_first_token_arguments_total": 9, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 5, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5098814229249012, + "grad_norm": 2.4101209365627443, + "learning_rate": 2.548180953306918e-06, + "loss": 0.2901, + "step": 172 + }, + { + "epoch": 0.5098814229249012, + "step": 172, + "train_accuracy_first_token_arguments": 0.9, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 0.8666666666666667, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 15 + }, + { + "epoch": 0.5098814229249012, + "step": 172, + "train_accuracy_first_token_arguments": 0.7678571428571429, + "train_accuracy_first_token_arguments_total": 56, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_histor": 0.5, + "train_accuracy_recipient_token_histor_total": 2, + "train_accuracy_recipient_token_organ": 1.0, + "train_accuracy_recipient_token_organ_total": 1, + "train_accuracy_recipient_token_public": 1.0, + "train_accuracy_recipient_token_public_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.5098814229249012, + "step": 172, + "train_accuracy_first_token_arguments": 0.5714285714285714, + "train_accuracy_first_token_arguments_total": 21, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.7272727272727273, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 1, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5128458498023716, + "grad_norm": 2.403125205150021, + "learning_rate": 2.524091595309952e-06, + "loss": 0.2281, + "step": 173 + }, + { + "epoch": 0.5128458498023716, + "step": 173, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 11, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_plant": 1.0, + "train_accuracy_recipient_token_plant_total": 3, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5128458498023716, + "step": 173, + "train_accuracy_first_token_arguments": 0.9285714285714286, + "train_accuracy_first_token_arguments_total": 28, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_evaluate": 1.0, + "train_accuracy_recipient_token_evaluate_total": 1, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_plan": 1.0, + "train_accuracy_recipient_token_plan_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.5128458498023716, + "step": 173, + "train_accuracy_first_token_arguments": 0.8947368421052632, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 5, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_update": 1.0, + "train_accuracy_recipient_token_update_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5158102766798419, + "grad_norm": 1.7135004510273728, + "learning_rate": 2.5e-06, + "loss": 0.1725, + "step": 174 + }, + { + "epoch": 0.5158102766798419, + "step": 174, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 7, + "train_accuracy_recipient_token": 0.9333333333333333, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_int": 1.0, + "train_accuracy_recipient_token_int_total": 1, + "train_total_number_recipient_token": 15 + }, + { + "epoch": 0.5158102766798419, + "step": 174, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.5158102766798419, + "step": 174, + "train_accuracy_first_token_arguments": 0.75, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 0.875, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.5187747035573123, + "grad_norm": 2.193026807831911, + "learning_rate": 2.475908404690049e-06, + "loss": 0.4577, + "step": 175 + }, + { + "epoch": 0.5187747035573123, + "step": 175, + "train_accuracy_first_token_arguments": 0.9090909090909091, + "train_accuracy_first_token_arguments_total": 22, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_ca": 1.0, + "train_accuracy_recipient_token_ca_total": 4, + "train_accuracy_recipient_token_decode": 1.0, + "train_accuracy_recipient_token_decode_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.5187747035573123, + "step": 175, + "train_accuracy_first_token_arguments": 0.7058823529411765, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_art": 1.0, + "train_accuracy_recipient_token_art_total": 2, + "train_accuracy_recipient_token_decode": 1.0, + "train_accuracy_recipient_token_decode_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.5187747035573123, + "step": 175, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 24, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.5217391304347826, + "grad_norm": 2.046983156954581, + "learning_rate": 2.4518190466930837e-06, + "loss": 0.4614, + "step": 176 + }, + { + "epoch": 0.5217391304347826, + "step": 176, + "train_accuracy_first_token_arguments": 0.8235294117647058, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 0.8571428571428571, + "train_accuracy_recipient_token_calculate_total": 7, + "train_accuracy_recipient_token_count": 1.0, + "train_accuracy_recipient_token_count_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.5217391304347826, + "step": 176, + "train_accuracy_first_token_arguments": 0.8666666666666667, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_create": 1.0, + "train_accuracy_recipient_token_create_total": 1, + "train_accuracy_recipient_token_draw": 1.0, + "train_accuracy_recipient_token_draw_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.5217391304347826, + "step": 176, + "train_accuracy_first_token_arguments": 0.9444444444444444, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_phone": 1.0, + "train_accuracy_recipient_token_phone_total": 2, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.5247035573122529, + "grad_norm": 2.7763963997279046, + "learning_rate": 2.427734163114319e-06, + "loss": 0.3222, + "step": 177 + }, + { + "epoch": 0.5247035573122529, + "step": 177, + "train_accuracy_first_token_arguments": 0.9473684210526315, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.8333333333333334, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_generate": 0.5, + "train_accuracy_recipient_token_generate_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.5247035573122529, + "step": 177, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 0.7647058823529411, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 0.5, + "train_accuracy_recipient_token_check_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_sum": 1.0, + "train_accuracy_recipient_token_sum_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.5247035573122529, + "step": 177, + "train_accuracy_first_token_arguments": 0.8823529411764706, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compress": 1.0, + "train_accuracy_recipient_token_compress_total": 1, + "train_accuracy_recipient_token_crop": 1.0, + "train_accuracy_recipient_token_crop_total": 1, + "train_accuracy_recipient_token_d": 1.0, + "train_accuracy_recipient_token_d_total": 2, + "train_accuracy_recipient_token_enh": 0.0, + "train_accuracy_recipient_token_enh_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.5276679841897233, + "grad_norm": 2.245421899066183, + "learning_rate": 2.403655990643444e-06, + "loss": 0.3072, + "step": 178 + }, + { + "epoch": 0.5276679841897233, + "step": 178, + "train_accuracy_first_token_arguments": 0.875, + "train_accuracy_first_token_arguments_total": 24, + "train_accuracy_recipient_token": 0.7894736842105263, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_customize": 0.5, + "train_accuracy_recipient_token_customize_total": 4, + "train_accuracy_recipient_token_get": 0.6, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.5276679841897233, + "step": 178, + "train_accuracy_first_token_arguments": 0.8333333333333334, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.7777777777777778, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 0.5714285714285714, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5276679841897233, + "step": 178, + "train_accuracy_first_token_arguments": 0.8823529411764706, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5306324110671937, + "grad_norm": 2.405264876175068, + "learning_rate": 2.3795867653469072e-06, + "loss": 0.344, + "step": 179 + }, + { + "epoch": 0.5306324110671937, + "step": 179, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 9, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 5, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 10 + }, + { + "epoch": 0.5306324110671937, + "step": 179, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 0.7142857142857143, + "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.5306324110671937, + "step": 179, + "train_accuracy_first_token_arguments": 0.7954545454545454, + "train_accuracy_first_token_arguments_total": 44, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_describe": 1.0, + "train_accuracy_recipient_token_describe_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.5335968379446641, + "grad_norm": 5.342437189948287, + "learning_rate": 2.355528722460261e-06, + "loss": 0.7124, + "step": 180 + }, + { + "epoch": 0.5335968379446641, + "step": 180, + "train_accuracy_first_token_arguments": 0.8666666666666667, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 0.8571428571428571, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5335968379446641, + "step": 180, + "train_accuracy_first_token_arguments": 0.8461538461538461, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_explain": 1.0, + "train_accuracy_recipient_token_explain_total": 3, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5335968379446641, + "step": 180, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 22, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_fish": 1.0, + "train_accuracy_recipient_token_fish_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_purchase": 1.0, + "train_accuracy_recipient_token_purchase_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.5365612648221344, + "grad_norm": 2.1777035417905934, + "learning_rate": 2.3314840961805806e-06, + "loss": 0.2801, + "step": 181 + }, + { + "epoch": 0.5365612648221344, + "step": 181, + "train_accuracy_first_token_arguments": 0.75, + "train_accuracy_first_token_arguments_total": 36, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 10, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.5365612648221344, + "step": 181, + "train_accuracy_first_token_arguments": 0.8, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.5365612648221344, + "step": 181, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 8, + "train_accuracy_recipient_token": 0.875, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 2, + "train_accuracy_recipient_token_extract": 1.0, + "train_accuracy_recipient_token_extract_total": 1, + "train_accuracy_recipient_token_monitor": 0.0, + "train_accuracy_recipient_token_monitor_total": 1, + "train_accuracy_recipient_token_recogn": 0.0, + "train_accuracy_recipient_token_recogn_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.5395256916996047, + "grad_norm": 1.9844826167678842, + "learning_rate": 2.3074551194589816e-06, + "loss": 0.3629, + "step": 182 + }, + { + "epoch": 0.5395256916996047, + "step": 182, + "train_accuracy_first_token_arguments": 0.9019607843137255, + "train_accuracy_first_token_arguments_total": 51, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_notify": 0.0, + "train_accuracy_recipient_token_notify_total": 1, + "train_accuracy_recipient_token_set": 1.0, + "train_accuracy_recipient_token_set_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.5395256916996047, + "step": 182, + "train_accuracy_first_token_arguments": 0.9090909090909091, + "train_accuracy_first_token_arguments_total": 22, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_sponsor": 1.0, + "train_accuracy_recipient_token_sponsor_total": 1, + "train_accuracy_recipient_token_vol": 1.0, + "train_accuracy_recipient_token_vol_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5395256916996047, + "step": 182, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 10, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_book": 0.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.5424901185770751, + "grad_norm": 2.429551580616199, + "learning_rate": 2.2834440237932537e-06, + "loss": 0.3087, + "step": 183 + }, + { + "epoch": 0.5424901185770751, + "step": 183, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_order": 0.0, + "train_accuracy_recipient_token_order_total": 1, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.5424901185770751, + "step": 183, + "train_accuracy_first_token_arguments": 0.9354838709677419, + "train_accuracy_first_token_arguments_total": 31, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 3, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.5424901185770751, + "step": 183, + "train_accuracy_first_token_arguments": 0.875, + "train_accuracy_first_token_arguments_total": 8, + "train_accuracy_recipient_token": 0.8, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 0.6666666666666666, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 0.0, + "train_accuracy_recipient_token_get_total": 1, + "train_total_number_recipient_token": 15 + }, + { + "epoch": 0.5454545454545454, + "grad_norm": 2.062437887240769, + "learning_rate": 2.259453039020626e-06, + "loss": 0.2448, + "step": 184 + }, + { + "epoch": 0.5454545454545454, + "step": 184, + "train_accuracy_first_token_arguments": 0.6976744186046512, + "train_accuracy_first_token_arguments_total": 43, + "train_accuracy_recipient_token": 0.9375, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_log": 1.0, + "train_accuracy_recipient_token_log_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.5454545454545454, + "step": 184, + "train_accuracy_first_token_arguments": 0.6153846153846154, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_interpret": 1.0, + "train_accuracy_recipient_token_interpret_total": 1, + "train_accuracy_recipient_token_stage": 1.0, + "train_accuracy_recipient_token_stage_total": 4, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.5454545454545454, + "step": 184, + "train_accuracy_first_token_arguments": 0.8846153846153846, + "train_accuracy_first_token_arguments_total": 26, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5484189723320159, + "grad_norm": 1.7733764406317996, + "learning_rate": 2.2354843931106933e-06, + "loss": 0.2303, + "step": 185 + }, + { + "epoch": 0.5484189723320159, + "step": 185, + "train_accuracy_first_token_arguments": 0.8918918918918919, + "train_accuracy_first_token_arguments_total": 37, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9166666666666666, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_count": 1.0, + "train_accuracy_recipient_token_count_total": 3, + "train_accuracy_recipient_token_estimate": 0.5, + "train_accuracy_recipient_token_estimate_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.5484189723320159, + "step": 185, + "train_accuracy_first_token_arguments": 0.6, + "train_accuracy_first_token_arguments_total": 55, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 0.5, + "train_accuracy_recipient_token_check_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_update": 1.0, + "train_accuracy_recipient_token_update_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.5484189723320159, + "step": 185, + "train_accuracy_first_token_arguments": 0.7352941176470589, + "train_accuracy_first_token_arguments_total": 34, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5513833992094862, + "grad_norm": 1.7231295422156472, + "learning_rate": 2.211540311958506e-06, + "loss": 0.2471, + "step": 186 + }, + { + "epoch": 0.5513833992094862, + "step": 186, + "train_accuracy_first_token_arguments": 0.8571428571428571, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.5513833992094862, + "step": 186, + "train_accuracy_first_token_arguments": 0.9230769230769231, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_art": 1.0, + "train_accuracy_recipient_token_art_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.5513833992094862, + "step": 186, + "train_accuracy_first_token_arguments": 0.6712328767123288, + "train_accuracy_first_token_arguments_total": 73, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_describe": 1.0, + "train_accuracy_recipient_token_describe_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.5543478260869565, + "grad_norm": 1.9589966620887675, + "learning_rate": 2.1876230191778598e-06, + "loss": 0.3246, + "step": 187 + }, + { + "epoch": 0.5543478260869565, + "step": 187, + "train_accuracy_first_token_arguments": 0.8260869565217391, + "train_accuracy_first_token_arguments_total": 23, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5543478260869565, + "step": 187, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 32, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.7272727272727273, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_convert": 1.0, + "train_accuracy_recipient_token_convert_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.5543478260869565, + "step": 187, + "train_accuracy_first_token_arguments": 0.75, + "train_accuracy_first_token_arguments_total": 8, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_decode": 1.0, + "train_accuracy_recipient_token_decode_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5573122529644269, + "grad_norm": 2.2200460526334824, + "learning_rate": 2.1637347358947984e-06, + "loss": 0.216, + "step": 188 + }, + { + "epoch": 0.5573122529644269, + "step": 188, + "train_accuracy_first_token_arguments": 0.95, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_explain": 1.0, + "train_accuracy_recipient_token_explain_total": 1, + "train_accuracy_recipient_token_gesture": 1.0, + "train_accuracy_recipient_token_gesture_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.5573122529644269, + "step": 188, + "train_accuracy_first_token_arguments": 0.9473684210526315, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8461538461538461, + "train_accuracy_recipient_token_all_total": 13, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5573122529644269, + "step": 188, + "train_accuracy_first_token_arguments": 0.75, + "train_accuracy_first_token_arguments_total": 24, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_add": 1.0, + "train_accuracy_recipient_token_add_total": 1, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_list": 1.0, + "train_accuracy_recipient_token_list_total": 1, + "train_accuracy_recipient_token_update": 1.0, + "train_accuracy_recipient_token_update_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5602766798418972, + "grad_norm": 2.440285593927014, + "learning_rate": 2.13987768054134e-06, + "loss": 0.3342, + "step": 189 + }, + { + "epoch": 0.5602766798418972, + "step": 189, + "train_accuracy_first_token_arguments": 0.9130434782608695, + "train_accuracy_first_token_arguments_total": 23, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_product": 0.75, + "train_accuracy_recipient_token_product_total": 4, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.5602766798418972, + "step": 189, + "train_accuracy_first_token_arguments": 0.9886363636363636, + "train_accuracy_first_token_arguments_total": 88, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 4, + "train_accuracy_recipient_token_hang": 1.0, + "train_accuracy_recipient_token_hang_total": 1, + "train_accuracy_recipient_token_re": 0.0, + "train_accuracy_recipient_token_re_total": 1, + "train_accuracy_recipient_token_remove": 1.0, + "train_accuracy_recipient_token_remove_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.5602766798418972, + "step": 189, + "train_accuracy_first_token_arguments": 0.7391304347826086, + "train_accuracy_first_token_arguments_total": 23, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5632411067193676, + "grad_norm": 2.2639594853684795, + "learning_rate": 2.1160540686494597e-06, + "loss": 0.2142, + "step": 190 + }, + { + "epoch": 0.5632411067193676, + "step": 190, + "train_accuracy_first_token_arguments": 0.7857142857142857, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5632411067193676, + "step": 190, + "train_accuracy_first_token_arguments": 0.9166666666666666, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5632411067193676, + "step": 190, + "train_accuracy_first_token_arguments": 0.7777777777777778, + "train_accuracy_first_token_arguments_total": 36, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 3, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5662055335968379, + "grad_norm": 3.2075927432234157, + "learning_rate": 2.0922661126453436e-06, + "loss": 0.4145, + "step": 191 + }, + { + "epoch": 0.5662055335968379, + "step": 191, + "train_accuracy_first_token_arguments": 0.7166666666666667, + "train_accuracy_first_token_arguments_total": 60, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compose": 1.0, + "train_accuracy_recipient_token_compose_total": 1, + "train_accuracy_recipient_token_create": 1.0, + "train_accuracy_recipient_token_create_total": 2, + "train_accuracy_recipient_token_generate": 1.0, + "train_accuracy_recipient_token_generate_total": 1, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.5662055335968379, + "step": 191, + "train_accuracy_first_token_arguments": 0.6385542168674698, + "train_accuracy_first_token_arguments_total": 83, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 2, + "train_accuracy_recipient_token_generate": 1.0, + "train_accuracy_recipient_token_generate_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.5662055335968379, + "step": 191, + "train_accuracy_first_token_arguments": 0.7121212121212122, + "train_accuracy_first_token_arguments_total": 66, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 2, + "train_accuracy_recipient_token_draw": 1.0, + "train_accuracy_recipient_token_draw_total": 4, + "train_accuracy_recipient_token_suggest": 1.0, + "train_accuracy_recipient_token_suggest_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.5691699604743083, + "grad_norm": 1.9745905408889657, + "learning_rate": 2.0685160216439205e-06, + "loss": 0.2894, + "step": 192 + }, + { + "epoch": 0.5691699604743083, + "step": 192, + "train_accuracy_first_token_arguments": 0.7647058823529411, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.875, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.8333333333333334, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.5691699604743083, + "step": 192, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 53, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 5, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.5691699604743083, + "step": 192, + "train_accuracy_first_token_arguments": 0.9545454545454546, + "train_accuracy_first_token_arguments_total": 22, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8333333333333334, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.5721343873517787, + "grad_norm": 2.517506728021728, + "learning_rate": 2.044806001243714e-06, + "loss": 0.2701, + "step": 193 + }, + { + "epoch": 0.5721343873517787, + "step": 193, + "train_accuracy_first_token_arguments": 0.6296296296296297, + "train_accuracy_first_token_arguments_total": 27, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_g": 1.0, + "train_accuracy_recipient_token_g_total": 3, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5721343873517787, + "step": 193, + "train_accuracy_first_token_arguments": 0.75, + "train_accuracy_first_token_arguments_total": 8, + "train_accuracy_recipient_token": 0.9333333333333333, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_create": 1.0, + "train_accuracy_recipient_token_create_total": 1, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 15 + }, + { + "epoch": 0.5721343873517787, + "step": 193, + "train_accuracy_first_token_arguments": 0.5471698113207547, + "train_accuracy_first_token_arguments_total": 53, + "train_accuracy_recipient_token": 0.875, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 2, + "train_accuracy_recipient_token_energy": 1.0, + "train_accuracy_recipient_token_energy_total": 1, + "train_accuracy_recipient_token_landscape": 1.0, + "train_accuracy_recipient_token_landscape_total": 1, + "train_accuracy_recipient_token_suggest": 1.0, + "train_accuracy_recipient_token_suggest_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.575098814229249, + "grad_norm": 2.477613864495461, + "learning_rate": 2.021138253322012e-06, + "loss": 0.3818, + "step": 194 + }, + { + "epoch": 0.575098814229249, + "step": 194, + "train_accuracy_first_token_arguments": 0.967741935483871, + "train_accuracy_first_token_arguments_total": 31, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_accuracy_recipient_token_sort": 1.0, + "train_accuracy_recipient_token_sort_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.575098814229249, + "step": 194, + "train_accuracy_first_token_arguments": 0.9705882352941176, + "train_accuracy_first_token_arguments_total": 34, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_order": 1.0, + "train_accuracy_recipient_token_order_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.575098814229249, + "step": 194, + "train_accuracy_first_token_arguments": 0.7368421052631579, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.5780632411067194, + "grad_norm": 1.7625337382807924, + "learning_rate": 1.9975149758303885e-06, + "loss": 0.148, + "step": 195 + }, + { + "epoch": 0.5780632411067194, + "step": 195, + "train_accuracy_first_token_arguments": 0.8055555555555556, + "train_accuracy_first_token_arguments_total": 36, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.5780632411067194, + "step": 195, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 23, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.5780632411067194, + "step": 195, + "train_accuracy_first_token_arguments": 0.9523809523809523, + "train_accuracy_first_token_arguments_total": 21, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_list": 1.0, + "train_accuracy_recipient_token_list_total": 1, + "train_accuracy_recipient_token_stat": 1.0, + "train_accuracy_recipient_token_stat_total": 2, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.5810276679841897, + "grad_norm": 1.8753430945988403, + "learning_rate": 1.97393836259058e-06, + "loss": 0.2221, + "step": 196 + }, + { + "epoch": 0.5810276679841897, + "step": 196, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 28, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.5810276679841897, + "step": 196, + "train_accuracy_first_token_arguments": 0.8367346938775511, + "train_accuracy_first_token_arguments_total": 49, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.5810276679841897, + "step": 196, + "train_accuracy_first_token_arguments": 0.9230769230769231, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.8571428571428571, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_monitor": 1.0, + "train_accuracy_recipient_token_monitor_total": 1, + "train_total_number_recipient_token": 14 + }, + { + "epoch": 0.58399209486166, + "grad_norm": 2.1633104244151853, + "learning_rate": 1.9504106030907605e-06, + "loss": 0.2133, + "step": 197 + }, + { + "epoch": 0.58399209486166, + "step": 197, + "train_accuracy_first_token_arguments": 0.9090909090909091, + "train_accuracy_first_token_arguments_total": 22, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_construction": 1.0, + "train_accuracy_recipient_token_construction_total": 2, + "train_accuracy_recipient_token_count": 1.0, + "train_accuracy_recipient_token_count_total": 5, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.58399209486166, + "step": 197, + "train_accuracy_first_token_arguments": 0.9411764705882353, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.58399209486166, + "step": 197, + "train_accuracy_first_token_arguments": 0.6527777777777778, + "train_accuracy_first_token_arguments_total": 72, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.5869565217391305, + "grad_norm": 2.0288710387741813, + "learning_rate": 1.9269338822822047e-06, + "loss": 0.1518, + "step": 198 + }, + { + "epoch": 0.5869565217391305, + "step": 198, + "train_accuracy_first_token_arguments": 0.8235294117647058, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 5, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_make": 1.0, + "train_accuracy_recipient_token_make_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.5869565217391305, + "step": 198, + "train_accuracy_first_token_arguments": 0.7777777777777778, + "train_accuracy_first_token_arguments_total": 27, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5869565217391305, + "step": 198, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 0.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_compile": 1.0, + "train_accuracy_recipient_token_compile_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.5899209486166008, + "grad_norm": 2.090710210779781, + "learning_rate": 1.9035103803763793e-06, + "loss": 0.1552, + "step": 199 + }, + { + "epoch": 0.5899209486166008, + "step": 199, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 21, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5899209486166008, + "step": 199, + "train_accuracy_first_token_arguments": 0.8571428571428571, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 1, + "train_accuracy_recipient_token_suggest": 1.0, + "train_accuracy_recipient_token_suggest_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.5899209486166008, + "step": 199, + "train_accuracy_first_token_arguments": 0.8125, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_building": 1.0, + "train_accuracy_recipient_token_building_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.5928853754940712, + "grad_norm": 1.7471771012162318, + "learning_rate": 1.8801422726424735e-06, + "loss": 0.1951, + "step": 200 + }, + { + "epoch": 0.5928853754940712, + "step": 200, + "train_accuracy_first_token_arguments": 0.9230769230769231, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.9375, + "train_accuracy_recipient_token_air": 1.0, + "train_accuracy_recipient_token_air_total": 1, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.5928853754940712, + "step": 200, + "train_accuracy_first_token_arguments": 0.8888888888888888, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_average": 1.0, + "train_accuracy_recipient_token_average_total": 1, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.5, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.5928853754940712, + "step": 200, + "train_accuracy_first_token_arguments": 0.8823529411764706, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_attach": 1.0, + "train_accuracy_recipient_token_attach_total": 1, + "train_accuracy_recipient_token_clean": 1.0, + "train_accuracy_recipient_token_clean_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_remove": 1.0, + "train_accuracy_recipient_token_remove_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.5958498023715415, + "grad_norm": 2.1505922703967726, + "learning_rate": 1.8568317292053894e-06, + "loss": 0.2902, + "step": 201 + }, + { + "epoch": 0.5958498023715415, + "step": 201, + "train_accuracy_first_token_arguments": 0.8235294117647058, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.7272727272727273, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_generate": 1.0, + "train_accuracy_recipient_token_generate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5958498023715415, + "step": 201, + "train_accuracy_first_token_arguments": 0.5925925925925926, + "train_accuracy_first_token_arguments_total": 27, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_ext": 1.0, + "train_accuracy_recipient_token_ext_total": 1, + "train_accuracy_recipient_token_fire": 1.0, + "train_accuracy_recipient_token_fire_total": 4, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.5958498023715415, + "step": 201, + "train_accuracy_first_token_arguments": 0.9032258064516129, + "train_accuracy_first_token_arguments_total": 31, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_d": 1.0, + "train_accuracy_recipient_token_d_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5988142292490118, + "grad_norm": 1.8633933589741076, + "learning_rate": 1.8335809148442074e-06, + "loss": 0.2595, + "step": 202 + }, + { + "epoch": 0.5988142292490118, + "step": 202, + "train_accuracy_first_token_arguments": 0.8888888888888888, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 2, + "train_accuracy_recipient_token_plant": 1.0, + "train_accuracy_recipient_token_plant_total": 3, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.5988142292490118, + "step": 202, + "train_accuracy_first_token_arguments": 0.8214285714285714, + "train_accuracy_first_token_arguments_total": 28, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_convert": 1.0, + "train_accuracy_recipient_token_convert_total": 1, + "train_accuracy_recipient_token_create": 0.0, + "train_accuracy_recipient_token_create_total": 1, + "train_accuracy_recipient_token_generate": 1.0, + "train_accuracy_recipient_token_generate_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5988142292490118, + "step": 202, + "train_accuracy_first_token_arguments": 0.52, + "train_accuracy_first_token_arguments_total": 25, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_calculate": 0.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_measure": 1.0, + "train_accuracy_recipient_token_measure_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.6017786561264822, + "grad_norm": 2.32947277265756, + "learning_rate": 1.8103919887911525e-06, + "loss": 0.3225, + "step": 203 + }, + { + "epoch": 0.6017786561264822, + "step": 203, + "train_accuracy_first_token_arguments": 0.8888888888888888, + "train_accuracy_first_token_arguments_total": 9, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.6017786561264822, + "step": 203, + "train_accuracy_first_token_arguments": 0.9, + "train_accuracy_first_token_arguments_total": 10, + "train_accuracy_recipient_token": 0.8125, + "train_accuracy_recipient_token_all": 0.7, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_d": 1.0, + "train_accuracy_recipient_token_d_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.6017786561264822, + "step": 203, + "train_accuracy_first_token_arguments": 0.9411764705882353, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 0.9166666666666666, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_bridge": 1.0, + "train_accuracy_recipient_token_bridge_total": 3, + "train_accuracy_recipient_token_find": 0.75, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_invest": 1.0, + "train_accuracy_recipient_token_invest_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.6047430830039525, + "grad_norm": 1.8953594708836015, + "learning_rate": 1.7872671045310703e-06, + "loss": 0.33, + "step": 204 + }, + { + "epoch": 0.6047430830039525, + "step": 204, + "train_accuracy_first_token_arguments": 0.9, + "train_accuracy_first_token_arguments_total": 10, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 3, + "train_accuracy_recipient_token_track": 1.0, + "train_accuracy_recipient_token_track_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.6047430830039525, + "step": 204, + "train_accuracy_first_token_arguments": 0.8235294117647058, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.6047430830039525, + "step": 204, + "train_accuracy_first_token_arguments": 0.8, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_book": 0.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.607707509881423, + "grad_norm": 1.4151900097248626, + "learning_rate": 1.7642084096014405e-06, + "loss": 0.1614, + "step": 205 + }, + { + "epoch": 0.607707509881423, + "step": 205, + "train_accuracy_first_token_arguments": 0.9705882352941176, + "train_accuracy_first_token_arguments_total": 34, + "train_accuracy_recipient_token": 0.875, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_organ": 0.0, + "train_accuracy_recipient_token_organ_total": 1, + "train_accuracy_recipient_token_review": 0.5, + "train_accuracy_recipient_token_review_total": 2, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.607707509881423, + "step": 205, + "train_accuracy_first_token_arguments": 0.9166666666666666, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_parse": 0.5, + "train_accuracy_recipient_token_parse_total": 2, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.607707509881423, + "step": 205, + "train_accuracy_first_token_arguments": 0.9230769230769231, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_convert": 1.0, + "train_accuracy_recipient_token_convert_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.6106719367588933, + "grad_norm": 2.1839121635772756, + "learning_rate": 1.7412180453929412e-06, + "loss": 0.2944, + "step": 206 + }, + { + "epoch": 0.6106719367588933, + "step": 206, + "train_accuracy_first_token_arguments": 0.8387096774193549, + "train_accuracy_first_token_arguments_total": 31, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 0.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.6106719367588933, + "step": 206, + "train_accuracy_first_token_arguments": 0.9180327868852459, + "train_accuracy_first_token_arguments_total": 61, + "train_accuracy_recipient_token": 0.8571428571428571, + "train_accuracy_recipient_token_all": 0.9285714285714286, + "train_accuracy_recipient_token_all_total": 14, + "train_accuracy_recipient_token_calculate": 0.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_find": 0.6666666666666666, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 21 + }, + { + "epoch": 0.6106719367588933, + "step": 206, + "train_accuracy_first_token_arguments": 0.8695652173913043, + "train_accuracy_first_token_arguments_total": 23, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_average": 1.0, + "train_accuracy_recipient_token_average_total": 1, + "train_accuracy_recipient_token_cat": 0.0, + "train_accuracy_recipient_token_cat_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.5, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.6136363636363636, + "grad_norm": 1.7289424673499263, + "learning_rate": 1.718298146950585e-06, + "loss": 0.1488, + "step": 207 + }, + { + "epoch": 0.6136363636363636, + "step": 207, + "train_accuracy_first_token_arguments": 0.7272727272727273, + "train_accuracy_first_token_arguments_total": 11, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8571428571428571, + "train_accuracy_recipient_token_all_total": 14, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.6136363636363636, + "step": 207, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.8666666666666667, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 15 + }, + { + "epoch": 0.6136363636363636, + "step": 207, + "train_accuracy_first_token_arguments": 0.8181818181818182, + "train_accuracy_first_token_arguments_total": 11, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_design": 1.0, + "train_accuracy_recipient_token_design_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_monitor": 1.0, + "train_accuracy_recipient_token_monitor_total": 1, + "train_accuracy_recipient_token_order": 0.0, + "train_accuracy_recipient_token_order_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.616600790513834, + "grad_norm": 2.1828846473287093, + "learning_rate": 1.6954508427754435e-06, + "loss": 0.2959, + "step": 208 + }, + { + "epoch": 0.616600790513834, + "step": 208, + "train_accuracy_first_token_arguments": 0.7608695652173914, + "train_accuracy_first_token_arguments_total": 46, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 0.5, + "train_accuracy_recipient_token_calculate_total": 2, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.616600790513834, + "step": 208, + "train_accuracy_first_token_arguments": 0.8888888888888888, + "train_accuracy_first_token_arguments_total": 9, + "train_accuracy_recipient_token": 0.75, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.616600790513834, + "step": 208, + "train_accuracy_first_token_arguments": 0.95, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 0.7894736842105263, + "train_accuracy_recipient_token_all": 0.7333333333333333, + "train_accuracy_recipient_token_all_total": 15, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.6195652173913043, + "grad_norm": 2.3743297354062536, + "learning_rate": 1.6726782546269793e-06, + "loss": 0.253, + "step": 209 + }, + { + "epoch": 0.6195652173913043, + "step": 209, + "train_accuracy_first_token_arguments": 0.8571428571428571, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_p": 1.0, + "train_accuracy_recipient_token_p_total": 3, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.6195652173913043, + "step": 209, + "train_accuracy_first_token_arguments": 0.8571428571428571, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.7142857142857143, + "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.6195652173913043, + "step": 209, + "train_accuracy_first_token_arguments": 0.8, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_locate": 1.0, + "train_accuracy_recipient_token_locate_total": 1, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.6225296442687747, + "grad_norm": 1.6892603326816198, + "learning_rate": 1.6499824973260086e-06, + "loss": 0.2491, + "step": 210 + }, + { + "epoch": 0.6225296442687747, + "step": 210, + "train_accuracy_first_token_arguments": 0.9230769230769231, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.6225296442687747, + "step": 210, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 26, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 3, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.6225296442687747, + "step": 210, + "train_accuracy_first_token_arguments": 0.75, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.875, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.6254940711462451, + "grad_norm": 2.3949353891643197, + "learning_rate": 1.6273656785582986e-06, + "loss": 0.3576, + "step": 211 + }, + { + "epoch": 0.6254940711462451, + "step": 211, + "train_accuracy_first_token_arguments": 0.6111111111111112, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.6254940711462451, + "step": 211, + "train_accuracy_first_token_arguments": 0.9166666666666666, + "train_accuracy_first_token_arguments_total": 24, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_check": 0.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_interpret": 1.0, + "train_accuracy_recipient_token_interpret_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.6254940711462451, + "step": 211, + "train_accuracy_first_token_arguments": 0.75, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 1, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.6284584980237155, + "grad_norm": 2.3847289223183123, + "learning_rate": 1.6048298986788345e-06, + "loss": 0.2644, + "step": 212 + }, + { + "epoch": 0.6284584980237155, + "eval_accuracy_first_token_arguments": 0.923923006416132, + "eval_accuracy_first_token_arguments_total": 1091, + "eval_accuracy_recipient_token": 0.9356984478935698, + "eval_accuracy_recipient_token_all": 0.9551656920077972, + "eval_accuracy_recipient_token_all_total": 513, + "eval_accuracy_recipient_token_calculate": 1.0, + "eval_accuracy_recipient_token_calculate_total": 13, + "eval_accuracy_recipient_token_find": 0.9473684210526315, + "eval_accuracy_recipient_token_find_total": 57, + "eval_accuracy_recipient_token_get": 0.9387755102040817, + "eval_accuracy_recipient_token_get_total": 196, + "eval_accuracy_recipient_token_identify": 0.6666666666666666, + "eval_accuracy_recipient_token_identify_total": 6, + "eval_loss": 0.22125016152858734, + "eval_perplexity": 1.034604144934006, + "eval_runtime": 113.1216, + "eval_samples_per_second": 0.893, + "eval_steps_per_second": 0.451, + "eval_total_number_recipient_token": 902, + "step": 212 + }, + { + "epoch": 0.6284584980237155, + "step": 212, + "train_accuracy_first_token_arguments": 0.9473684210526315, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.6284584980237155, + "step": 212, + "train_accuracy_first_token_arguments": 0.9411764705882353, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.8333333333333334, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.6284584980237155, + "step": 212, + "train_accuracy_first_token_arguments": 0.8333333333333334, + "train_accuracy_first_token_arguments_total": 42, + "train_accuracy_recipient_token": 0.7894736842105263, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 0.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 0.75, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_locate": 0.0, + "train_accuracy_recipient_token_locate_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.6314229249011858, + "grad_norm": 1.4624962550695582, + "learning_rate": 1.58237725051677e-06, + "loss": 0.1802, + "step": 213 + }, + { + "epoch": 0.6314229249011858, + "step": 213, + "train_accuracy_first_token_arguments": 0.8, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.5, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_sum": 1.0, + "train_accuracy_recipient_token_sum_total": 2, + "train_accuracy_recipient_token_translate": 0.5, + "train_accuracy_recipient_token_translate_total": 2, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.6314229249011858, + "step": 213, + "train_accuracy_first_token_arguments": 0.9230769230769231, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.875, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_register": 1.0, + "train_accuracy_recipient_token_register_total": 1, + "train_accuracy_recipient_token_verify": 1.0, + "train_accuracy_recipient_token_verify_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.6314229249011858, + "step": 213, + "train_accuracy_first_token_arguments": 0.8918918918918919, + "train_accuracy_first_token_arguments_total": 37, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_filter": 1.0, + "train_accuracy_recipient_token_filter_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_identify": 0.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.6343873517786561, + "grad_norm": 1.6957397755168004, + "learning_rate": 1.5600098191810682e-06, + "loss": 0.275, + "step": 214 + }, + { + "epoch": 0.6343873517786561, + "step": 214, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 28, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.6343873517786561, + "step": 214, + "train_accuracy_first_token_arguments": 0.8873239436619719, + "train_accuracy_first_token_arguments_total": 71, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_boat": 1.0, + "train_accuracy_recipient_token_boat_total": 3, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_suggest": 1.0, + "train_accuracy_recipient_token_suggest_total": 2, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.6343873517786561, + "step": 214, + "train_accuracy_first_token_arguments": 0.8214285714285714, + "train_accuracy_first_token_arguments_total": 28, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_hare": 0.6666666666666666, + "train_accuracy_recipient_token_hare_total": 3, + "train_accuracy_recipient_token_invest": 1.0, + "train_accuracy_recipient_token_invest_total": 1, + "train_accuracy_recipient_token_plane": 1.0, + "train_accuracy_recipient_token_plane_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.6373517786561265, + "grad_norm": 1.64437830735148, + "learning_rate": 1.5377296818668638e-06, + "loss": 0.1791, + "step": 215 + }, + { + "epoch": 0.6373517786561265, + "step": 215, + "train_accuracy_first_token_arguments": 0.7407407407407407, + "train_accuracy_first_token_arguments_total": 27, + "train_accuracy_recipient_token": 0.8125, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_update": 0.0, + "train_accuracy_recipient_token_update_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.6373517786561265, + "step": 215, + "train_accuracy_first_token_arguments": 0.6764705882352942, + "train_accuracy_first_token_arguments_total": 34, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_evaluate": 0.0, + "train_accuracy_recipient_token_evaluate_total": 1, + "train_accuracy_recipient_token_find": 0.3333333333333333, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_pack": 1.0, + "train_accuracy_recipient_token_pack_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.6373517786561265, + "step": 215, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_ind": 1.0, + "train_accuracy_recipient_token_ind_total": 4, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.6403162055335968, + "grad_norm": 2.485352785980747, + "learning_rate": 1.5155389076625663e-06, + "loss": 0.293, + "step": 216 + }, + { + "epoch": 0.6403162055335968, + "step": 216, + "train_accuracy_first_token_arguments": 0.9473684210526315, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_architecture": 1.0, + "train_accuracy_recipient_token_architecture_total": 1, + "train_accuracy_recipient_token_compare": 0.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_get": 0.7142857142857143, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.6403162055335968, + "step": 216, + "train_accuracy_first_token_arguments": 0.967741935483871, + "train_accuracy_first_token_arguments_total": 31, + "train_accuracy_recipient_token": 0.85, + "train_accuracy_recipient_token_all": 0.8461538461538461, + "train_accuracy_recipient_token_all_total": 13, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.8333333333333334, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.6403162055335968, + "step": 216, + "train_accuracy_first_token_arguments": 0.8421052631578947, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.6432806324110671, + "grad_norm": 1.78457120796898, + "learning_rate": 1.4934395573577016e-06, + "loss": 0.3178, + "step": 217 + }, + { + "epoch": 0.6432806324110671, + "step": 217, + "train_accuracy_first_token_arguments": 0.9565217391304348, + "train_accuracy_first_token_arguments_total": 23, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 9, + "train_accuracy_recipient_token_compare": 0.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.6432806324110671, + "step": 217, + "train_accuracy_first_token_arguments": 0.7692307692307693, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.6432806324110671, + "step": 217, + "train_accuracy_first_token_arguments": 0.9333333333333333, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_inspect": 1.0, + "train_accuracy_recipient_token_inspect_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.6462450592885376, + "grad_norm": 2.027667840038845, + "learning_rate": 1.4714336832515386e-06, + "loss": 0.167, + "step": 218 + }, + { + "epoch": 0.6462450592885376, + "step": 218, + "train_accuracy_first_token_arguments": 0.9285714285714286, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_language": 1.0, + "train_accuracy_recipient_token_language_total": 1, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 4, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.6462450592885376, + "step": 218, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 0.6666666666666666, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.6462450592885376, + "step": 218, + "train_accuracy_first_token_arguments": 0.85, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_art": 1.0, + "train_accuracy_recipient_token_art_total": 4, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.6492094861660079, + "grad_norm": 1.4967382429323417, + "learning_rate": 1.449523328962496e-06, + "loss": 0.2031, + "step": 219 + }, + { + "epoch": 0.6492094861660079, + "step": 219, + "train_accuracy_first_token_arguments": 0.7692307692307693, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.3333333333333333, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.6492094861660079, + "step": 219, + "train_accuracy_first_token_arguments": 0.8461538461538461, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_explain": 0.8, + "train_accuracy_recipient_token_explain_total": 5, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.6492094861660079, + "step": 219, + "train_accuracy_first_token_arguments": 0.8125, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_b": 1.0, + "train_accuracy_recipient_token_b_total": 1, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.6521739130434783, + "grad_norm": 1.8670991248523237, + "learning_rate": 1.4277105292383594e-06, + "loss": 0.1995, + "step": 220 + }, + { + "epoch": 0.6521739130434783, + "step": 220, + "train_accuracy_first_token_arguments": 0.9090909090909091, + "train_accuracy_first_token_arguments_total": 11, + "train_accuracy_recipient_token": 0.9375, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_inspect": 1.0, + "train_accuracy_recipient_token_inspect_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.6521739130434783, + "step": 220, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 10, + "train_accuracy_recipient_token": 0.8125, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.6521739130434783, + "step": 220, + "train_accuracy_first_token_arguments": 0.8333333333333334, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 0.8333333333333334, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.6551383399209486, + "grad_norm": 2.550246191316389, + "learning_rate": 1.4059973097673187e-06, + "loss": 0.2752, + "step": 221 + }, + { + "epoch": 0.6551383399209486, + "step": 221, + "train_accuracy_first_token_arguments": 0.9333333333333333, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_invest": 1.0, + "train_accuracy_recipient_token_invest_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.6551383399209486, + "step": 221, + "train_accuracy_first_token_arguments": 0.625, + "train_accuracy_first_token_arguments_total": 24, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_load": 1.0, + "train_accuracy_recipient_token_load_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.6551383399209486, + "step": 221, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 9, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.658102766798419, + "grad_norm": 1.8305223962446249, + "learning_rate": 1.3843856869898486e-06, + "loss": 0.2692, + "step": 222 + }, + { + "epoch": 0.658102766798419, + "step": 222, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_device": 1.0, + "train_accuracy_recipient_token_device_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.658102766798419, + "step": 222, + "train_accuracy_first_token_arguments": 0.8823529411764706, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.9, + "train_accuracy_recipient_token_get_total": 10, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.658102766798419, + "step": 222, + "train_accuracy_first_token_arguments": 0.88, + "train_accuracy_first_token_arguments_total": 50, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_park": 1.0, + "train_accuracy_recipient_token_park_total": 2, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.6610671936758893, + "grad_norm": 1.7979076335597703, + "learning_rate": 1.3628776679114516e-06, + "loss": 0.1334, + "step": 223 + }, + { + "epoch": 0.6610671936758893, + "step": 223, + "train_accuracy_first_token_arguments": 0.7894736842105263, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.8666666666666667, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 0.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 15 + }, + { + "epoch": 0.6610671936758893, + "step": 223, + "train_accuracy_first_token_arguments": 0.5, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_explain": 1.0, + "train_accuracy_recipient_token_explain_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_translate": 0.0, + "train_accuracy_recipient_token_translate_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.6610671936758893, + "step": 223, + "train_accuracy_first_token_arguments": 0.9210526315789473, + "train_accuracy_first_token_arguments_total": 38, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_local": 1.0, + "train_accuracy_recipient_token_local_total": 1, + "train_accuracy_recipient_token_plan": 1.0, + "train_accuracy_recipient_token_plan_total": 1, + "train_accuracy_recipient_token_restaurant": 1.0, + "train_accuracy_recipient_token_restaurant_total": 2, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.6640316205533597, + "grad_norm": 2.7045237044612875, + "learning_rate": 1.3414752499162676e-06, + "loss": 0.3284, + "step": 224 + }, + { + "epoch": 0.6640316205533597, + "step": 224, + "train_accuracy_first_token_arguments": 0.9310344827586207, + "train_accuracy_first_token_arguments_total": 29, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 2, + "train_accuracy_recipient_token_display": 1.0, + "train_accuracy_recipient_token_display_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_suggest": 0.0, + "train_accuracy_recipient_token_suggest_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.6640316205533597, + "step": 224, + "train_accuracy_first_token_arguments": 0.8333333333333334, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_classify": 0.0, + "train_accuracy_recipient_token_classify_total": 1, + "train_accuracy_recipient_token_decode": 1.0, + "train_accuracy_recipient_token_decode_total": 1, + "train_accuracy_recipient_token_g": 1.0, + "train_accuracy_recipient_token_g_total": 2, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.6640316205533597, + "step": 224, + "train_accuracy_first_token_arguments": 0.7692307692307693, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.6669960474308301, + "grad_norm": 1.8337644487638454, + "learning_rate": 1.3201804205815872e-06, + "loss": 0.2451, + "step": 225 + }, + { + "epoch": 0.6669960474308301, + "step": 225, + "train_accuracy_first_token_arguments": 0.9523809523809523, + "train_accuracy_first_token_arguments_total": 21, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_art": 1.0, + "train_accuracy_recipient_token_art_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.6669960474308301, + "step": 225, + "train_accuracy_first_token_arguments": 0.9166666666666666, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_christ": 1.0, + "train_accuracy_recipient_token_christ_total": 1, + "train_accuracy_recipient_token_explain": 0.0, + "train_accuracy_recipient_token_explain_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.6669960474308301, + "step": 225, + "train_accuracy_first_token_arguments": 0.8333333333333334, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compose": 1.0, + "train_accuracy_recipient_token_compose_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.6699604743083004, + "grad_norm": 1.601543108317484, + "learning_rate": 1.2989951574932693e-06, + "loss": 0.2814, + "step": 226 + }, + { + "epoch": 0.6699604743083004, + "step": 226, + "train_accuracy_first_token_arguments": 0.967741935483871, + "train_accuracy_first_token_arguments_total": 31, + "train_accuracy_recipient_token": 0.8, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_configure": 0.0, + "train_accuracy_recipient_token_configure_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 15 + }, + { + "epoch": 0.6699604743083004, + "step": 226, + "train_accuracy_first_token_arguments": 0.8888888888888888, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 0.9047619047619048, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.6666666666666666, + "train_accuracy_recipient_token_find_total": 6, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 21 + }, + { + "epoch": 0.6699604743083004, + "step": 226, + "train_accuracy_first_token_arguments": 0.9333333333333333, + "train_accuracy_first_token_arguments_total": 30, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_calculate": 0.75, + "train_accuracy_recipient_token_calculate_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.6729249011857708, + "grad_norm": 1.9791094562714513, + "learning_rate": 1.277921428062091e-06, + "loss": 0.2229, + "step": 227 + }, + { + "epoch": 0.6729249011857708, + "step": 227, + "train_accuracy_first_token_arguments": 0.9629629629629629, + "train_accuracy_first_token_arguments_total": 54, + "train_accuracy_recipient_token": 0.7777777777777778, + "train_accuracy_recipient_token_all": 0.6666666666666666, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.6729249011857708, + "step": 227, + "train_accuracy_first_token_arguments": 0.7777777777777778, + "train_accuracy_first_token_arguments_total": 9, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_d": 1.0, + "train_accuracy_recipient_token_d_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 14 + }, + { + "epoch": 0.6729249011857708, + "step": 227, + "train_accuracy_first_token_arguments": 0.6862745098039216, + "train_accuracy_first_token_arguments_total": 51, + "train_accuracy_recipient_token": 0.9375, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_customize": 1.0, + "train_accuracy_recipient_token_customize_total": 1, + "train_accuracy_recipient_token_draw": 1.0, + "train_accuracy_recipient_token_draw_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.6758893280632411, + "grad_norm": 2.308708795701363, + "learning_rate": 1.2569611893410374e-06, + "loss": 0.3605, + "step": 228 + }, + { + "epoch": 0.6758893280632411, + "step": 228, + "train_accuracy_first_token_arguments": 0.6666666666666666, + "train_accuracy_first_token_arguments_total": 21, + "train_accuracy_recipient_token": 0.7894736842105263, + "train_accuracy_recipient_token_all": 0.7142857142857143, + "train_accuracy_recipient_token_all_total": 14, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_event": 1.0, + "train_accuracy_recipient_token_event_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.6758893280632411, + "step": 228, + "train_accuracy_first_token_arguments": 0.7619047619047619, + "train_accuracy_first_token_arguments_total": 21, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_generate": 1.0, + "train_accuracy_recipient_token_generate_total": 5, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.6758893280632411, + "step": 228, + "train_accuracy_first_token_arguments": 0.7, + "train_accuracy_first_token_arguments_total": 10, + "train_accuracy_recipient_token": 0.8571428571428571, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_customize": 1.0, + "train_accuracy_recipient_token_customize_total": 1, + "train_accuracy_recipient_token_remove": 1.0, + "train_accuracy_recipient_token_remove_total": 1, + "train_accuracy_recipient_token_translate": 0.0, + "train_accuracy_recipient_token_translate_total": 1, + "train_total_number_recipient_token": 14 + }, + { + "epoch": 0.6788537549407114, + "grad_norm": 2.4044180995153486, + "learning_rate": 1.2361163878435594e-06, + "loss": 0.4432, + "step": 229 + }, + { + "epoch": 0.6788537549407114, + "step": 229, + "train_accuracy_first_token_arguments": 0.9615384615384616, + "train_accuracy_first_token_arguments_total": 26, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 1, + "train_total_number_recipient_token": 15 + }, + { + "epoch": 0.6788537549407114, + "step": 229, + "train_accuracy_first_token_arguments": 0.9230769230769231, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 10, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.6788537549407114, + "step": 229, + "train_accuracy_first_token_arguments": 0.7666666666666667, + "train_accuracy_first_token_arguments_total": 30, + "train_accuracy_recipient_token": 0.85, + "train_accuracy_recipient_token_all": 0.75, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.6818181818181818, + "grad_norm": 2.0846069978627417, + "learning_rate": 1.2153889593628032e-06, + "loss": 0.2884, + "step": 230 + }, + { + "epoch": 0.6818181818181818, + "step": 230, + "train_accuracy_first_token_arguments": 0.5185185185185185, + "train_accuracy_first_token_arguments_total": 27, + "train_accuracy_recipient_token": 0.8571428571428571, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 14 + }, + { + "epoch": 0.6818181818181818, + "step": 230, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.6818181818181818, + "step": 230, + "train_accuracy_first_token_arguments": 0.6060606060606061, + "train_accuracy_first_token_arguments_total": 33, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_wine": 1.0, + "train_accuracy_recipient_token_wine_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.6847826086956522, + "grad_norm": 1.903972306272471, + "learning_rate": 1.1947808287918406e-06, + "loss": 0.233, + "step": 231 + }, + { + "epoch": 0.6847826086956522, + "step": 231, + "train_accuracy_first_token_arguments": 0.75, + "train_accuracy_first_token_arguments_total": 24, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_decode": 1.0, + "train_accuracy_recipient_token_decode_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.6847826086956522, + "step": 231, + "train_accuracy_first_token_arguments": 0.9411764705882353, + "train_accuracy_first_token_arguments_total": 51, + "train_accuracy_recipient_token": 0.7647058823529411, + "train_accuracy_recipient_token_all": 0.7333333333333333, + "train_accuracy_recipient_token_all_total": 15, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.6847826086956522, + "step": 231, + "train_accuracy_first_token_arguments": 0.76, + "train_accuracy_first_token_arguments_total": 25, + "train_accuracy_recipient_token": 0.875, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.6877470355731226, + "grad_norm": 3.071022026928369, + "learning_rate": 1.1742939099449126e-06, + "loss": 0.3897, + "step": 232 + }, + { + "epoch": 0.6877470355731226, + "step": 232, + "train_accuracy_first_token_arguments": 0.6440677966101694, + "train_accuracy_first_token_arguments_total": 59, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_draw": 1.0, + "train_accuracy_recipient_token_draw_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_inspect": 1.0, + "train_accuracy_recipient_token_inspect_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.6877470355731226, + "step": 232, + "train_accuracy_first_token_arguments": 0.5454545454545454, + "train_accuracy_first_token_arguments_total": 33, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_dec": 1.0, + "train_accuracy_recipient_token_dec_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_verify": 1.0, + "train_accuracy_recipient_token_verify_total": 1, + "train_total_number_recipient_token": 15 + }, + { + "epoch": 0.6877470355731226, + "step": 232, + "train_accuracy_first_token_arguments": 0.8571428571428571, + "train_accuracy_first_token_arguments_total": 21, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_convert": 1.0, + "train_accuracy_recipient_token_convert_total": 4, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.6907114624505929, + "grad_norm": 2.858676916448621, + "learning_rate": 1.153930105379695e-06, + "loss": 0.2713, + "step": 233 + }, + { + "epoch": 0.6907114624505929, + "step": 233, + "train_accuracy_first_token_arguments": 0.9411764705882353, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_create": 1.0, + "train_accuracy_recipient_token_create_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.6907114624505929, + "step": 233, + "train_accuracy_first_token_arguments": 0.9459459459459459, + "train_accuracy_first_token_arguments_total": 37, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.8571428571428571, + "train_accuracy_recipient_token_find_total": 7, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.6907114624505929, + "step": 233, + "train_accuracy_first_token_arguments": 0.8260869565217391, + "train_accuracy_first_token_arguments_total": 23, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_enh": 1.0, + "train_accuracy_recipient_token_enh_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_optimize": 1.0, + "train_accuracy_recipient_token_optimize_total": 2, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.6936758893280632, + "grad_norm": 1.4873714467096886, + "learning_rate": 1.1336913062206157e-06, + "loss": 0.1483, + "step": 234 + }, + { + "epoch": 0.6936758893280632, + "step": 234, + "train_accuracy_first_token_arguments": 0.9, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8571428571428571, + "train_accuracy_recipient_token_all_total": 14, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.6936758893280632, + "step": 234, + "train_accuracy_first_token_arguments": 0.9090909090909091, + "train_accuracy_first_token_arguments_total": 11, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.8333333333333334, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.6936758893280632, + "step": 234, + "train_accuracy_first_token_arguments": 0.8333333333333334, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_detect": 1.0, + "train_accuracy_recipient_token_detect_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_review": 1.0, + "train_accuracy_recipient_token_review_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.6966403162055336, + "grad_norm": 1.5877712208219548, + "learning_rate": 1.1135793919832336e-06, + "loss": 0.2483, + "step": 235 + }, + { + "epoch": 0.6966403162055336, + "step": 235, + "train_accuracy_first_token_arguments": 0.9411764705882353, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_join": 1.0, + "train_accuracy_recipient_token_join_total": 1, + "train_accuracy_recipient_token_vol": 1.0, + "train_accuracy_recipient_token_vol_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.6966403162055336, + "step": 235, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 11, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 9, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.6966403162055336, + "step": 235, + "train_accuracy_first_token_arguments": 0.9047619047619048, + "train_accuracy_first_token_arguments_total": 21, + "train_accuracy_recipient_token": 0.85, + "train_accuracy_recipient_token_all": 0.7692307692307693, + "train_accuracy_recipient_token_all_total": 13, + "train_accuracy_recipient_token_detect": 1.0, + "train_accuracy_recipient_token_detect_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.6996047430830039, + "grad_norm": 1.392788536484163, + "learning_rate": 1.09359623039969e-06, + "loss": 0.1911, + "step": 236 + }, + { + "epoch": 0.6996047430830039, + "step": 236, + "train_accuracy_first_token_arguments": 0.9090909090909091, + "train_accuracy_first_token_arguments_total": 22, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_optimize": 0.0, + "train_accuracy_recipient_token_optimize_total": 1, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.6996047430830039, + "step": 236, + "train_accuracy_first_token_arguments": 0.8108108108108109, + "train_accuracy_first_token_arguments_total": 37, + "train_accuracy_recipient_token": 0.7857142857142857, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_design": 1.0, + "train_accuracy_recipient_token_design_total": 1, + "train_accuracy_recipient_token_im": 1.0, + "train_accuracy_recipient_token_im_total": 1, + "train_accuracy_recipient_token_install": 0.0, + "train_accuracy_recipient_token_install_total": 1, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 14 + }, + { + "epoch": 0.6996047430830039, + "step": 236, + "train_accuracy_first_token_arguments": 0.8695652173913043, + "train_accuracy_first_token_arguments_total": 23, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.7025691699604744, + "grad_norm": 1.9185387350295975, + "learning_rate": 1.0737436772452602e-06, + "loss": 0.2789, + "step": 237 + }, + { + "epoch": 0.7025691699604744, + "step": 237, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.7142857142857143, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.7025691699604744, + "step": 237, + "train_accuracy_first_token_arguments": 0.9285714285714286, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.7025691699604744, + "step": 237, + "train_accuracy_first_token_arguments": 0.7435897435897436, + "train_accuracy_first_token_arguments_total": 39, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_generate": 1.0, + "train_accuracy_recipient_token_generate_total": 5, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.7055335968379447, + "grad_norm": 1.6960089025056184, + "learning_rate": 1.054023576166014e-06, + "loss": 0.2385, + "step": 238 + }, + { + "epoch": 0.7055335968379447, + "step": 238, + "train_accuracy_first_token_arguments": 0.9, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.7055335968379447, + "step": 238, + "train_accuracy_first_token_arguments": 0.8947368421052632, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.9333333333333333, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_count": 1.0, + "train_accuracy_recipient_token_count_total": 1, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_track": 1.0, + "train_accuracy_recipient_token_track_total": 1, + "train_total_number_recipient_token": 15 + }, + { + "epoch": 0.7055335968379447, + "step": 238, + "train_accuracy_first_token_arguments": 0.9090909090909091, + "train_accuracy_first_token_arguments_total": 44, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.708498023715415, + "grad_norm": 2.366641831793585, + "learning_rate": 1.0344377585076e-06, + "loss": 0.4772, + "step": 239 + }, + { + "epoch": 0.708498023715415, + "step": 239, + "train_accuracy_first_token_arguments": 0.9444444444444444, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.708498023715415, + "step": 239, + "train_accuracy_first_token_arguments": 0.75, + "train_accuracy_first_token_arguments_total": 28, + "train_accuracy_recipient_token": 0.9333333333333333, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_merge": 1.0, + "train_accuracy_recipient_token_merge_total": 1, + "train_total_number_recipient_token": 15 + }, + { + "epoch": 0.708498023715415, + "step": 239, + "train_accuracy_first_token_arguments": 0.9411764705882353, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 4, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.7114624505928854, + "grad_norm": 2.0428633070085302, + "learning_rate": 1.0149880431451736e-06, + "loss": 0.2845, + "step": 240 + }, + { + "epoch": 0.7114624505928854, + "step": 240, + "train_accuracy_first_token_arguments": 0.9473684210526315, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_book": 0.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_order": 1.0, + "train_accuracy_recipient_token_order_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.7114624505928854, + "step": 240, + "train_accuracy_first_token_arguments": 0.8421052631578947, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.6666666666666666, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_inspect": 1.0, + "train_accuracy_recipient_token_inspect_total": 1, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.7114624505928854, + "step": 240, + "train_accuracy_first_token_arguments": 0.9, + "train_accuracy_first_token_arguments_total": 30, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_count": 1.0, + "train_accuracy_recipient_token_count_total": 1, + "train_accuracy_recipient_token_d": 1.0, + "train_accuracy_recipient_token_d_total": 1, + "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_vehicle": 0.6666666666666666, + "train_accuracy_recipient_token_vehicle_total": 3, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.7144268774703557, + "grad_norm": 1.870434102867515, + "learning_rate": 9.956762363144892e-07, + "loss": 0.1578, + "step": 241 + }, + { + "epoch": 0.7144268774703557, + "step": 241, + "train_accuracy_first_token_arguments": 0.7222222222222222, + "train_accuracy_first_token_arguments_total": 36, + "train_accuracy_recipient_token": 0.7894736842105263, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.7144268774703557, + "step": 241, + "train_accuracy_first_token_arguments": 0.8235294117647058, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_recommend": 0.0, + "train_accuracy_recipient_token_recommend_total": 1, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.7144268774703557, + "step": 241, + "train_accuracy_first_token_arguments": 0.7272727272727273, + "train_accuracy_first_token_arguments_total": 11, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 0.5, + "train_accuracy_recipient_token_analyze_total": 2, + "train_accuracy_recipient_token_building": 1.0, + "train_accuracy_recipient_token_building_total": 3, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_object": 1.0, + "train_accuracy_recipient_token_object_total": 3, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.717391304347826, + "grad_norm": 2.192061804607294, + "learning_rate": 9.765041314441529e-07, + "loss": 0.3503, + "step": 242 + }, + { + "epoch": 0.717391304347826, + "step": 242, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.8888888888888888, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.717391304347826, + "step": 242, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 11, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_join": 1.0, + "train_accuracy_recipient_token_join_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.717391304347826, + "step": 242, + "train_accuracy_first_token_arguments": 0.9090909090909091, + "train_accuracy_first_token_arguments_total": 11, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compile": 1.0, + "train_accuracy_recipient_token_compile_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.7203557312252964, + "grad_norm": 2.540677501201676, + "learning_rate": 9.574735089890765e-07, + "loss": 0.3029, + "step": 243 + }, + { + "epoch": 0.7203557312252964, + "step": 243, + "train_accuracy_first_token_arguments": 0.8, + "train_accuracy_first_token_arguments_total": 10, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.7203557312252964, + "step": 243, + "train_accuracy_first_token_arguments": 0.8235294117647058, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_order": 1.0, + "train_accuracy_recipient_token_order_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.7203557312252964, + "step": 243, + "train_accuracy_first_token_arguments": 0.6875, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_exp": 1.0, + "train_accuracy_recipient_token_exp_total": 4, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.7233201581027668, + "grad_norm": 1.681752257297182, + "learning_rate": 9.385861362651322e-07, + "loss": 0.2678, + "step": 244 + }, + { + "epoch": 0.7233201581027668, + "step": 244, + "train_accuracy_first_token_arguments": 0.717391304347826, + "train_accuracy_first_token_arguments_total": 46, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_create": 1.0, + "train_accuracy_recipient_token_create_total": 1, + "train_accuracy_recipient_token_design": 1.0, + "train_accuracy_recipient_token_design_total": 2, + "train_accuracy_recipient_token_draw": 1.0, + "train_accuracy_recipient_token_draw_total": 1, + "train_accuracy_recipient_token_evaluate": 0.0, + "train_accuracy_recipient_token_evaluate_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.7233201581027668, + "step": 244, + "train_accuracy_first_token_arguments": 0.9130434782608695, + "train_accuracy_first_token_arguments_total": 23, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 3, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_list": 1.0, + "train_accuracy_recipient_token_list_total": 3, + "train_accuracy_recipient_token_use": 1.0, + "train_accuracy_recipient_token_use_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.7233201581027668, + "step": 244, + "train_accuracy_first_token_arguments": 0.9423076923076923, + "train_accuracy_first_token_arguments_total": 52, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_design": 1.0, + "train_accuracy_recipient_token_design_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.7262845849802372, + "grad_norm": 1.6081692582595575, + "learning_rate": 9.198437672850249e-07, + "loss": 0.2291, + "step": 245 + }, + { + "epoch": 0.7262845849802372, + "step": 245, + "train_accuracy_first_token_arguments": 0.8, + "train_accuracy_first_token_arguments_total": 10, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_create": 1.0, + "train_accuracy_recipient_token_create_total": 2, + "train_accuracy_recipient_token_generate": 1.0, + "train_accuracy_recipient_token_generate_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.7262845849802372, + "step": 245, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.7262845849802372, + "step": 245, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9166666666666666, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.7292490118577075, + "grad_norm": 1.786148716474924, + "learning_rate": 9.012481425954053e-07, + "loss": 0.2712, + "step": 246 + }, + { + "epoch": 0.7292490118577075, + "step": 246, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 9, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.7292490118577075, + "step": 246, + "train_accuracy_first_token_arguments": 0.9333333333333333, + "train_accuracy_first_token_arguments_total": 30, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 2, + "train_accuracy_recipient_token_inventory": 1.0, + "train_accuracy_recipient_token_inventory_total": 1, + "train_accuracy_recipient_token_sales": 1.0, + "train_accuracy_recipient_token_sales_total": 1, + "train_accuracy_recipient_token_shelf": 1.0, + "train_accuracy_recipient_token_shelf_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.7292490118577075, + "step": 246, + "train_accuracy_first_token_arguments": 0.875, + "train_accuracy_first_token_arguments_total": 24, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_buy": 0.0, + "train_accuracy_recipient_token_buy_total": 1, + "train_accuracy_recipient_token_find": 0.8571428571428571, + "train_accuracy_recipient_token_find_total": 7, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.7322134387351779, + "grad_norm": 1.7149609341130112, + "learning_rate": 8.828009891152301e-07, + "loss": 0.2038, + "step": 247 + }, + { + "epoch": 0.7322134387351779, + "step": 247, + "train_accuracy_first_token_arguments": 0.7857142857142857, + "train_accuracy_first_token_arguments_total": 28, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_locate": 1.0, + "train_accuracy_recipient_token_locate_total": 1, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.7322134387351779, + "step": 247, + "train_accuracy_first_token_arguments": 0.8571428571428571, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_m": 1.0, + "train_accuracy_recipient_token_m_total": 4, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.7322134387351779, + "step": 247, + "train_accuracy_first_token_arguments": 0.9761904761904762, + "train_accuracy_first_token_arguments_total": 42, + "train_accuracy_recipient_token": 0.7777777777777778, + "train_accuracy_recipient_token_all": 0.7857142857142857, + "train_accuracy_recipient_token_all_total": 14, + "train_accuracy_recipient_token_calculate": 0.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.7351778656126482, + "grad_norm": 1.6828278608062666, + "learning_rate": 8.64504019975386e-07, + "loss": 0.218, + "step": 248 + }, + { + "epoch": 0.7351778656126482, + "step": 248, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.7351778656126482, + "step": 248, + "train_accuracy_first_token_arguments": 0.9166666666666666, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 0.6, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.7351778656126482, + "step": 248, + "train_accuracy_first_token_arguments": 0.9523809523809523, + "train_accuracy_first_token_arguments_total": 21, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9285714285714286, + "train_accuracy_recipient_token_all_total": 14, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_translate": 0.0, + "train_accuracy_recipient_token_translate_total": 1, + "train_accuracy_recipient_token_vehicle": 1.0, + "train_accuracy_recipient_token_vehicle_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.7381422924901185, + "grad_norm": 2.180566937973925, + "learning_rate": 8.463589343595976e-07, + "loss": 0.3458, + "step": 249 + }, + { + "epoch": 0.7381422924901185, + "step": 249, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 27, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9166666666666666, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_construction": 1.0, + "train_accuracy_recipient_token_construction_total": 2, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.7381422924901185, + "step": 249, + "train_accuracy_first_token_arguments": 0.9375, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.7142857142857143, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.7381422924901185, + "step": 249, + "train_accuracy_first_token_arguments": 0.9285714285714286, + "train_accuracy_first_token_arguments_total": 28, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.741106719367589, + "grad_norm": 1.7757645324819282, + "learning_rate": 8.28367417346633e-07, + "loss": 0.2223, + "step": 250 + }, + { + "epoch": 0.741106719367589, + "step": 250, + "train_accuracy_first_token_arguments": 0.75, + "train_accuracy_first_token_arguments_total": 8, + "train_accuracy_recipient_token": 0.8, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 5, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 10 + }, + { + "epoch": 0.741106719367589, + "step": 250, + "train_accuracy_first_token_arguments": 0.8571428571428571, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_animal": 1.0, + "train_accuracy_recipient_token_animal_total": 3, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_locate": 0.0, + "train_accuracy_recipient_token_locate_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.741106719367589, + "step": 250, + "train_accuracy_first_token_arguments": 0.95, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 0.7857142857142857, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.5, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 14 + }, + { + "epoch": 0.7440711462450593, + "grad_norm": 11.280998921270033, + "learning_rate": 8.105311397538085e-07, + "loss": 0.5545, + "step": 251 + }, + { + "epoch": 0.7440711462450593, + "step": 251, + "train_accuracy_first_token_arguments": 0.7142857142857143, + "train_accuracy_first_token_arguments_total": 28, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_compose": 1.0, + "train_accuracy_recipient_token_compose_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.7440711462450593, + "step": 251, + "train_accuracy_first_token_arguments": 0.75, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.7440711462450593, + "step": 251, + "train_accuracy_first_token_arguments": 0.9090909090909091, + "train_accuracy_first_token_arguments_total": 11, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.7470355731225297, + "grad_norm": 2.177175537608394, + "learning_rate": 7.928517579818312e-07, + "loss": 0.2545, + "step": 252 + }, + { + "epoch": 0.7470355731225297, + "step": 252, + "train_accuracy_first_token_arguments": 0.8, + "train_accuracy_first_token_arguments_total": 30, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.75, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.7470355731225297, + "step": 252, + "train_accuracy_first_token_arguments": 0.6666666666666666, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_evaluate": 0.6666666666666666, + "train_accuracy_recipient_token_evaluate_total": 3, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.7470355731225297, + "step": 252, + "train_accuracy_first_token_arguments": 0.9473684210526315, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_te": 1.0, + "train_accuracy_recipient_token_te_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.75, + "grad_norm": 1.527663654975652, + "learning_rate": 7.753309138609705e-07, + "loss": 0.1884, + "step": 253 + }, + { + "epoch": 0.75, + "step": 253, + "train_accuracy_first_token_arguments": 0.7884615384615384, + "train_accuracy_first_token_arguments_total": 52, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_evaluate": 0.0, + "train_accuracy_recipient_token_evaluate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_suggest": 1.0, + "train_accuracy_recipient_token_suggest_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.75, + "step": 253, + "train_accuracy_first_token_arguments": 0.6764705882352942, + "train_accuracy_first_token_arguments_total": 34, + "train_accuracy_recipient_token": 0.7777777777777778, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 0.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_review": 1.0, + "train_accuracy_recipient_token_review_total": 1, + "train_accuracy_recipient_token_stamp": 0.75, + "train_accuracy_recipient_token_stamp_total": 4, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.75, + "step": 253, + "train_accuracy_first_token_arguments": 0.8125, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_convert": 1.0, + "train_accuracy_recipient_token_convert_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_sum": 1.0, + "train_accuracy_recipient_token_sum_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.7529644268774703, + "grad_norm": 1.9700572495421322, + "learning_rate": 7.579702344985899e-07, + "loss": 0.3442, + "step": 254 + }, + { + "epoch": 0.7529644268774703, + "step": 254, + "train_accuracy_first_token_arguments": 0.84, + "train_accuracy_first_token_arguments_total": 25, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 2, + "train_accuracy_recipient_token_decode": 1.0, + "train_accuracy_recipient_token_decode_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.7529644268774703, + "step": 254, + "train_accuracy_first_token_arguments": 0.8125, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_adjust": 1.0, + "train_accuracy_recipient_token_adjust_total": 1, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_discover": 1.0, + "train_accuracy_recipient_token_discover_total": 1, + "train_accuracy_recipient_token_enh": 1.0, + "train_accuracy_recipient_token_enh_total": 1, + "train_accuracy_recipient_token_event": 1.0, + "train_accuracy_recipient_token_event_total": 4, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.7529644268774703, + "step": 254, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_compress": 1.0, + "train_accuracy_recipient_token_compress_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.7559288537549407, + "grad_norm": 2.171316152952409, + "learning_rate": 7.407713321280377e-07, + "loss": 0.1791, + "step": 255 + }, + { + "epoch": 0.7559288537549407, + "step": 255, + "train_accuracy_first_token_arguments": 0.8947368421052632, + "train_accuracy_first_token_arguments_total": 38, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_customize": 1.0, + "train_accuracy_recipient_token_customize_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_style": 1.0, + "train_accuracy_recipient_token_style_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.7559288537549407, + "step": 255, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 26, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.7559288537549407, + "step": 255, + "train_accuracy_first_token_arguments": 0.9285714285714286, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_find": 0.6666666666666666, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_rec": 1.0, + "train_accuracy_recipient_token_rec_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.758893280632411, + "grad_norm": 1.7774995166747949, + "learning_rate": 7.237358039589271e-07, + "loss": 0.2064, + "step": 256 + }, + { + "epoch": 0.758893280632411, + "step": 256, + "train_accuracy_first_token_arguments": 0.8529411764705882, + "train_accuracy_first_token_arguments_total": 34, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_order": 1.0, + "train_accuracy_recipient_token_order_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.758893280632411, + "step": 256, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.875, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_clean": 1.0, + "train_accuracy_recipient_token_clean_total": 1, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_rent": 1.0, + "train_accuracy_recipient_token_rent_total": 2, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.758893280632411, + "step": 256, + "train_accuracy_first_token_arguments": 0.9090909090909091, + "train_accuracy_first_token_arguments_total": 11, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.7618577075098815, + "grad_norm": 1.7945920171198126, + "learning_rate": 7.068652320288081e-07, + "loss": 0.1652, + "step": 257 + }, + { + "epoch": 0.7618577075098815, + "step": 257, + "train_accuracy_first_token_arguments": 0.9166666666666666, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.8888888888888888, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.7618577075098815, + "step": 257, + "train_accuracy_first_token_arguments": 0.8105263157894737, + "train_accuracy_first_token_arguments_total": 95, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.7618577075098815, + "step": 257, + "train_accuracy_first_token_arguments": 0.8666666666666667, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_calculate": 0.5, + "train_accuracy_recipient_token_calculate_total": 2, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.7648221343873518, + "grad_norm": 2.033550650864743, + "learning_rate": 6.901611830562469e-07, + "loss": 0.2407, + "step": 258 + }, + { + "epoch": 0.7648221343873518, + "step": 258, + "train_accuracy_first_token_arguments": 0.9047619047619048, + "train_accuracy_first_token_arguments_total": 21, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_ap": 1.0, + "train_accuracy_recipient_token_ap_total": 1, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.7648221343873518, + "step": 258, + "train_accuracy_first_token_arguments": 0.72, + "train_accuracy_first_token_arguments_total": 25, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_create": 1.0, + "train_accuracy_recipient_token_create_total": 1, + "train_accuracy_recipient_token_discover": 1.0, + "train_accuracy_recipient_token_discover_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_interpret": 1.0, + "train_accuracy_recipient_token_interpret_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.7648221343873518, + "step": 258, + "train_accuracy_first_token_arguments": 0.9090909090909091, + "train_accuracy_first_token_arguments_total": 22, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_event": 0.6666666666666666, + "train_accuracy_recipient_token_event_total": 3, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_suggest": 1.0, + "train_accuracy_recipient_token_suggest_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.7677865612648221, + "grad_norm": 1.6980201197958136, + "learning_rate": 6.736252082953307e-07, + "loss": 0.2026, + "step": 259 + }, + { + "epoch": 0.7677865612648221, + "step": 259, + "train_accuracy_first_token_arguments": 0.9459459459459459, + "train_accuracy_first_token_arguments_total": 37, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.7677865612648221, + "step": 259, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 22, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_order": 1.0, + "train_accuracy_recipient_token_order_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.7677865612648221, + "step": 259, + "train_accuracy_first_token_arguments": 0.92, + "train_accuracy_first_token_arguments_total": 50, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.7707509881422925, + "grad_norm": 1.4870369650234745, + "learning_rate": 6.572588433916082e-07, + "loss": 0.1544, + "step": 260 + }, + { + "epoch": 0.7707509881422925, + "step": 260, + "train_accuracy_first_token_arguments": 0.8064516129032258, + "train_accuracy_first_token_arguments_total": 31, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 1, + "train_accuracy_recipient_token_restore": 0.0, + "train_accuracy_recipient_token_restore_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.7707509881422925, + "step": 260, + "train_accuracy_first_token_arguments": 0.7413793103448276, + "train_accuracy_first_token_arguments_total": 58, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_evaluate": 1.0, + "train_accuracy_recipient_token_evaluate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.7707509881422925, + "step": 260, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 29, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.7737154150197628, + "grad_norm": 1.6094384600411795, + "learning_rate": 6.410636082394772e-07, + "loss": 0.2592, + "step": 261 + }, + { + "epoch": 0.7737154150197628, + "step": 261, + "train_accuracy_first_token_arguments": 0.7692307692307693, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.8, + "train_accuracy_recipient_token_find_total": 5, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_locate": 1.0, + "train_accuracy_recipient_token_locate_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.7737154150197628, + "step": 261, + "train_accuracy_first_token_arguments": 0.7666666666666667, + "train_accuracy_first_token_arguments_total": 30, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_ass": 1.0, + "train_accuracy_recipient_token_ass_total": 1, + "train_accuracy_recipient_token_environment": 1.0, + "train_accuracy_recipient_token_environment_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_w": 1.0, + "train_accuracy_recipient_token_w_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.7737154150197628, + "step": 261, + "train_accuracy_first_token_arguments": 0.975, + "train_accuracy_first_token_arguments_total": 40, + "train_accuracy_recipient_token": 0.9523809523809523, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_f": 1.0, + "train_accuracy_recipient_token_f_total": 4, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_locate": 1.0, + "train_accuracy_recipient_token_locate_total": 1, + "train_accuracy_recipient_token_recommend": 0.5, + "train_accuracy_recipient_token_recommend_total": 2, + "train_total_number_recipient_token": 21 + }, + { + "epoch": 0.7766798418972332, + "grad_norm": 1.997171944251961, + "learning_rate": 6.250410068410367e-07, + "loss": 0.2894, + "step": 262 + }, + { + "epoch": 0.7766798418972332, + "step": 262, + "train_accuracy_first_token_arguments": 0.9333333333333333, + "train_accuracy_first_token_arguments_total": 30, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.7766798418972332, + "step": 262, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 11, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.7766798418972332, + "step": 262, + "train_accuracy_first_token_arguments": 0.6956521739130435, + "train_accuracy_first_token_arguments_total": 69, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 2, + "train_accuracy_recipient_token_enh": 1.0, + "train_accuracy_recipient_token_enh_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.7796442687747036, + "grad_norm": 1.5690225372388775, + "learning_rate": 6.091925271664156e-07, + "loss": 0.1593, + "step": 263 + }, + { + "epoch": 0.7796442687747036, + "step": 263, + "train_accuracy_first_token_arguments": 0.75, + "train_accuracy_first_token_arguments_total": 24, + "train_accuracy_recipient_token": 0.85, + "train_accuracy_recipient_token_all": 0.7692307692307693, + "train_accuracy_recipient_token_all_total": 13, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.7796442687747036, + "step": 263, + "train_accuracy_first_token_arguments": 0.8333333333333334, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_locate": 1.0, + "train_accuracy_recipient_token_locate_total": 2, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 2, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.7796442687747036, + "step": 263, + "train_accuracy_first_token_arguments": 0.9285714285714286, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 10, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.782608695652174, + "grad_norm": 1.769966354853594, + "learning_rate": 5.93519641015591e-07, + "loss": 0.2617, + "step": 264 + }, + { + "epoch": 0.782608695652174, + "step": 264, + "train_accuracy_first_token_arguments": 0.9, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_stamp": 1.0, + "train_accuracy_recipient_token_stamp_total": 4, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.782608695652174, + "step": 264, + "train_accuracy_first_token_arguments": 0.6875, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.782608695652174, + "step": 264, + "train_accuracy_first_token_arguments": 0.8529411764705882, + "train_accuracy_first_token_arguments_total": 34, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_city": 1.0, + "train_accuracy_recipient_token_city_total": 1, + "train_accuracy_recipient_token_generate": 1.0, + "train_accuracy_recipient_token_generate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_urban": 1.0, + "train_accuracy_recipient_token_urban_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.7855731225296443, + "grad_norm": 2.0500780698786483, + "learning_rate": 5.780238038817035e-07, + "loss": 0.4104, + "step": 265 + }, + { + "epoch": 0.7855731225296443, + "step": 265, + "train_accuracy_first_token_arguments": 0.9444444444444444, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_gift": 1.0, + "train_accuracy_recipient_token_gift_total": 1, + "train_accuracy_recipient_token_personal": 1.0, + "train_accuracy_recipient_token_personal_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.7855731225296443, + "step": 265, + "train_accuracy_first_token_arguments": 0.8333333333333334, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_accuracy_recipient_token_initialize": 0.0, + "train_accuracy_recipient_token_initialize_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.7855731225296443, + "step": 265, + "train_accuracy_first_token_arguments": 0.782608695652174, + "train_accuracy_first_token_arguments_total": 23, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 0.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_verify": 1.0, + "train_accuracy_recipient_token_verify_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.7885375494071146, + "grad_norm": 2.3088267661053083, + "learning_rate": 5.627064548158903e-07, + "loss": 0.215, + "step": 266 + }, + { + "epoch": 0.7885375494071146, + "step": 266, + "train_accuracy_first_token_arguments": 0.7692307692307693, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_building": 1.0, + "train_accuracy_recipient_token_building_total": 1, + "train_accuracy_recipient_token_flag": 1.0, + "train_accuracy_recipient_token_flag_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.7885375494071146, + "step": 266, + "train_accuracy_first_token_arguments": 0.8333333333333334, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_deb": 1.0, + "train_accuracy_recipient_token_deb_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.7885375494071146, + "step": 266, + "train_accuracy_first_token_arguments": 0.7692307692307693, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_building": 1.0, + "train_accuracy_recipient_token_building_total": 4, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.791501976284585, + "grad_norm": 2.3252209612398764, + "learning_rate": 5.47569016293649e-07, + "loss": 0.2737, + "step": 267 + }, + { + "epoch": 0.791501976284585, + "step": 267, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_configure": 1.0, + "train_accuracy_recipient_token_configure_total": 5, + "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.791501976284585, + "step": 267, + "train_accuracy_first_token_arguments": 0.6363636363636364, + "train_accuracy_first_token_arguments_total": 22, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 1, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.791501976284585, + "step": 267, + "train_accuracy_first_token_arguments": 0.8666666666666667, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 1, + "train_accuracy_recipient_token_rel": 1.0, + "train_accuracy_recipient_token_rel_total": 2, + "train_accuracy_recipient_token_y": 1.0, + "train_accuracy_recipient_token_y_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.7944664031620553, + "grad_norm": 1.6931063369189647, + "learning_rate": 5.326128940827313e-07, + "loss": 0.172, + "step": 268 + }, + { + "epoch": 0.7944664031620553, + "step": 268, + "train_accuracy_first_token_arguments": 0.9166666666666666, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.7944664031620553, + "step": 268, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.7944664031620553, + "step": 268, + "train_accuracy_first_token_arguments": 0.9375, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 0.9047619047619048, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_count": 1.0, + "train_accuracy_recipient_token_count_total": 1, + "train_accuracy_recipient_token_find": 0.75, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_locate": 1.0, + "train_accuracy_recipient_token_locate_total": 1, + "train_total_number_recipient_token": 21 + }, + { + "epoch": 0.7974308300395256, + "grad_norm": 1.608131055033889, + "learning_rate": 5.178394771125969e-07, + "loss": 0.1891, + "step": 269 + }, + { + "epoch": 0.7974308300395256, + "step": 269, + "train_accuracy_first_token_arguments": 0.7058823529411765, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.8333333333333334, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_short": 1.0, + "train_accuracy_recipient_token_short_total": 1, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.7974308300395256, + "step": 269, + "train_accuracy_first_token_arguments": 0.9047619047619048, + "train_accuracy_first_token_arguments_total": 21, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_recipe": 1.0, + "train_accuracy_recipient_token_recipe_total": 4, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.7974308300395256, + "step": 269, + "train_accuracy_first_token_arguments": 0.9047619047619048, + "train_accuracy_first_token_arguments_total": 21, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.8003952569169961, + "grad_norm": 1.9940648204984408, + "learning_rate": 5.032501373454266e-07, + "loss": 0.2175, + "step": 270 + }, + { + "epoch": 0.8003952569169961, + "step": 270, + "train_accuracy_first_token_arguments": 0.9354838709677419, + "train_accuracy_first_token_arguments_total": 31, + "train_accuracy_recipient_token": 0.8125, + "train_accuracy_recipient_token_all": 0.8333333333333334, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_crop": 1.0, + "train_accuracy_recipient_token_crop_total": 1, + "train_accuracy_recipient_token_f": 1.0, + "train_accuracy_recipient_token_f_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_har": 0.0, + "train_accuracy_recipient_token_har_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.8003952569169961, + "step": 270, + "train_accuracy_first_token_arguments": 0.9259259259259259, + "train_accuracy_first_token_arguments_total": 27, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.8888888888888888, + "train_accuracy_recipient_token_all_total": 9, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_land": 1.0, + "train_accuracy_recipient_token_land_total": 4, + "train_accuracy_recipient_token_mount": 1.0, + "train_accuracy_recipient_token_mount_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.8003952569169961, + "step": 270, + "train_accuracy_first_token_arguments": 0.9090909090909091, + "train_accuracy_first_token_arguments_total": 11, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_go": 1.0, + "train_accuracy_recipient_token_go_total": 3, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.8033596837944664, + "grad_norm": 1.941454727605692, + "learning_rate": 4.888462296487129e-07, + "loss": 0.2125, + "step": 271 + }, + { + "epoch": 0.8033596837944664, + "step": 271, + "train_accuracy_first_token_arguments": 0.9411764705882353, + "train_accuracy_first_token_arguments_total": 34, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.8033596837944664, + "step": 271, + "train_accuracy_first_token_arguments": 0.7407407407407407, + "train_accuracy_first_token_arguments_total": 27, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.8033596837944664, + "step": 271, + "train_accuracy_first_token_arguments": 0.7272727272727273, + "train_accuracy_first_token_arguments_total": 33, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 2, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 2, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.8063241106719368, + "grad_norm": 1.9052339985594693, + "learning_rate": 4.746290916694368e-07, + "loss": 0.2996, + "step": 272 + }, + { + "epoch": 0.8063241106719368, + "step": 272, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 10, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.7272727272727273, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.8063241106719368, + "step": 272, + "train_accuracy_first_token_arguments": 0.8333333333333334, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 1, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.8063241106719368, + "step": 272, + "train_accuracy_first_token_arguments": 0.8181818181818182, + "train_accuracy_first_token_arguments_total": 11, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 2, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.8092885375494071, + "grad_norm": 2.933908602104762, + "learning_rate": 4.6060004370984763e-07, + "loss": 0.3296, + "step": 273 + }, + { + "epoch": 0.8092885375494071, + "step": 273, + "train_accuracy_first_token_arguments": 0.9629629629629629, + "train_accuracy_first_token_arguments_total": 27, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_convert": 1.0, + "train_accuracy_recipient_token_convert_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.8092885375494071, + "step": 273, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 9, + "train_accuracy_recipient_token": 0.7857142857142857, + "train_accuracy_recipient_token_all": 0.7, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_data": 1.0, + "train_accuracy_recipient_token_data_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 14 + }, + { + "epoch": 0.8092885375494071, + "step": 273, + "train_accuracy_first_token_arguments": 0.9, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.8122529644268774, + "grad_norm": 2.787651925675771, + "learning_rate": 4.467603886048452e-07, + "loss": 0.2215, + "step": 274 + }, + { + "epoch": 0.8122529644268774, + "step": 274, + "train_accuracy_first_token_arguments": 0.9583333333333334, + "train_accuracy_first_token_arguments_total": 24, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_h": 1.0, + "train_accuracy_recipient_token_h_total": 4, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.8122529644268774, + "step": 274, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 0.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.8122529644268774, + "step": 274, + "train_accuracy_first_token_arguments": 0.9473684210526315, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_crop": 1.0, + "train_accuracy_recipient_token_crop_total": 1, + "train_accuracy_recipient_token_f": 1.0, + "train_accuracy_recipient_token_f_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_har": 0.0, + "train_accuracy_recipient_token_har_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.8152173913043478, + "grad_norm": 2.0435081993694, + "learning_rate": 4.331114116009938e-07, + "loss": 0.1421, + "step": 275 + }, + { + "epoch": 0.8152173913043478, + "step": 275, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 10, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 10, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.8152173913043478, + "step": 275, + "train_accuracy_first_token_arguments": 0.9333333333333333, + "train_accuracy_first_token_arguments_total": 30, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.8152173913043478, + "step": 275, + "train_accuracy_first_token_arguments": 0.5925925925925926, + "train_accuracy_first_token_arguments_total": 27, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_get": 0.5, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_sum": 1.0, + "train_accuracy_recipient_token_sum_total": 1, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.8181818181818182, + "grad_norm": 2.30706005395235, + "learning_rate": 4.196543802371641e-07, + "loss": 0.2433, + "step": 276 + }, + { + "epoch": 0.8181818181818182, + "step": 276, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.8181818181818182, + "step": 276, + "train_accuracy_first_token_arguments": 0.7333333333333333, + "train_accuracy_first_token_arguments_total": 45, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_rail": 1.0, + "train_accuracy_recipient_token_rail_total": 4, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.8181818181818182, + "step": 276, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.8211462450592886, + "grad_norm": 1.9811813651011425, + "learning_rate": 4.063905442268201e-07, + "loss": 0.2254, + "step": 277 + }, + { + "epoch": 0.8211462450592886, + "step": 277, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 28, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 0.75, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.8211462450592886, + "step": 277, + "train_accuracy_first_token_arguments": 0.8095238095238095, + "train_accuracy_first_token_arguments_total": 21, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_motor": 1.0, + "train_accuracy_recipient_token_motor_total": 3, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.8211462450592886, + "step": 277, + "train_accuracy_first_token_arguments": 0.9130434782608695, + "train_accuracy_first_token_arguments_total": 23, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 4, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.8241106719367589, + "grad_norm": 1.7163941442053223, + "learning_rate": 3.9332113534196194e-07, + "loss": 0.2561, + "step": 278 + }, + { + "epoch": 0.8241106719367589, + "step": 278, + "train_accuracy_first_token_arguments": 0.9444444444444444, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 0.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_get": 0.875, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.8241106719367589, + "step": 278, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 25, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.8241106719367589, + "step": 278, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 9, + "train_accuracy_recipient_token_be": 0.6666666666666666, + "train_accuracy_recipient_token_be_total": 3, + "train_accuracy_recipient_token_find": 0.8, + "train_accuracy_recipient_token_find_total": 5, + "train_accuracy_recipient_token_suggest": 1.0, + "train_accuracy_recipient_token_suggest_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.8270750988142292, + "grad_norm": 1.6379930568149086, + "learning_rate": 3.80447367298738e-07, + "loss": 0.1022, + "step": 279 + }, + { + "epoch": 0.8270750988142292, + "step": 279, + "train_accuracy_first_token_arguments": 0.8823529411764706, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_create": 1.0, + "train_accuracy_recipient_token_create_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 5, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.8270750988142292, + "step": 279, + "train_accuracy_first_token_arguments": 0.9047619047619048, + "train_accuracy_first_token_arguments_total": 21, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_gr": 1.0, + "train_accuracy_recipient_token_gr_total": 2, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_interpret": 1.0, + "train_accuracy_recipient_token_interpret_total": 1, + "train_accuracy_recipient_token_remove": 1.0, + "train_accuracy_recipient_token_remove_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.8270750988142292, + "step": 279, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 22, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.8300395256916996, + "grad_norm": 1.6541675532013322, + "learning_rate": 3.677704356447254e-07, + "loss": 0.1926, + "step": 280 + }, + { + "epoch": 0.8300395256916996, + "step": 280, + "train_accuracy_first_token_arguments": 0.7903225806451613, + "train_accuracy_first_token_arguments_total": 62, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_f": 1.0, + "train_accuracy_recipient_token_f_total": 2, + "train_accuracy_recipient_token_generate": 1.0, + "train_accuracy_recipient_token_generate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_port": 1.0, + "train_accuracy_recipient_token_port_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.8300395256916996, + "step": 280, + "train_accuracy_first_token_arguments": 0.9833333333333333, + "train_accuracy_first_token_arguments_total": 120, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 2, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.8300395256916996, + "step": 280, + "train_accuracy_first_token_arguments": 0.8846153846153846, + "train_accuracy_first_token_arguments_total": 26, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 9, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 3, + "train_accuracy_recipient_token_convert": 1.0, + "train_accuracy_recipient_token_convert_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_set": 1.0, + "train_accuracy_recipient_token_set_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.8330039525691699, + "grad_norm": 2.430973200540749, + "learning_rate": 3.5529151764790715e-07, + "loss": 0.282, + "step": 281 + }, + { + "epoch": 0.8330039525691699, + "step": 281, + "train_accuracy_first_token_arguments": 0.9166666666666666, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.9375, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_aqu": 1.0, + "train_accuracy_recipient_token_aqu_total": 1, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_fish": 1.0, + "train_accuracy_recipient_token_fish_total": 3, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.8330039525691699, + "step": 281, + "train_accuracy_first_token_arguments": 0.9393939393939394, + "train_accuracy_first_token_arguments_total": 33, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9166666666666666, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_wild": 1.0, + "train_accuracy_recipient_token_wild_total": 2, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.8330039525691699, + "step": 281, + "train_accuracy_first_token_arguments": 0.8181818181818182, + "train_accuracy_first_token_arguments_total": 11, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 10, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.8359683794466403, + "grad_norm": 2.0245615896487577, + "learning_rate": 3.43011772187343e-07, + "loss": 0.3178, + "step": 282 + }, + { + "epoch": 0.8359683794466403, + "step": 282, + "train_accuracy_first_token_arguments": 0.8947368421052632, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.8333333333333334, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.8359683794466403, + "step": 282, + "train_accuracy_first_token_arguments": 0.8571428571428571, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_pill": 1.0, + "train_accuracy_recipient_token_pill_total": 2, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.8359683794466403, + "step": 282, + "train_accuracy_first_token_arguments": 0.8947368421052632, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_evaluate": 1.0, + "train_accuracy_recipient_token_evaluate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_plan": 1.0, + "train_accuracy_recipient_token_plan_total": 1, + "train_accuracy_recipient_token_suggest": 1.0, + "train_accuracy_recipient_token_suggest_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.8389328063241107, + "grad_norm": 2.1514655537590763, + "learning_rate": 3.3093233964554464e-07, + "loss": 0.2761, + "step": 283 + }, + { + "epoch": 0.8389328063241107, + "step": 283, + "train_accuracy_first_token_arguments": 0.9375, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 0.5, + "train_accuracy_recipient_token_calculate_total": 2, + "train_accuracy_recipient_token_estimate": 0.5, + "train_accuracy_recipient_token_estimate_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_optimize": 1.0, + "train_accuracy_recipient_token_optimize_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.8389328063241107, + "step": 283, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 27, + "train_accuracy_recipient_token": 0.75, + "train_accuracy_recipient_token_all": 0.6363636363636364, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_conduct": 1.0, + "train_accuracy_recipient_token_conduct_total": 1, + "train_accuracy_recipient_token_d": 1.0, + "train_accuracy_recipient_token_d_total": 1, + "train_accuracy_recipient_token_landscape": 1.0, + "train_accuracy_recipient_token_landscape_total": 1, + "train_accuracy_recipient_token_organ": 1.0, + "train_accuracy_recipient_token_organ_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.8389328063241107, + "step": 283, + "train_accuracy_first_token_arguments": 0.8421052631578947, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.8888888888888888, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.841897233201581, + "grad_norm": 2.2087100686535135, + "learning_rate": 3.190543418025749e-07, + "loss": 0.2386, + "step": 284 + }, + { + "epoch": 0.841897233201581, + "step": 284, + "train_accuracy_first_token_arguments": 0.8666666666666667, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_dec": 1.0, + "train_accuracy_recipient_token_dec_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.841897233201581, + "step": 284, + "train_accuracy_first_token_arguments": 0.8571428571428571, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 0.875, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.841897233201581, + "step": 284, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 10, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.8888888888888888, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.8448616600790514, + "grad_norm": 2.1863171999841655, + "learning_rate": 3.0737888173187067e-07, + "loss": 0.1506, + "step": 285 + }, + { + "epoch": 0.8448616600790514, + "step": 285, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 11, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_l": 1.0, + "train_accuracy_recipient_token_l_total": 1, + "train_accuracy_recipient_token_remove": 1.0, + "train_accuracy_recipient_token_remove_total": 1, + "train_accuracy_recipient_token_write": 1.0, + "train_accuracy_recipient_token_write_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.8448616600790514, + "step": 285, + "train_accuracy_first_token_arguments": 0.6666666666666666, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_detect": 1.0, + "train_accuracy_recipient_token_detect_total": 1, + "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.8448616600790514, + "step": 285, + "train_accuracy_first_token_arguments": 0.76, + "train_accuracy_first_token_arguments_total": 25, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_write": 1.0, + "train_accuracy_recipient_token_write_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.8478260869565217, + "grad_norm": 1.3956004649176736, + "learning_rate": 2.9590704369780313e-07, + "loss": 0.1174, + "step": 286 + }, + { + "epoch": 0.8478260869565217, + "step": 286, + "train_accuracy_first_token_arguments": 0.8235294117647058, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 0.9166666666666666, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_paint": 1.0, + "train_accuracy_recipient_token_paint_total": 1, + "train_accuracy_recipient_token_rep": 1.0, + "train_accuracy_recipient_token_rep_total": 1, + "train_accuracy_recipient_token_surface": 1.0, + "train_accuracy_recipient_token_surface_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.8478260869565217, + "step": 286, + "train_accuracy_first_token_arguments": 0.9333333333333333, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.875, + "train_accuracy_recipient_token_all": 0.8333333333333334, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.8478260869565217, + "step": 286, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.8507905138339921, + "grad_norm": 1.8154026629925726, + "learning_rate": 2.8463989305498596e-07, + "loss": 0.2343, + "step": 287 + }, + { + "epoch": 0.8507905138339921, + "step": 287, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.8507905138339921, + "step": 287, + "train_accuracy_first_token_arguments": 0.8260869565217391, + "train_accuracy_first_token_arguments_total": 23, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_order": 1.0, + "train_accuracy_recipient_token_order_total": 1, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 2, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.8507905138339921, + "step": 287, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 10, + "train_accuracy_recipient_token": 0.875, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.8537549407114624, + "grad_norm": 2.1946988449253766, + "learning_rate": 2.7357847614933876e-07, + "loss": 0.2904, + "step": 288 + }, + { + "epoch": 0.8537549407114624, + "step": 288, + "train_accuracy_first_token_arguments": 0.8666666666666667, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_f": 1.0, + "train_accuracy_recipient_token_f_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_mul": 1.0, + "train_accuracy_recipient_token_mul_total": 1, + "train_accuracy_recipient_token_weed": 1.0, + "train_accuracy_recipient_token_weed_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.8537549407114624, + "step": 288, + "train_accuracy_first_token_arguments": 0.8846153846153846, + "train_accuracy_first_token_arguments_total": 26, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_histor": 1.0, + "train_accuracy_recipient_token_histor_total": 1, + "train_accuracy_recipient_token_land": 1.0, + "train_accuracy_recipient_token_land_total": 3, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.8537549407114624, + "step": 288, + "train_accuracy_first_token_arguments": 0.85, + "train_accuracy_first_token_arguments_total": 40, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9166666666666666, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_check": 0.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.8567193675889329, + "grad_norm": 1.8692598511995904, + "learning_rate": 2.6272382022091704e-07, + "loss": 0.2141, + "step": 289 + }, + { + "epoch": 0.8567193675889329, + "step": 289, + "train_accuracy_first_token_arguments": 0.9310344827586207, + "train_accuracy_first_token_arguments_total": 29, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_repair": 1.0, + "train_accuracy_recipient_token_repair_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.8567193675889329, + "step": 289, + "train_accuracy_first_token_arguments": 0.75, + "train_accuracy_first_token_arguments_total": 36, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_design": 1.0, + "train_accuracy_recipient_token_design_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 5, + "train_accuracy_recipient_token_play": 1.0, + "train_accuracy_recipient_token_play_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.8567193675889329, + "step": 289, + "train_accuracy_first_token_arguments": 0.92, + "train_accuracy_first_token_arguments_total": 25, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_train": 0.75, + "train_accuracy_recipient_token_train_total": 4, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.8596837944664032, + "grad_norm": 1.4764081056636353, + "learning_rate": 2.520769333085141e-07, + "loss": 0.1611, + "step": 290 + }, + { + "epoch": 0.8596837944664032, + "step": 290, + "train_accuracy_first_token_arguments": 0.9626168224299065, + "train_accuracy_first_token_arguments_total": 107, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_check": 0.5, + "train_accuracy_recipient_token_check_total": 2, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_mach": 0.6666666666666666, + "train_accuracy_recipient_token_mach_total": 3, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.8596837944664032, + "step": 290, + "train_accuracy_first_token_arguments": 0.8666666666666667, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.9, + "train_accuracy_recipient_token_get_total": 10, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.8596837944664032, + "step": 290, + "train_accuracy_first_token_arguments": 0.75, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_order": 1.0, + "train_accuracy_recipient_token_order_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.8626482213438735, + "grad_norm": 1.920082186381157, + "learning_rate": 2.4163880415604913e-07, + "loss": 0.266, + "step": 291 + }, + { + "epoch": 0.8626482213438735, + "step": 291, + "train_accuracy_first_token_arguments": 0.7333333333333333, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_pig": 1.0, + "train_accuracy_recipient_token_pig_total": 3, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.8626482213438735, + "step": 291, + "train_accuracy_first_token_arguments": 0.9680851063829787, + "train_accuracy_first_token_arguments_total": 94, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_accuracy_recipient_token_track": 1.0, + "train_accuracy_recipient_token_track_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.8626482213438735, + "step": 291, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 10, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.8656126482213439, + "grad_norm": 2.4496922640460923, + "learning_rate": 2.3141040212074445e-07, + "loss": 0.3954, + "step": 292 + }, + { + "epoch": 0.8656126482213439, + "step": 292, + "train_accuracy_first_token_arguments": 0.8, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_inspect": 1.0, + "train_accuracy_recipient_token_inspect_total": 1, + "train_accuracy_recipient_token_remove": 1.0, + "train_accuracy_recipient_token_remove_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.8656126482213439, + "step": 292, + "train_accuracy_first_token_arguments": 0.8666666666666667, + "train_accuracy_first_token_arguments_total": 30, + "train_accuracy_recipient_token": 0.8260869565217391, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 3, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_multiply": 0.5, + "train_accuracy_recipient_token_multiply_total": 4, + "train_accuracy_recipient_token_suggest": 1.0, + "train_accuracy_recipient_token_suggest_total": 2, + "train_total_number_recipient_token": 23 + }, + { + "epoch": 0.8656126482213439, + "step": 292, + "train_accuracy_first_token_arguments": 0.90625, + "train_accuracy_first_token_arguments_total": 32, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 9, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 2, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 5, + "train_accuracy_recipient_token_join": 0.0, + "train_accuracy_recipient_token_join_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.8685770750988142, + "grad_norm": 2.4189295709433183, + "learning_rate": 2.2139267708310457e-07, + "loss": 0.1385, + "step": 293 + }, + { + "epoch": 0.8685770750988142, + "step": 293, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 10, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_locate": 1.0, + "train_accuracy_recipient_token_locate_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.8685770750988142, + "step": 293, + "train_accuracy_first_token_arguments": 0.9629629629629629, + "train_accuracy_first_token_arguments_total": 27, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_order": 1.0, + "train_accuracy_recipient_token_order_total": 1, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.8685770750988142, + "step": 293, + "train_accuracy_first_token_arguments": 0.8571428571428571, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_text": 1.0, + "train_accuracy_recipient_token_text_total": 6, + "train_total_number_recipient_token": 21 + }, + { + "epoch": 0.8715415019762845, + "grad_norm": 2.1088624243937426, + "learning_rate": 2.1158655935870325e-07, + "loss": 0.2126, + "step": 294 + }, + { + "epoch": 0.8715415019762845, + "step": 294, + "train_accuracy_first_token_arguments": 0.8947368421052632, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.8715415019762845, + "step": 294, + "train_accuracy_first_token_arguments": 0.7230769230769231, + "train_accuracy_first_token_arguments_total": 65, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_draw": 1.0, + "train_accuracy_recipient_token_draw_total": 5, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.8715415019762845, + "step": 294, + "train_accuracy_first_token_arguments": 0.9444444444444444, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 10, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.8745059288537549, + "grad_norm": 2.032905445433328, + "learning_rate": 2.0199295961178893e-07, + "loss": 0.2617, + "step": 295 + }, + { + "epoch": 0.8745059288537549, + "step": 295, + "train_accuracy_first_token_arguments": 0.9166666666666666, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_locate": 1.0, + "train_accuracy_recipient_token_locate_total": 1, + "train_accuracy_recipient_token_suggest": 1.0, + "train_accuracy_recipient_token_suggest_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.8745059288537549, + "step": 295, + "train_accuracy_first_token_arguments": 0.9210526315789473, + "train_accuracy_first_token_arguments_total": 38, + "train_accuracy_recipient_token": 0.9333333333333333, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_design": 1.0, + "train_accuracy_recipient_token_design_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_extract": 0.5, + "train_accuracy_recipient_token_extract_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_total_number_recipient_token": 15 + }, + { + "epoch": 0.8745059288537549, + "step": 295, + "train_accuracy_first_token_arguments": 0.7272727272727273, + "train_accuracy_first_token_arguments_total": 22, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_att": 1.0, + "train_accuracy_recipient_token_att_total": 1, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_explain": 1.0, + "train_accuracy_recipient_token_explain_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.8774703557312253, + "grad_norm": 2.284386546160239, + "learning_rate": 1.9261276877071354e-07, + "loss": 0.3743, + "step": 296 + }, + { + "epoch": 0.8774703557312253, + "step": 296, + "train_accuracy_first_token_arguments": 0.8, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_learn": 1.0, + "train_accuracy_recipient_token_learn_total": 1, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.8774703557312253, + "step": 296, + "train_accuracy_first_token_arguments": 0.9285714285714286, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.8774703557312253, + "step": 296, + "train_accuracy_first_token_arguments": 0.9069767441860465, + "train_accuracy_first_token_arguments_total": 43, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8333333333333334, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.8804347826086957, + "grad_norm": 1.6291217601575605, + "learning_rate": 1.8344685794519507e-07, + "loss": 0.1641, + "step": 297 + }, + { + "epoch": 0.8804347826086957, + "step": 297, + "train_accuracy_first_token_arguments": 0.9649122807017544, + "train_accuracy_first_token_arguments_total": 57, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 3, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_get": 0.5, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 2, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.8804347826086957, + "step": 297, + "train_accuracy_first_token_arguments": 0.825, + "train_accuracy_first_token_arguments_total": 40, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_accuracy_recipient_token_sort": 0.0, + "train_accuracy_recipient_token_sort_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.8804347826086957, + "step": 297, + "train_accuracy_first_token_arguments": 0.8648648648648649, + "train_accuracy_first_token_arguments_total": 37, + "train_accuracy_recipient_token": 0.85, + "train_accuracy_recipient_token_all": 0.9230769230769231, + "train_accuracy_recipient_token_all_total": 13, + "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_restaurant": 1.0, + "train_accuracy_recipient_token_restaurant_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.883399209486166, + "grad_norm": 1.7794723846623823, + "learning_rate": 1.744960783454186e-07, + "loss": 0.2034, + "step": 298 + }, + { + "epoch": 0.883399209486166, + "step": 298, + "train_accuracy_first_token_arguments": 0.8421052631578947, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 9, + "train_accuracy_recipient_token_identify": 0.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.883399209486166, + "step": 298, + "train_accuracy_first_token_arguments": 0.7872340425531915, + "train_accuracy_first_token_arguments_total": 47, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8333333333333334, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_building": 1.0, + "train_accuracy_recipient_token_building_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_stat": 1.0, + "train_accuracy_recipient_token_stat_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.883399209486166, + "step": 298, + "train_accuracy_first_token_arguments": 0.8837209302325582, + "train_accuracy_first_token_arguments_total": 43, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compile": 1.0, + "train_accuracy_recipient_token_compile_total": 1, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 1, + "train_accuracy_recipient_token_restaurant": 1.0, + "train_accuracy_recipient_token_restaurant_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.8863636363636364, + "grad_norm": 1.9655423422483125, + "learning_rate": 1.6576126120299046e-07, + "loss": 0.3209, + "step": 299 + }, + { + "epoch": 0.8863636363636364, + "step": 299, + "train_accuracy_first_token_arguments": 0.8333333333333334, + "train_accuracy_first_token_arguments_total": 30, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_draw": 1.0, + "train_accuracy_recipient_token_draw_total": 5, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.8863636363636364, + "step": 299, + "train_accuracy_first_token_arguments": 0.7777777777777778, + "train_accuracy_first_token_arguments_total": 9, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_exp": 1.0, + "train_accuracy_recipient_token_exp_total": 3, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_translate": 0.0, + "train_accuracy_recipient_token_translate_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.8863636363636364, + "step": 299, + "train_accuracy_first_token_arguments": 0.9473684210526315, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_plant": 1.0, + "train_accuracy_recipient_token_plant_total": 3, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.8893280632411067, + "grad_norm": 2.3395853773637407, + "learning_rate": 1.5724321769374023e-07, + "loss": 0.3217, + "step": 300 + }, + { + "epoch": 0.8893280632411067, + "step": 300, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 9, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.7, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.8893280632411067, + "step": 300, + "train_accuracy_first_token_arguments": 0.9, + "train_accuracy_first_token_arguments_total": 30, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_bird": 1.0, + "train_accuracy_recipient_token_bird_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.8893280632411067, + "step": 300, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 10, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.892292490118577, + "grad_norm": 2.265776160882912, + "learning_rate": 1.4894273886239208e-07, + "loss": 0.3318, + "step": 301 + }, + { + "epoch": 0.892292490118577, + "step": 301, + "train_accuracy_first_token_arguments": 0.9375, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.892292490118577, + "step": 301, + "train_accuracy_first_token_arguments": 0.8888888888888888, + "train_accuracy_first_token_arguments_total": 9, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8333333333333334, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.892292490118577, + "step": 301, + "train_accuracy_first_token_arguments": 0.9, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.8333333333333334, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_list": 1.0, + "train_accuracy_recipient_token_list_total": 1, + "train_accuracy_recipient_token_stat": 1.0, + "train_accuracy_recipient_token_stat_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.8952569169960475, + "grad_norm": 2.3746950651302927, + "learning_rate": 1.4086059554910186e-07, + "loss": 0.3669, + "step": 302 + }, + { + "epoch": 0.8952569169960475, + "step": 302, + "train_accuracy_first_token_arguments": 0.6590909090909091, + "train_accuracy_first_token_arguments_total": 44, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_object": 1.0, + "train_accuracy_recipient_token_object_total": 1, + "train_accuracy_recipient_token_people": 1.0, + "train_accuracy_recipient_token_people_total": 2, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.8952569169960475, + "step": 302, + "train_accuracy_first_token_arguments": 0.9285714285714286, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_verify": 1.0, + "train_accuracy_recipient_token_verify_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.8952569169960475, + "step": 302, + "train_accuracy_first_token_arguments": 0.9090909090909091, + "train_accuracy_first_token_arguments_total": 11, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_building": 1.0, + "train_accuracy_recipient_token_building_total": 4, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.8982213438735178, + "grad_norm": 2.5356426721293994, + "learning_rate": 1.3299753831787193e-07, + "loss": 0.2913, + "step": 303 + }, + { + "epoch": 0.8982213438735178, + "step": 303, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.8982213438735178, + "step": 303, + "train_accuracy_first_token_arguments": 0.8947368421052632, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 0.875, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.8982213438735178, + "step": 303, + "train_accuracy_first_token_arguments": 0.9230769230769231, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 2, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.9011857707509882, + "grad_norm": 2.3231357183648993, + "learning_rate": 1.2535429738684822e-07, + "loss": 0.2729, + "step": 304 + }, + { + "epoch": 0.9011857707509882, + "step": 304, + "train_accuracy_first_token_arguments": 0.8571428571428571, + "train_accuracy_first_token_arguments_total": 21, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 3, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.9011857707509882, + "step": 304, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 6, + "train_accuracy_recipient_token": 0.8, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_recogn": 1.0, + "train_accuracy_recipient_token_recogn_total": 1, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 1, + "train_total_number_recipient_token": 15 + }, + { + "epoch": 0.9011857707509882, + "step": 304, + "train_accuracy_first_token_arguments": 0.9444444444444444, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.9041501976284585, + "grad_norm": 2.0889543788782716, + "learning_rate": 1.1793158256050708e-07, + "loss": 0.3167, + "step": 305 + }, + { + "epoch": 0.9041501976284585, + "step": 305, + "train_accuracy_first_token_arguments": 0.9285714285714286, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 10, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.9041501976284585, + "step": 305, + "train_accuracy_first_token_arguments": 0.8059701492537313, + "train_accuracy_first_token_arguments_total": 67, + "train_accuracy_recipient_token": 0.9375, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_event": 1.0, + "train_accuracy_recipient_token_event_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.9041501976284585, + "step": 305, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_car": 1.0, + "train_accuracy_recipient_token_car_total": 1, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.9071146245059288, + "grad_norm": 1.546141851099797, + "learning_rate": 1.1073008316373812e-07, + "loss": 0.1956, + "step": 306 + }, + { + "epoch": 0.9071146245059288, + "step": 306, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 30, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_stat": 1.0, + "train_accuracy_recipient_token_stat_total": 3, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.9071146245059288, + "step": 306, + "train_accuracy_first_token_arguments": 0.9166666666666666, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.9071146245059288, + "step": 306, + "train_accuracy_first_token_arguments": 0.7857142857142857, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.9523809523809523, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 2, + "train_accuracy_recipient_token_search": 0.0, + "train_accuracy_recipient_token_search_total": 1, + "train_total_number_recipient_token": 21 + }, + { + "epoch": 0.9100790513833992, + "grad_norm": 2.2381733984414756, + "learning_rate": 1.0375046797782868e-07, + "loss": 0.2507, + "step": 307 + }, + { + "epoch": 0.9100790513833992, + "step": 307, + "train_accuracy_first_token_arguments": 0.8947368421052632, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 2, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_count": 1.0, + "train_accuracy_recipient_token_count_total": 1, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.9100790513833992, + "step": 307, + "train_accuracy_first_token_arguments": 0.8, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_gr": 1.0, + "train_accuracy_recipient_token_gr_total": 3, + "train_accuracy_recipient_token_identify": 0.5, + "train_accuracy_recipient_token_identify_total": 2, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.9100790513833992, + "step": 307, + "train_accuracy_first_token_arguments": 0.9565217391304348, + "train_accuracy_first_token_arguments_total": 23, + "train_accuracy_recipient_token": 0.9523809523809523, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.9090909090909091, + "train_accuracy_recipient_token_get_total": 11, + "train_total_number_recipient_token": 21 + }, + { + "epoch": 0.9130434782608695, + "grad_norm": 1.6564222632639842, + "learning_rate": 9.699338517835611e-08, + "loss": 0.2674, + "step": 308 + }, + { + "epoch": 0.9130434782608695, + "step": 308, + "train_accuracy_first_token_arguments": 0.9166666666666666, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_verify": 1.0, + "train_accuracy_recipient_token_verify_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.9130434782608695, + "step": 308, + "train_accuracy_first_token_arguments": 0.8571428571428571, + "train_accuracy_first_token_arguments_total": 28, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_arr": 0.0, + "train_accuracy_recipient_token_arr_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_house": 1.0, + "train_accuracy_recipient_token_house_total": 3, + "train_accuracy_recipient_token_landscape": 1.0, + "train_accuracy_recipient_token_landscape_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.9130434782608695, + "step": 308, + "train_accuracy_first_token_arguments": 0.8125, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_market": 1.0, + "train_accuracy_recipient_token_market_total": 3, + "train_accuracy_recipient_token_verify": 1.0, + "train_accuracy_recipient_token_verify_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.91600790513834, + "grad_norm": 1.503775394527991, + "learning_rate": 9.045946227499298e-08, + "loss": 0.1276, + "step": 309 + }, + { + "epoch": 0.91600790513834, + "step": 309, + "train_accuracy_first_token_arguments": 0.8947368421052632, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 0.8888888888888888, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.91600790513834, + "step": 309, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 10, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.91600790513834, + "step": 309, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 11, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_research": 1.0, + "train_accuracy_recipient_token_research_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.9189723320158103, + "grad_norm": 2.2000639611250423, + "learning_rate": 8.414930605323445e-08, + "loss": 0.2596, + "step": 310 + }, + { + "epoch": 0.9189723320158103, + "step": 310, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 4, + "train_accuracy_recipient_token": 0.9285714285714286, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_extract": 0.5, + "train_accuracy_recipient_token_extract_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 14 + }, + { + "epoch": 0.9189723320158103, + "step": 310, + "train_accuracy_first_token_arguments": 0.7272727272727273, + "train_accuracy_first_token_arguments_total": 11, + "train_accuracy_recipient_token": 0.8125, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_explain": 0.6666666666666666, + "train_accuracy_recipient_token_explain_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_suggest": 1.0, + "train_accuracy_recipient_token_suggest_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.9189723320158103, + "step": 310, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 23, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.9219367588932806, + "grad_norm": 2.4294485936947194, + "learning_rate": 7.806350251804484e-08, + "loss": 0.3957, + "step": 311 + }, + { + "epoch": 0.9219367588932806, + "step": 311, + "train_accuracy_first_token_arguments": 0.7619047619047619, + "train_accuracy_first_token_arguments_total": 42, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.9219367588932806, + "step": 311, + "train_accuracy_first_token_arguments": 0.8235294117647058, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_property": 1.0, + "train_accuracy_recipient_token_property_total": 1, + "train_accuracy_recipient_token_schedule": 0.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.9219367588932806, + "step": 311, + "train_accuracy_first_token_arguments": 0.9444444444444444, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.924901185770751, + "grad_norm": 2.1844882128572, + "learning_rate": 7.220261683943935e-08, + "loss": 0.26, + "step": 312 + }, + { + "epoch": 0.924901185770751, + "step": 312, + "train_accuracy_first_token_arguments": 0.6666666666666666, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_ap": 1.0, + "train_accuracy_recipient_token_ap_total": 1, + "train_accuracy_recipient_token_ger": 1.0, + "train_accuracy_recipient_token_ger_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.924901185770751, + "step": 312, + "train_accuracy_first_token_arguments": 0.875, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_sand": 0.75, + "train_accuracy_recipient_token_sand_total": 4, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.924901185770751, + "step": 312, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.8888888888888888, + "train_accuracy_recipient_token_get_total": 9, + "train_accuracy_recipient_token_identify": 0.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.9278656126482213, + "grad_norm": 1.842757768341881, + "learning_rate": 6.6567193299997e-08, + "loss": 0.1724, + "step": 313 + }, + { + "epoch": 0.9278656126482213, + "step": 313, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 8, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.9278656126482213, + "step": 313, + "train_accuracy_first_token_arguments": 0.5833333333333334, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_arr": 0.0, + "train_accuracy_recipient_token_arr_total": 1, + "train_accuracy_recipient_token_marine": 1.0, + "train_accuracy_recipient_token_marine_total": 1, + "train_accuracy_recipient_token_o": 1.0, + "train_accuracy_recipient_token_o_total": 1, + "train_accuracy_recipient_token_under": 1.0, + "train_accuracy_recipient_token_under_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.9278656126482213, + "step": 313, + "train_accuracy_first_token_arguments": 0.8571428571428571, + "train_accuracy_first_token_arguments_total": 35, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 0.5, + "train_accuracy_recipient_token_compare_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.9308300395256917, + "grad_norm": 2.101888459848309, + "learning_rate": 6.115775524431711e-08, + "loss": 0.3051, + "step": 314 + }, + { + "epoch": 0.9308300395256917, + "step": 314, + "train_accuracy_first_token_arguments": 0.9090909090909091, + "train_accuracy_first_token_arguments_total": 22, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_plan": 1.0, + "train_accuracy_recipient_token_plan_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.9308300395256917, + "step": 314, + "train_accuracy_first_token_arguments": 0.7857142857142857, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_stat": 0.5, + "train_accuracy_recipient_token_stat_total": 2, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.9308300395256917, + "step": 314, + "train_accuracy_first_token_arguments": 0.9411764705882353, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_arch": 1.0, + "train_accuracy_recipient_token_arch_total": 2, + "train_accuracy_recipient_token_building": 1.0, + "train_accuracy_recipient_token_building_total": 3, + "train_accuracy_recipient_token_gr": 1.0, + "train_accuracy_recipient_token_gr_total": 3, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.9337944664031621, + "grad_norm": 1.8025117128414663, + "learning_rate": 5.597480503041486e-08, + "loss": 0.3276, + "step": 315 + }, + { + "epoch": 0.9337944664031621, + "step": 315, + "train_accuracy_first_token_arguments": 0.7647058823529411, + "train_accuracy_first_token_arguments_total": 102, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_inter": 1.0, + "train_accuracy_recipient_token_inter_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.9337944664031621, + "step": 315, + "train_accuracy_first_token_arguments": 0.8461538461538461, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.9337944664031621, + "step": 315, + "train_accuracy_first_token_arguments": 0.9473684210526315, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_fabric": 1.0, + "train_accuracy_recipient_token_fabric_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_send": 1.0, + "train_accuracy_recipient_token_send_total": 1, + "train_accuracy_recipient_token_set": 1.0, + "train_accuracy_recipient_token_set_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.9367588932806324, + "grad_norm": 2.303122934120546, + "learning_rate": 5.1018823983070285e-08, + "loss": 0.2539, + "step": 316 + }, + { + "epoch": 0.9367588932806324, + "step": 316, + "train_accuracy_first_token_arguments": 0.7, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.9367588932806324, + "step": 316, + "train_accuracy_first_token_arguments": 0.9166666666666666, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_format": 1.0, + "train_accuracy_recipient_token_format_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_sum": 1.0, + "train_accuracy_recipient_token_sum_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.9367588932806324, + "step": 316, + "train_accuracy_first_token_arguments": 0.7, + "train_accuracy_first_token_arguments_total": 10, + "train_accuracy_recipient_token": 0.7777777777777778, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_arr": 0.0, + "train_accuracy_recipient_token_arr_total": 1, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.9397233201581028, + "grad_norm": 1.7793933316663082, + "learning_rate": 4.629027234912986e-08, + "loss": 0.1763, + "step": 317 + }, + { + "epoch": 0.9397233201581028, + "step": 317, + "train_accuracy_first_token_arguments": 0.95, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.9166666666666666, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_compare": 0.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.9397233201581028, + "step": 317, + "train_accuracy_first_token_arguments": 0.7428571428571429, + "train_accuracy_first_token_arguments_total": 35, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_explain": 1.0, + "train_accuracy_recipient_token_explain_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_plane": 1.0, + "train_accuracy_recipient_token_plane_total": 3, + "train_accuracy_recipient_token_zi": 0.0, + "train_accuracy_recipient_token_zi_total": 2, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.9397233201581028, + "step": 317, + "train_accuracy_first_token_arguments": 0.8888888888888888, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_arr": 1.0, + "train_accuracy_recipient_token_arr_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.9426877470355731, + "grad_norm": 1.5774943505040582, + "learning_rate": 4.178958925476401e-08, + "loss": 0.1834, + "step": 318 + }, + { + "epoch": 0.9426877470355731, + "eval_accuracy_first_token_arguments": 0.9248395967002749, + "eval_accuracy_first_token_arguments_total": 1091, + "eval_accuracy_recipient_token": 0.9401330376940134, + "eval_accuracy_recipient_token_all": 0.9220272904483431, + "eval_accuracy_recipient_token_all_total": 513, + "eval_accuracy_recipient_token_calculate": 1.0, + "eval_accuracy_recipient_token_calculate_total": 13, + "eval_accuracy_recipient_token_find": 1.0, + "eval_accuracy_recipient_token_find_total": 57, + "eval_accuracy_recipient_token_get": 0.9795918367346939, + "eval_accuracy_recipient_token_get_total": 196, + "eval_accuracy_recipient_token_identify": 0.8333333333333334, + "eval_accuracy_recipient_token_identify_total": 6, + "eval_loss": 0.21341286599636078, + "eval_perplexity": 1.0333647517086502, + "eval_runtime": 112.5857, + "eval_samples_per_second": 0.897, + "eval_steps_per_second": 0.453, + "eval_total_number_recipient_token": 902, + "step": 318 + }, + { + "epoch": 0.9426877470355731, + "step": 318, + "train_accuracy_first_token_arguments": 0.7777777777777778, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_identify": 0.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.9426877470355731, + "step": 318, + "train_accuracy_first_token_arguments": 0.9375, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 0.8571428571428571, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.9426877470355731, + "step": 318, + "train_accuracy_first_token_arguments": 0.8461538461538461, + "train_accuracy_first_token_arguments_total": 26, + "train_accuracy_recipient_token": 0.85, + "train_accuracy_recipient_token_all": 0.75, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 4, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.9456521739130435, + "grad_norm": 2.4406440061460377, + "learning_rate": 3.7517192664685844e-08, + "loss": 0.3195, + "step": 319 + }, + { + "epoch": 0.9456521739130435, + "step": 319, + "train_accuracy_first_token_arguments": 0.9375, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.9456521739130435, + "step": 319, + "train_accuracy_first_token_arguments": 0.7142857142857143, + "train_accuracy_first_token_arguments_total": 21, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 2, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.9456521739130435, + "step": 319, + "train_accuracy_first_token_arguments": 0.9375, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_plant": 1.0, + "train_accuracy_recipient_token_plant_total": 3, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.9486166007905138, + "grad_norm": 2.041210168988887, + "learning_rate": 3.347347934333778e-08, + "loss": 0.2436, + "step": 320 + }, + { + "epoch": 0.9486166007905138, + "step": 320, + "train_accuracy_first_token_arguments": 0.9375, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 0.85, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 15, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.9486166007905138, + "step": 320, + "train_accuracy_first_token_arguments": 0.5925925925925926, + "train_accuracy_first_token_arguments_total": 54, + "train_accuracy_recipient_token": 0.8666666666666667, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_predict": 1.0, + "train_accuracy_recipient_token_predict_total": 1, + "train_accuracy_recipient_token_stamp": 0.5, + "train_accuracy_recipient_token_stamp_total": 2, + "train_total_number_recipient_token": 15 + }, + { + "epoch": 0.9486166007905138, + "step": 320, + "train_accuracy_first_token_arguments": 0.9565217391304348, + "train_accuracy_first_token_arguments_total": 23, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.9515810276679841, + "grad_norm": 2.151431626558894, + "learning_rate": 2.9658824818044328e-08, + "loss": 0.309, + "step": 321 + }, + { + "epoch": 0.9515810276679841, + "step": 321, + "train_accuracy_first_token_arguments": 0.9811320754716981, + "train_accuracy_first_token_arguments_total": 53, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_optimize": 0.0, + "train_accuracy_recipient_token_optimize_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.9515810276679841, + "step": 321, + "train_accuracy_first_token_arguments": 0.90625, + "train_accuracy_first_token_arguments_total": 32, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.9515810276679841, + "step": 321, + "train_accuracy_first_token_arguments": 0.9545454545454546, + "train_accuracy_first_token_arguments_total": 22, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.9545454545454546, + "grad_norm": 2.5639718605032216, + "learning_rate": 2.607358334413779e-08, + "loss": 0.3681, + "step": 322 + }, + { + "epoch": 0.9545454545454546, + "step": 322, + "train_accuracy_first_token_arguments": 0.8148148148148148, + "train_accuracy_first_token_arguments_total": 27, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_create": 1.0, + "train_accuracy_recipient_token_create_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ne": 1.0, + "train_accuracy_recipient_token_ne_total": 3, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.9545454545454546, + "step": 322, + "train_accuracy_first_token_arguments": 0.8888888888888888, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 0.9333333333333333, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_convert": 1.0, + "train_accuracy_recipient_token_convert_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 15 + }, + { + "epoch": 0.9545454545454546, + "step": 322, + "train_accuracy_first_token_arguments": 0.9, + "train_accuracy_first_token_arguments_total": 30, + "train_accuracy_recipient_token": 0.9375, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_customize": 1.0, + "train_accuracy_recipient_token_customize_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.9575098814229249, + "grad_norm": 2.8167178374609607, + "learning_rate": 2.2718087872060925e-08, + "loss": 0.2343, + "step": 323 + }, + { + "epoch": 0.9575098814229249, + "step": 323, + "train_accuracy_first_token_arguments": 0.6744186046511628, + "train_accuracy_first_token_arguments_total": 172, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_pool": 1.0, + "train_accuracy_recipient_token_pool_total": 4, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.9575098814229249, + "step": 323, + "train_accuracy_first_token_arguments": 0.9285714285714286, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_robot": 1.0, + "train_accuracy_recipient_token_robot_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.9575098814229249, + "step": 323, + "train_accuracy_first_token_arguments": 0.9583333333333334, + "train_accuracy_first_token_arguments_total": 24, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 0.875, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.9604743083003953, + "grad_norm": 1.6492281178442556, + "learning_rate": 1.9592650016444503e-08, + "loss": 0.1848, + "step": 324 + }, + { + "epoch": 0.9604743083003953, + "step": 324, + "train_accuracy_first_token_arguments": 0.896551724137931, + "train_accuracy_first_token_arguments_total": 29, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_learn": 1.0, + "train_accuracy_recipient_token_learn_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.9604743083003953, + "step": 324, + "train_accuracy_first_token_arguments": 0.9166666666666666, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 0.8888888888888888, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.9604743083003953, + "step": 324, + "train_accuracy_first_token_arguments": 0.5384615384615384, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 0.9047619047619048, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_count": 0.0, + "train_accuracy_recipient_token_count_total": 1, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_subtract": 1.0, + "train_accuracy_recipient_token_subtract_total": 1, + "train_accuracy_recipient_token_track": 1.0, + "train_accuracy_recipient_token_track_total": 1, + "train_total_number_recipient_token": 21 + }, + { + "epoch": 0.9634387351778656, + "grad_norm": 1.754485036239026, + "learning_rate": 1.6697560027171543e-08, + "loss": 0.1782, + "step": 325 + }, + { + "epoch": 0.9634387351778656, + "step": 325, + "train_accuracy_first_token_arguments": 0.9642857142857143, + "train_accuracy_first_token_arguments_total": 28, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 0.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_compile": 1.0, + "train_accuracy_recipient_token_compile_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.9634387351778656, + "step": 325, + "train_accuracy_first_token_arguments": 0.7857142857142857, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_extract": 1.0, + "train_accuracy_recipient_token_extract_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_recogn": 0.0, + "train_accuracy_recipient_token_recogn_total": 1, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.9634387351778656, + "step": 325, + "train_accuracy_first_token_arguments": 0.5333333333333333, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.9333333333333333, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_verify": 1.0, + "train_accuracy_recipient_token_verify_total": 1, + "train_total_number_recipient_token": 15 + }, + { + "epoch": 0.9664031620553359, + "grad_norm": 2.1745450667504445, + "learning_rate": 1.4033086762419989e-08, + "loss": 0.3463, + "step": 326 + }, + { + "epoch": 0.9664031620553359, + "step": 326, + "train_accuracy_first_token_arguments": 0.9322033898305084, + "train_accuracy_first_token_arguments_total": 59, + "train_accuracy_recipient_token": 0.7777777777777778, + "train_accuracy_recipient_token_all": 0.7272727272727273, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 3, + "train_accuracy_recipient_token_check": 0.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.9664031620553359, + "step": 326, + "train_accuracy_first_token_arguments": 0.9375, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compile": 1.0, + "train_accuracy_recipient_token_compile_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.9664031620553359, + "step": 326, + "train_accuracy_first_token_arguments": 0.6875, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_download": 1.0, + "train_accuracy_recipient_token_download_total": 1, + "train_accuracy_recipient_token_get": 0.8333333333333334, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_search": 0.0, + "train_accuracy_recipient_token_search_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.9693675889328063, + "grad_norm": 2.1355214484085425, + "learning_rate": 1.1599477663696845e-08, + "loss": 0.246, + "step": 327 + }, + { + "epoch": 0.9693675889328063, + "step": 327, + "train_accuracy_first_token_arguments": 0.9230769230769231, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.9693675889328063, + "step": 327, + "train_accuracy_first_token_arguments": 0.8125, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_deb": 1.0, + "train_accuracy_recipient_token_deb_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_plan": 1.0, + "train_accuracy_recipient_token_plan_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.9693675889328063, + "step": 327, + "train_accuracy_first_token_arguments": 0.7727272727272727, + "train_accuracy_first_token_arguments_total": 22, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.9723320158102767, + "grad_norm": 1.6782990768747863, + "learning_rate": 9.396958732856843e-09, + "loss": 0.2385, + "step": 328 + }, + { + "epoch": 0.9723320158102767, + "step": 328, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 29, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_find": 0.75, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.9723320158102767, + "step": 328, + "train_accuracy_first_token_arguments": 0.85, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.9723320158102767, + "step": 328, + "train_accuracy_first_token_arguments": 0.631578947368421, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_park": 1.0, + "train_accuracy_recipient_token_park_total": 4, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.9752964426877471, + "grad_norm": 2.45983910491564, + "learning_rate": 7.425734511117e-09, + "loss": 0.3068, + "step": 329 + }, + { + "epoch": 0.9752964426877471, + "step": 329, + "train_accuracy_first_token_arguments": 0.95, + "train_accuracy_first_token_arguments_total": 40, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_explain": 1.0, + "train_accuracy_recipient_token_explain_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.9752964426877471, + "step": 329, + "train_accuracy_first_token_arguments": 0.95, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.9752964426877471, + "step": 329, + "train_accuracy_first_token_arguments": 0.8666666666666667, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_art": 1.0, + "train_accuracy_recipient_token_art_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.9782608695652174, + "grad_norm": 1.9407711881422764, + "learning_rate": 5.685988060059045e-09, + "loss": 0.2393, + "step": 330 + }, + { + "epoch": 0.9782608695652174, + "step": 330, + "train_accuracy_first_token_arguments": 0.8571428571428571, + "train_accuracy_first_token_arguments_total": 28, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.8888888888888888, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.9782608695652174, + "step": 330, + "train_accuracy_first_token_arguments": 0.7142857142857143, + "train_accuracy_first_token_arguments_total": 21, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_filter": 1.0, + "train_accuracy_recipient_token_filter_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.9782608695652174, + "step": 330, + "train_accuracy_first_token_arguments": 0.9166666666666666, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_motor": 1.0, + "train_accuracy_recipient_token_motor_total": 3, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.9812252964426877, + "grad_norm": 1.9607276731011434, + "learning_rate": 4.17788094463023e-09, + "loss": 0.1726, + "step": 331 + }, + { + "epoch": 0.9812252964426877, + "step": 331, + "train_accuracy_first_token_arguments": 0.8666666666666667, + "train_accuracy_first_token_arguments_total": 15, + "train_accuracy_recipient_token": 0.6842105263157895, + "train_accuracy_recipient_token_all": 0.6666666666666666, + "train_accuracy_recipient_token_all_total": 15, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.9812252964426877, + "step": 331, + "train_accuracy_first_token_arguments": 0.8571428571428571, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_building": 1.0, + "train_accuracy_recipient_token_building_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.9812252964426877, + "step": 331, + "train_accuracy_first_token_arguments": 0.8, + "train_accuracy_first_token_arguments_total": 30, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_schedule": 0.5, + "train_accuracy_recipient_token_schedule_total": 2, + "train_accuracy_recipient_token_table": 0.5, + "train_accuracy_recipient_token_table_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.9841897233201581, + "grad_norm": 1.9808670794420047, + "learning_rate": 2.9015532181397854e-09, + "loss": 0.2188, + "step": 332 + }, + { + "epoch": 0.9841897233201581, + "step": 332, + "train_accuracy_first_token_arguments": 0.8421052631578947, + "train_accuracy_first_token_arguments_total": 19, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_generate": 0.5, + "train_accuracy_recipient_token_generate_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_sum": 1.0, + "train_accuracy_recipient_token_sum_total": 1, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 2, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.9841897233201581, + "step": 332, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 14, + "train_accuracy_recipient_token": 0.7857142857142857, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_evaluate": 0.0, + "train_accuracy_recipient_token_evaluate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_total_number_recipient_token": 14 + }, + { + "epoch": 0.9841897233201581, + "step": 332, + "train_accuracy_first_token_arguments": 0.96875, + "train_accuracy_first_token_arguments_total": 64, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_property": 1.0, + "train_accuracy_recipient_token_property_total": 4, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.9871541501976284, + "grad_norm": 2.0556535601550325, + "learning_rate": 1.857123409250705e-09, + "loss": 0.2487, + "step": 333 + }, + { + "epoch": 0.9871541501976284, + "step": 333, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_explain": 1.0, + "train_accuracy_recipient_token_explain_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.9871541501976284, + "step": 333, + "train_accuracy_first_token_arguments": 0.8571428571428571, + "train_accuracy_first_token_arguments_total": 21, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 2, + "train_accuracy_recipient_token_calculate": 0.5, + "train_accuracy_recipient_token_calculate_total": 2, + "train_accuracy_recipient_token_find": 0.6666666666666666, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_recommend": 1.0, + "train_accuracy_recipient_token_recommend_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.9871541501976284, + "step": 333, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 17, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_artifact": 1.0, + "train_accuracy_recipient_token_artifact_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.9901185770750988, + "grad_norm": 2.092295914240381, + "learning_rate": 1.0446885109746673e-09, + "loss": 0.1504, + "step": 334 + }, + { + "epoch": 0.9901185770750988, + "step": 334, + "train_accuracy_first_token_arguments": 0.8823529411764706, + "train_accuracy_first_token_arguments_total": 34, + "train_accuracy_recipient_token": 0.7222222222222222, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_schedule": 0.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_accuracy_recipient_token_track": 0.0, + "train_accuracy_recipient_token_track_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.9901185770750988, + "step": 334, + "train_accuracy_first_token_arguments": 0.9230769230769231, + "train_accuracy_first_token_arguments_total": 13, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analyze": 1.0, + "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_identify": 1.0, + "train_accuracy_recipient_token_identify_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.9901185770750988, + "step": 334, + "train_accuracy_first_token_arguments": 0.9166666666666666, + "train_accuracy_first_token_arguments_total": 12, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.9930830039525692, + "grad_norm": 2.0272702589635965, + "learning_rate": 4.6432397166285e-10, + "loss": 0.1775, + "step": 335 + }, + { + "epoch": 0.9930830039525692, + "step": 335, + "train_accuracy_first_token_arguments": 0.4375, + "train_accuracy_first_token_arguments_total": 32, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_write": 1.0, + "train_accuracy_recipient_token_write_total": 3, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.9930830039525692, + "step": 335, + "train_accuracy_first_token_arguments": 0.8769230769230769, + "train_accuracy_first_token_arguments_total": 65, + "train_accuracy_recipient_token": 0.7222222222222222, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.3333333333333333, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_optimize": 0.0, + "train_accuracy_recipient_token_optimize_total": 1, + "train_accuracy_recipient_token_plan": 0.0, + "train_accuracy_recipient_token_plan_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.9930830039525692, + "step": 335, + "train_accuracy_first_token_arguments": 0.85, + "train_accuracy_first_token_arguments_total": 20, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 15, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.9960474308300395, + "grad_norm": 1.9403721941845617, + "learning_rate": 1.160836880001459e-10, + "loss": 0.3228, + "step": 336 + }, + { + "epoch": 0.9960474308300395, + "step": 336, + "train_accuracy_first_token_arguments": 0.868421052631579, + "train_accuracy_first_token_arguments_total": 38, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.8333333333333334, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 2, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.9960474308300395, + "step": 336, + "train_accuracy_first_token_arguments": 0.9375, + "train_accuracy_first_token_arguments_total": 16, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_explain": 1.0, + "train_accuracy_recipient_token_explain_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.9960474308300395, + "step": 336, + "train_accuracy_first_token_arguments": 1.0, + "train_accuracy_first_token_arguments_total": 18, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.8, + "train_accuracy_recipient_token_find_total": 5, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.9990118577075099, + "grad_norm": 1.5288849501967092, + "learning_rate": 0.0, + "loss": 0.1852, + "step": 337 + }, + { + "epoch": 0.9990118577075099, + "step": 337, + "total_flos": 337785957384192.0, + "train_loss": 0.2736482885984002, + "train_runtime": 6297.3, + "train_samples_per_second": 0.321, + "train_steps_per_second": 0.054 + } + ], + "logging_steps": 1.0, + "max_steps": 337, + "num_input_tokens_seen": 0, + "num_train_epochs": 1, + "save_steps": 100.0, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 337785957384192.0, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}