{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.9990118577075099, "eval_steps": 106, "global_step": 337, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0, "step": 0, "train_accuracy_first_token_arguments": 0.7777777777777778, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 0.8, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_accuracy_recipient_token_set": 0.0, "train_accuracy_recipient_token_set_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0, "step": 0, "train_accuracy_first_token_arguments": 0.6129032258064516, "train_accuracy_first_token_arguments_total": 31, "train_accuracy_recipient_token": 0.6111111111111112, "train_accuracy_recipient_token_all": 0.6, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.3333333333333333, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_identify": 0.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0, "step": 0, "train_accuracy_first_token_arguments": 0.5555555555555556, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 0.6842105263157895, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_histor": 0.5, "train_accuracy_recipient_token_histor_total": 2, "train_accuracy_recipient_token_med": 1.0, "train_accuracy_recipient_token_med_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.0029644268774703555, "grad_norm": 3.2711954063798103, "learning_rate": 4.5454545454545457e-07, "loss": 0.3754, "step": 1 }, { "epoch": 0.0029644268774703555, "step": 1, "train_accuracy_first_token_arguments": 0.5245901639344263, "train_accuracy_first_token_arguments_total": 61, "train_accuracy_recipient_token": 0.6, "train_accuracy_recipient_token_all": 0.6, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 0.5, "train_accuracy_recipient_token_estimate_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_prior": 0.0, "train_accuracy_recipient_token_prior_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.0029644268774703555, "step": 1, "train_accuracy_first_token_arguments": 0.8076923076923077, "train_accuracy_first_token_arguments_total": 26, "train_accuracy_recipient_token": 0.6666666666666666, "train_accuracy_recipient_token_all": 0.6, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_search": 0.5, "train_accuracy_recipient_token_search_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.0029644268774703555, "step": 1, "train_accuracy_first_token_arguments": 0.8181818181818182, "train_accuracy_first_token_arguments_total": 22, "train_accuracy_recipient_token": 0.7894736842105263, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 0.75, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_search": 0.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.005928853754940711, "grad_norm": 2.9342358559874504, "learning_rate": 9.090909090909091e-07, "loss": 0.2983, "step": 2 }, { "epoch": 0.005928853754940711, "step": 2, "train_accuracy_first_token_arguments": 0.9166666666666666, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.7368421052631579, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_c": 0.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_get": 0.875, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.005928853754940711, "step": 2, "train_accuracy_first_token_arguments": 0.6666666666666666, "train_accuracy_first_token_arguments_total": 24, "train_accuracy_recipient_token": 0.8, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 2, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_interpret": 0.6666666666666666, "train_accuracy_recipient_token_interpret_total": 3, "train_total_number_recipient_token": 20 }, { "epoch": 0.005928853754940711, "step": 2, "train_accuracy_first_token_arguments": 0.8387096774193549, "train_accuracy_first_token_arguments_total": 31, "train_accuracy_recipient_token": 0.7368421052631579, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_check": 0.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 19 }, { "epoch": 0.008893280632411068, "grad_norm": 3.7666597300364524, "learning_rate": 1.3636363636363636e-06, "loss": 0.3456, "step": 3 }, { "epoch": 0.008893280632411068, "step": 3, "train_accuracy_first_token_arguments": 0.7241379310344828, "train_accuracy_first_token_arguments_total": 29, "train_accuracy_recipient_token": 0.6666666666666666, "train_accuracy_recipient_token_all": 0.6, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.008893280632411068, "step": 3, "train_accuracy_first_token_arguments": 0.7333333333333333, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.631578947368421, "train_accuracy_recipient_token_all": 0.6, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_list": 0.0, "train_accuracy_recipient_token_list_total": 1, "train_accuracy_recipient_token_vendor": 1.0, "train_accuracy_recipient_token_vendor_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.008893280632411068, "step": 3, "train_accuracy_first_token_arguments": 0.7058823529411765, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.55, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_count": 0.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_get": 0.5714285714285714, "train_accuracy_recipient_token_get_total": 7, "train_accuracy_recipient_token_identify": 0.0, "train_accuracy_recipient_token_identify_total": 2, "train_total_number_recipient_token": 20 }, { "epoch": 0.011857707509881422, "grad_norm": 4.455480512503962, "learning_rate": 1.8181818181818183e-06, "loss": 0.3511, "step": 4 }, { "epoch": 0.011857707509881422, "step": 4, "train_accuracy_first_token_arguments": 0.975609756097561, "train_accuracy_first_token_arguments_total": 41, "train_accuracy_recipient_token": 0.6, "train_accuracy_recipient_token_all": 0.5, "train_accuracy_recipient_token_all_total": 8, "train_accuracy_recipient_token_d": 0.0, "train_accuracy_recipient_token_d_total": 1, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 1, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_total_number_recipient_token": 15 }, { "epoch": 0.011857707509881422, "step": 4, "train_accuracy_first_token_arguments": 0.8125, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 0.6, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 0.5, "train_accuracy_recipient_token_analyze_total": 2, "train_accuracy_recipient_token_get": 0.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_locate": 0.0, "train_accuracy_recipient_token_locate_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 15 }, { "epoch": 0.011857707509881422, "step": 4, "train_accuracy_first_token_arguments": 0.75, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 0.5625, "train_accuracy_recipient_token_all": 0.6, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 0.0, "train_accuracy_recipient_token_analyze_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_rate": 1.0, "train_accuracy_recipient_token_rate_total": 1, "train_accuracy_recipient_token_recommend": 0.0, "train_accuracy_recipient_token_recommend_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.014822134387351778, "grad_norm": 4.251122308361701, "learning_rate": 2.2727272727272728e-06, "loss": 0.5231, "step": 5 }, { "epoch": 0.014822134387351778, "step": 5, "train_accuracy_first_token_arguments": 0.7391304347826086, "train_accuracy_first_token_arguments_total": 23, "train_accuracy_recipient_token": 0.6875, "train_accuracy_recipient_token_all": 0.6, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_fruit": 1.0, "train_accuracy_recipient_token_fruit_total": 3, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_select": 1.0, "train_accuracy_recipient_token_select_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.014822134387351778, "step": 5, "train_accuracy_first_token_arguments": 0.8888888888888888, "train_accuracy_first_token_arguments_total": 9, "train_accuracy_recipient_token": 0.6111111111111112, "train_accuracy_recipient_token_all": 0.6, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.7142857142857143, "train_accuracy_recipient_token_get_total": 7, "train_accuracy_recipient_token_identify": 0.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.014822134387351778, "step": 5, "train_accuracy_first_token_arguments": 0.6216216216216216, "train_accuracy_first_token_arguments_total": 37, "train_accuracy_recipient_token": 0.7894736842105263, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compose": 0.0, "train_accuracy_recipient_token_compose_total": 1, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_g": 1.0, "train_accuracy_recipient_token_g_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.017786561264822136, "grad_norm": 2.9047536525561006, "learning_rate": 2.7272727272727272e-06, "loss": 0.3063, "step": 6 }, { "epoch": 0.017786561264822136, "step": 6, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.7222222222222222, "train_accuracy_recipient_token_all": 0.6363636363636364, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.017786561264822136, "step": 6, "train_accuracy_first_token_arguments": 0.8648648648648649, "train_accuracy_first_token_arguments_total": 37, "train_accuracy_recipient_token": 0.5294117647058824, "train_accuracy_recipient_token_all": 0.6363636363636364, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 0.3333333333333333, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_identify": 0.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_validate": 0.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.017786561264822136, "step": 6, "train_accuracy_first_token_arguments": 0.76, "train_accuracy_first_token_arguments_total": 25, "train_accuracy_recipient_token": 0.75, "train_accuracy_recipient_token_all": 0.8333333333333334, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_evaluate": 0.0, "train_accuracy_recipient_token_evaluate_total": 1, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.020750988142292492, "grad_norm": 3.2977375008302783, "learning_rate": 3.181818181818182e-06, "loss": 0.4802, "step": 7 }, { "epoch": 0.020750988142292492, "step": 7, "train_accuracy_first_token_arguments": 0.9090909090909091, "train_accuracy_first_token_arguments_total": 11, "train_accuracy_recipient_token": 0.7777777777777778, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.5714285714285714, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { "epoch": 0.020750988142292492, "step": 7, "train_accuracy_first_token_arguments": 0.7692307692307693, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.75, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 16 }, { "epoch": 0.020750988142292492, "step": 7, "train_accuracy_first_token_arguments": 0.8421052631578947, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.7368421052631579, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.5555555555555556, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.023715415019762844, "grad_norm": 2.6597485903906652, "learning_rate": 3.6363636363636366e-06, "loss": 0.4234, "step": 8 }, { "epoch": 0.023715415019762844, "step": 8, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 0.6666666666666666, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 0.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_count": 0.5, "train_accuracy_recipient_token_count_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_identify": 0.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.023715415019762844, "step": 8, "train_accuracy_first_token_arguments": 0.9, "train_accuracy_first_token_arguments_total": 10, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.023715415019762844, "step": 8, "train_accuracy_first_token_arguments": 0.8867924528301887, "train_accuracy_first_token_arguments_total": 53, "train_accuracy_recipient_token": 0.7368421052631579, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_describe": 1.0, "train_accuracy_recipient_token_describe_total": 2, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_sum": 0.0, "train_accuracy_recipient_token_sum_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.0266798418972332, "grad_norm": 2.552647042706623, "learning_rate": 4.0909090909090915e-06, "loss": 0.3278, "step": 9 }, { "epoch": 0.0266798418972332, "step": 9, "train_accuracy_first_token_arguments": 0.7894736842105263, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.7777777777777778, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_prior": 1.0, "train_accuracy_recipient_token_prior_total": 1, "train_accuracy_recipient_token_schedule": 0.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.0266798418972332, "step": 9, "train_accuracy_first_token_arguments": 0.9333333333333333, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.5, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 0.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_load": 0.0, "train_accuracy_recipient_token_load_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.0266798418972332, "step": 9, "train_accuracy_first_token_arguments": 0.7142857142857143, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_property": 1.0, "train_accuracy_recipient_token_property_total": 3, "train_accuracy_recipient_token_real": 1.0, "train_accuracy_recipient_token_real_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.029644268774703556, "grad_norm": 2.3528382096065767, "learning_rate": 4.5454545454545455e-06, "loss": 0.2897, "step": 10 }, { "epoch": 0.029644268774703556, "step": 10, "train_accuracy_first_token_arguments": 0.76, "train_accuracy_first_token_arguments_total": 25, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_assemble": 1.0, "train_accuracy_recipient_token_assemble_total": 1, "train_accuracy_recipient_token_clean": 1.0, "train_accuracy_recipient_token_clean_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_repair": 0.5, "train_accuracy_recipient_token_repair_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.029644268774703556, "step": 10, "train_accuracy_first_token_arguments": 0.8260869565217391, "train_accuracy_first_token_arguments_total": 23, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 1, "train_accuracy_recipient_token_search": 0.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.029644268774703556, "step": 10, "train_accuracy_first_token_arguments": 0.6333333333333333, "train_accuracy_first_token_arguments_total": 30, "train_accuracy_recipient_token": 0.6666666666666666, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_suggest": 0.5, "train_accuracy_recipient_token_suggest_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.03260869565217391, "grad_norm": 2.351257384059913, "learning_rate": 5e-06, "loss": 0.2987, "step": 11 }, { "epoch": 0.03260869565217391, "step": 11, "train_accuracy_first_token_arguments": 0.5106382978723404, "train_accuracy_first_token_arguments_total": 47, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_add": 1.0, "train_accuracy_recipient_token_add_total": 5, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.03260869565217391, "step": 11, "train_accuracy_first_token_arguments": 0.9130434782608695, "train_accuracy_first_token_arguments_total": 23, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_h": 1.0, "train_accuracy_recipient_token_h_total": 3, "train_total_number_recipient_token": 20 }, { "epoch": 0.03260869565217391, "step": 11, "train_accuracy_first_token_arguments": 0.7142857142857143, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_fabric": 1.0, "train_accuracy_recipient_token_fabric_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.03557312252964427, "grad_norm": 2.266360756598947, "learning_rate": 4.999883916312e-06, "loss": 0.3119, "step": 12 }, { "epoch": 0.03557312252964427, "step": 12, "train_accuracy_first_token_arguments": 0.696969696969697, "train_accuracy_first_token_arguments_total": 33, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 5, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 10 }, { "epoch": 0.03557312252964427, "step": 12, "train_accuracy_first_token_arguments": 0.8048780487804879, "train_accuracy_first_token_arguments_total": 41, "train_accuracy_recipient_token": 0.7058823529411765, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 2, "train_accuracy_recipient_token_explain": 0.0, "train_accuracy_recipient_token_explain_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_translate": 0.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.03557312252964427, "step": 12, "train_accuracy_first_token_arguments": 0.8333333333333334, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.038537549407114624, "grad_norm": 10.867702658903724, "learning_rate": 4.999535676028338e-06, "loss": 0.4711, "step": 13 }, { "epoch": 0.038537549407114624, "step": 13, "train_accuracy_first_token_arguments": 0.9333333333333333, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.7058823529411765, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_suggest": 0.0, "train_accuracy_recipient_token_suggest_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.038537549407114624, "step": 13, "train_accuracy_first_token_arguments": 0.875, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_verify": 1.0, "train_accuracy_recipient_token_verify_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.038537549407114624, "step": 13, "train_accuracy_first_token_arguments": 0.9285714285714286, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_track": 1.0, "train_accuracy_recipient_token_track_total": 1, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.041501976284584984, "grad_norm": 2.8220096523893234, "learning_rate": 4.998955311489025e-06, "loss": 0.3082, "step": 14 }, { "epoch": 0.041501976284584984, "step": 14, "train_accuracy_first_token_arguments": 0.8636363636363636, "train_accuracy_first_token_arguments_total": 22, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 18 }, { "epoch": 0.041501976284584984, "step": 14, "train_accuracy_first_token_arguments": 0.7777777777777778, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_buy": 1.0, "train_accuracy_recipient_token_buy_total": 1, "train_accuracy_recipient_token_exp": 1.0, "train_accuracy_recipient_token_exp_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 17 }, { "epoch": 0.041501976284584984, "step": 14, "train_accuracy_first_token_arguments": 0.8285714285714286, "train_accuracy_first_token_arguments_total": 35, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_filter": 1.0, "train_accuracy_recipient_token_filter_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_locate": 1.0, "train_accuracy_recipient_token_locate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.044466403162055336, "grad_norm": 2.258107289871962, "learning_rate": 4.998142876590749e-06, "loss": 0.3097, "step": 15 }, { "epoch": 0.044466403162055336, "step": 15, "train_accuracy_first_token_arguments": 0.75, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_clean": 1.0, "train_accuracy_recipient_token_clean_total": 1, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 5, "train_accuracy_recipient_token_remove": 1.0, "train_accuracy_recipient_token_remove_total": 1, "train_accuracy_recipient_token_replace": 1.0, "train_accuracy_recipient_token_replace_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.044466403162055336, "step": 15, "train_accuracy_first_token_arguments": 0.75, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 0.7058823529411765, "train_accuracy_recipient_token_all": 0.6, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_identify": 0.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_register": 1.0, "train_accuracy_recipient_token_register_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.044466403162055336, "step": 15, "train_accuracy_first_token_arguments": 0.9545454545454546, "train_accuracy_first_token_arguments_total": 22, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_schedule": 0.0, "train_accuracy_recipient_token_schedule_total": 1, "train_accuracy_recipient_token_suggest": 1.0, "train_accuracy_recipient_token_suggest_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.04743083003952569, "grad_norm": 3.2749980475323635, "learning_rate": 4.997098446781861e-06, "loss": 0.2419, "step": 16 }, { "epoch": 0.04743083003952569, "step": 16, "train_accuracy_first_token_arguments": 0.85, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.04743083003952569, "step": 16, "train_accuracy_first_token_arguments": 0.7096774193548387, "train_accuracy_first_token_arguments_total": 62, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.04743083003952569, "step": 16, "train_accuracy_first_token_arguments": 0.9285714285714286, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_histor": 1.0, "train_accuracy_recipient_token_histor_total": 1, "train_accuracy_recipient_token_math": 1.0, "train_accuracy_recipient_token_math_total": 1, "train_accuracy_recipient_token_symbol": 1.0, "train_accuracy_recipient_token_symbol_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.05039525691699605, "grad_norm": 2.3223843516425937, "learning_rate": 4.9958221190553705e-06, "loss": 0.2414, "step": 17 }, { "epoch": 0.05039525691699605, "step": 17, "train_accuracy_first_token_arguments": 0.9583333333333334, "train_accuracy_first_token_arguments_total": 24, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 18 }, { "epoch": 0.05039525691699605, "step": 17, "train_accuracy_first_token_arguments": 0.92, "train_accuracy_first_token_arguments_total": 25, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.05039525691699605, "step": 17, "train_accuracy_first_token_arguments": 0.8571428571428571, "train_accuracy_first_token_arguments_total": 21, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_nutrition": 1.0, "train_accuracy_recipient_token_nutrition_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.0533596837944664, "grad_norm": 2.7061439489759453, "learning_rate": 4.994314011939941e-06, "loss": 0.2925, "step": 18 }, { "epoch": 0.0533596837944664, "step": 18, "train_accuracy_first_token_arguments": 0.9666666666666667, "train_accuracy_first_token_arguments_total": 30, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_building": 1.0, "train_accuracy_recipient_token_building_total": 1, "train_accuracy_recipient_token_find": 0.75, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_property": 1.0, "train_accuracy_recipient_token_property_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.0533596837944664, "step": 18, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.0533596837944664, "step": 18, "train_accuracy_first_token_arguments": 0.6428571428571429, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_decode": 0.0, "train_accuracy_recipient_token_decode_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { "epoch": 0.05632411067193676, "grad_norm": 2.8315352410694907, "learning_rate": 4.992574265488883e-06, "loss": 0.3402, "step": 19 }, { "epoch": 0.05632411067193676, "step": 19, "train_accuracy_first_token_arguments": 0.7916666666666666, "train_accuracy_first_token_arguments_total": 24, "train_accuracy_recipient_token": 0.7894736842105263, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_plan": 0.0, "train_accuracy_recipient_token_plan_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.05632411067193676, "step": 19, "train_accuracy_first_token_arguments": 0.8, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.05632411067193676, "step": 19, "train_accuracy_first_token_arguments": 0.8571428571428571, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.85, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_traffic": 1.0, "train_accuracy_recipient_token_traffic_total": 2, "train_accuracy_recipient_token_verify": 1.0, "train_accuracy_recipient_token_verify_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.05928853754940711, "grad_norm": 3.725703665798811, "learning_rate": 4.990603041267144e-06, "loss": 0.2884, "step": 20 }, { "epoch": 0.05928853754940711, "step": 20, "train_accuracy_first_token_arguments": 0.7241379310344828, "train_accuracy_first_token_arguments_total": 29, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.8, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_accuracy_recipient_token_suggest": 1.0, "train_accuracy_recipient_token_suggest_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.05928853754940711, "step": 20, "train_accuracy_first_token_arguments": 0.8, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8888888888888888, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.05928853754940711, "step": 20, "train_accuracy_first_token_arguments": 0.8, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.5625, "train_accuracy_recipient_token_all": 0.6363636363636364, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_deploy": 1.0, "train_accuracy_recipient_token_deploy_total": 1, "train_accuracy_recipient_token_identify": 0.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_install": 1.0, "train_accuracy_recipient_token_install_total": 1, "train_accuracy_recipient_token_schedule": 0.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.06225296442687747, "grad_norm": 2.4295313275927413, "learning_rate": 4.988400522336304e-06, "loss": 0.2966, "step": 21 }, { "epoch": 0.06225296442687747, "step": 21, "train_accuracy_first_token_arguments": 0.9473684210526315, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.7647058823529411, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_order": 0.0, "train_accuracy_recipient_token_order_total": 1, "train_accuracy_recipient_token_prior": 1.0, "train_accuracy_recipient_token_prior_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.06225296442687747, "step": 21, "train_accuracy_first_token_arguments": 0.8461538461538461, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.7647058823529411, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.7142857142857143, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.06225296442687747, "step": 21, "train_accuracy_first_token_arguments": 0.9, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_translate": 0.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.06521739130434782, "grad_norm": 3.0399009498719938, "learning_rate": 4.985966913237581e-06, "loss": 0.3759, "step": 22 }, { "epoch": 0.06521739130434782, "step": 22, "train_accuracy_first_token_arguments": 0.9230769230769231, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.7777777777777778, "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_estimate": 0.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_style": 1.0, "train_accuracy_recipient_token_style_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.06521739130434782, "step": 22, "train_accuracy_first_token_arguments": 0.5, "train_accuracy_first_token_arguments_total": 6, "train_accuracy_recipient_token": 0.7333333333333333, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_c": 0.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 15 }, { "epoch": 0.06521739130434782, "step": 22, "train_accuracy_first_token_arguments": 0.7804878048780488, "train_accuracy_first_token_arguments_total": 41, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_list": 0.0, "train_accuracy_recipient_token_list_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.06818181818181818, "grad_norm": 2.939285198394914, "learning_rate": 4.9833024399728295e-06, "loss": 0.4389, "step": 23 }, { "epoch": 0.06818181818181818, "step": 23, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 17 }, { "epoch": 0.06818181818181818, "step": 23, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.5, "train_accuracy_recipient_token_all": 0.6363636363636364, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 0.25, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_track": 0.0, "train_accuracy_recipient_token_track_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.06818181818181818, "step": 23, "train_accuracy_first_token_arguments": 0.9375, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 0.8125, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 16 }, { "epoch": 0.07114624505928854, "grad_norm": 3.2811864983396806, "learning_rate": 4.980407349983556e-06, "loss": 0.2645, "step": 24 }, { "epoch": 0.07114624505928854, "step": 24, "train_accuracy_first_token_arguments": 0.9166666666666666, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.07114624505928854, "step": 24, "train_accuracy_first_token_arguments": 0.9016393442622951, "train_accuracy_first_token_arguments_total": 61, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.07114624505928854, "step": 24, "train_accuracy_first_token_arguments": 0.76, "train_accuracy_first_token_arguments_total": 25, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 1, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_accuracy_recipient_token_rate": 1.0, "train_accuracy_recipient_token_rate_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.0741106719367589, "grad_norm": 1.8226124894214228, "learning_rate": 4.9772819121279395e-06, "loss": 0.2212, "step": 25 }, { "epoch": 0.0741106719367589, "step": 25, "train_accuracy_first_token_arguments": 0.9193548387096774, "train_accuracy_first_token_arguments_total": 62, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { "epoch": 0.0741106719367589, "step": 25, "train_accuracy_first_token_arguments": 0.76, "train_accuracy_first_token_arguments_total": 25, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.0741106719367589, "step": 25, "train_accuracy_first_token_arguments": 0.6666666666666666, "train_accuracy_first_token_arguments_total": 24, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_plan": 0.0, "train_accuracy_recipient_token_plan_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.07707509881422925, "grad_norm": 1.524883365948892, "learning_rate": 4.973926416655863e-06, "loss": 0.1674, "step": 26 }, { "epoch": 0.07707509881422925, "step": 26, "train_accuracy_first_token_arguments": 0.8333333333333334, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.7777777777777778, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_animal": 0.6666666666666666, "train_accuracy_recipient_token_animal_total": 3, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_im": 1.0, "train_accuracy_recipient_token_im_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.07707509881422925, "step": 26, "train_accuracy_first_token_arguments": 0.8260869565217391, "train_accuracy_first_token_arguments_total": 23, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_plan": 0.0, "train_accuracy_recipient_token_plan_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.07707509881422925, "step": 26, "train_accuracy_first_token_arguments": 0.75, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 0.9047619047619048, "train_accuracy_recipient_token_all": 0.9166666666666666, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_get": 0.875, "train_accuracy_recipient_token_get_total": 8, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 21 }, { "epoch": 0.0800395256916996, "grad_norm": 2.0177323408527643, "learning_rate": 4.970341175181957e-06, "loss": 0.2729, "step": 27 }, { "epoch": 0.0800395256916996, "step": 27, "train_accuracy_first_token_arguments": 0.8888888888888888, "train_accuracy_first_token_arguments_total": 9, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.0800395256916996, "step": 27, "train_accuracy_first_token_arguments": 0.8787878787878788, "train_accuracy_first_token_arguments_total": 33, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_draw": 1.0, "train_accuracy_recipient_token_draw_total": 5, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.0800395256916996, "step": 27, "train_accuracy_first_token_arguments": 0.7692307692307693, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.85, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 0.875, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 20 }, { "epoch": 0.08300395256916997, "grad_norm": 2.0988150229289273, "learning_rate": 4.966526520656663e-06, "loss": 0.3187, "step": 28 }, { "epoch": 0.08300395256916997, "step": 28, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.7692307692307693, "train_accuracy_recipient_token_all_total": 13, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.08300395256916997, "step": 28, "train_accuracy_first_token_arguments": 0.7857142857142857, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.08300395256916997, "step": 28, "train_accuracy_first_token_arguments": 0.7241379310344828, "train_accuracy_first_token_arguments_total": 29, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_rec": 1.0, "train_accuracy_recipient_token_rec_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.08596837944664032, "grad_norm": 2.2452954917456447, "learning_rate": 4.9624828073353144e-06, "loss": 0.2444, "step": 29 }, { "epoch": 0.08596837944664032, "step": 29, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 24, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 18 }, { "epoch": 0.08596837944664032, "step": 29, "train_accuracy_first_token_arguments": 0.9, "train_accuracy_first_token_arguments_total": 10, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.08596837944664032, "step": 29, "train_accuracy_first_token_arguments": 0.75, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.8, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_im": 1.0, "train_accuracy_recipient_token_im_total": 1, "train_accuracy_recipient_token_install": 0.0, "train_accuracy_recipient_token_install_total": 1, "train_total_number_recipient_token": 15 }, { "epoch": 0.08893280632411067, "grad_norm": 2.1813680120346604, "learning_rate": 4.958210410745237e-06, "loss": 0.2787, "step": 30 }, { "epoch": 0.08893280632411067, "step": 30, "train_accuracy_first_token_arguments": 0.9259259259259259, "train_accuracy_first_token_arguments_total": 54, "train_accuracy_recipient_token": 0.8125, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_event": 1.0, "train_accuracy_recipient_token_event_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 16 }, { "epoch": 0.08893280632411067, "step": 30, "train_accuracy_first_token_arguments": 0.8235294117647058, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.7777777777777778, "train_accuracy_recipient_token_all": 0.6666666666666666, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.08893280632411067, "step": 30, "train_accuracy_first_token_arguments": 0.88, "train_accuracy_first_token_arguments_total": 25, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_hire": 1.0, "train_accuracy_recipient_token_hire_total": 1, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.09189723320158102, "grad_norm": 1.9535807669288685, "learning_rate": 4.95370972765087e-06, "loss": 0.384, "step": 31 }, { "epoch": 0.09189723320158102, "step": 31, "train_accuracy_first_token_arguments": 0.6666666666666666, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_convert": 1.0, "train_accuracy_recipient_token_convert_total": 5, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.09189723320158102, "step": 31, "train_accuracy_first_token_arguments": 0.75, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_evaluate": 1.0, "train_accuracy_recipient_token_evaluate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_stat": 1.0, "train_accuracy_recipient_token_stat_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.09189723320158102, "step": 31, "train_accuracy_first_token_arguments": 0.7368421052631579, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_filter": 0.8, "train_accuracy_recipient_token_filter_total": 5, "train_accuracy_recipient_token_food": 1.0, "train_accuracy_recipient_token_food_total": 4, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.09486166007905138, "grad_norm": 2.5184069445357213, "learning_rate": 4.94898117601693e-06, "loss": 0.255, "step": 32 }, { "epoch": 0.09486166007905138, "step": 32, "train_accuracy_first_token_arguments": 0.7272727272727273, "train_accuracy_first_token_arguments_total": 22, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.09486166007905138, "step": 32, "train_accuracy_first_token_arguments": 0.8461538461538461, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_schedule": 0.0, "train_accuracy_recipient_token_schedule_total": 1, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.09486166007905138, "step": 32, "train_accuracy_first_token_arguments": 0.9444444444444444, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_suggest": 1.0, "train_accuracy_recipient_token_suggest_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.09782608695652174, "grad_norm": 1.7872388531938566, "learning_rate": 4.944025194969586e-06, "loss": 0.241, "step": 33 }, { "epoch": 0.09782608695652174, "step": 33, "train_accuracy_first_token_arguments": 0.6428571428571429, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { "epoch": 0.09782608695652174, "step": 33, "train_accuracy_first_token_arguments": 0.9528301886792453, "train_accuracy_first_token_arguments_total": 106, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_assert": 1.0, "train_accuracy_recipient_token_assert_total": 5, "train_accuracy_recipient_token_describe": 1.0, "train_accuracy_recipient_token_describe_total": 1, "train_accuracy_recipient_token_mount": 1.0, "train_accuracy_recipient_token_mount_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.09782608695652174, "step": 33, "train_accuracy_first_token_arguments": 0.9811320754716981, "train_accuracy_first_token_arguments_total": 53, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9166666666666666, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { "epoch": 0.1007905138339921, "grad_norm": 1.826928371292299, "learning_rate": 4.938842244755683e-06, "loss": 0.3343, "step": 34 }, { "epoch": 0.1007905138339921, "step": 34, "train_accuracy_first_token_arguments": 0.75, "train_accuracy_first_token_arguments_total": 8, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.1007905138339921, "step": 34, "train_accuracy_first_token_arguments": 0.8888888888888888, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8888888888888888, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.1007905138339921, "step": 34, "train_accuracy_first_token_arguments": 0.9333333333333333, "train_accuracy_first_token_arguments_total": 30, "train_accuracy_recipient_token": 0.7894736842105263, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_compare": 0.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.10375494071146245, "grad_norm": 2.586283846482814, "learning_rate": 4.933432806700004e-06, "loss": 0.2806, "step": 35 }, { "epoch": 0.10375494071146245, "step": 35, "train_accuracy_first_token_arguments": 0.6956521739130435, "train_accuracy_first_token_arguments_total": 23, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_validate": 0.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.10375494071146245, "step": 35, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 22, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 20 }, { "epoch": 0.10375494071146245, "step": 35, "train_accuracy_first_token_arguments": 0.6666666666666666, "train_accuracy_first_token_arguments_total": 33, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_art": 0.6666666666666666, "train_accuracy_recipient_token_art_total": 3, "train_accuracy_recipient_token_inspect": 1.0, "train_accuracy_recipient_token_inspect_total": 1, "train_accuracy_recipient_token_repair": 1.0, "train_accuracy_recipient_token_repair_total": 1, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.1067193675889328, "grad_norm": 2.28772413603275, "learning_rate": 4.927797383160561e-06, "loss": 0.3219, "step": 36 }, { "epoch": 0.1067193675889328, "step": 36, "train_accuracy_first_token_arguments": 0.9090909090909091, "train_accuracy_first_token_arguments_total": 11, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 20 }, { "epoch": 0.1067193675889328, "step": 36, "train_accuracy_first_token_arguments": 0.8, "train_accuracy_first_token_arguments_total": 25, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.1067193675889328, "step": 36, "train_accuracy_first_token_arguments": 0.6333333333333333, "train_accuracy_first_token_arguments_total": 30, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 0.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.10968379446640317, "grad_norm": 1.7442627302155183, "learning_rate": 4.921936497481956e-06, "loss": 0.3006, "step": 37 }, { "epoch": 0.10968379446640317, "step": 37, "train_accuracy_first_token_arguments": 0.8, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.10968379446640317, "step": 37, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_track": 1.0, "train_accuracy_recipient_token_track_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.10968379446640317, "step": 37, "train_accuracy_first_token_arguments": 0.9166666666666666, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 21 }, { "epoch": 0.11264822134387352, "grad_norm": 2.0180662009823, "learning_rate": 4.915850693946766e-06, "loss": 0.266, "step": 38 }, { "epoch": 0.11264822134387352, "step": 38, "train_accuracy_first_token_arguments": 0.8387096774193549, "train_accuracy_first_token_arguments_total": 31, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_log": 1.0, "train_accuracy_recipient_token_log_total": 1, "train_accuracy_recipient_token_plant": 1.0, "train_accuracy_recipient_token_plant_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.11264822134387352, "step": 38, "train_accuracy_first_token_arguments": 0.6274509803921569, "train_accuracy_first_token_arguments_total": 51, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_plan": 1.0, "train_accuracy_recipient_token_plan_total": 1, "train_accuracy_recipient_token_urban": 1.0, "train_accuracy_recipient_token_urban_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.11264822134387352, "step": 38, "train_accuracy_first_token_arguments": 0.6862745098039216, "train_accuracy_first_token_arguments_total": 51, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_upgrade": 1.0, "train_accuracy_recipient_token_upgrade_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.11561264822134387, "grad_norm": 2.0013783198828907, "learning_rate": 4.909540537725007e-06, "loss": 0.3129, "step": 39 }, { "epoch": 0.11561264822134387, "step": 39, "train_accuracy_first_token_arguments": 0.8888888888888888, "train_accuracy_first_token_arguments_total": 9, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 17 }, { "epoch": 0.11561264822134387, "step": 39, "train_accuracy_first_token_arguments": 0.9375, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 2, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.11561264822134387, "step": 39, "train_accuracy_first_token_arguments": 0.7560975609756098, "train_accuracy_first_token_arguments_total": 41, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_track": 1.0, "train_accuracy_recipient_token_track_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.11857707509881422, "grad_norm": 2.383877337768723, "learning_rate": 4.903006614821645e-06, "loss": 0.2901, "step": 40 }, { "epoch": 0.11857707509881422, "step": 40, "train_accuracy_first_token_arguments": 0.7666666666666667, "train_accuracy_first_token_arguments_total": 30, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_accuracy_recipient_token_translate": 0.5, "train_accuracy_recipient_token_translate_total": 2, "train_total_number_recipient_token": 20 }, { "epoch": 0.11857707509881422, "step": 40, "train_accuracy_first_token_arguments": 0.8571428571428571, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 20 }, { "epoch": 0.11857707509881422, "step": 40, "train_accuracy_first_token_arguments": 0.9183673469387755, "train_accuracy_first_token_arguments_total": 49, "train_accuracy_recipient_token": 0.7, "train_accuracy_recipient_token_all": 0.6666666666666666, "train_accuracy_recipient_token_all_total": 15, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_setup": 0.0, "train_accuracy_recipient_token_setup_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.12154150197628459, "grad_norm": 2.3845477671070725, "learning_rate": 4.8962495320221714e-06, "loss": 0.2275, "step": 41 }, { "epoch": 0.12154150197628459, "step": 41, "train_accuracy_first_token_arguments": 0.65, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.12154150197628459, "step": 41, "train_accuracy_first_token_arguments": 0.8888888888888888, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.12154150197628459, "step": 41, "train_accuracy_first_token_arguments": 0.5769230769230769, "train_accuracy_first_token_arguments_total": 52, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 1, "train_accuracy_recipient_token_merge": 1.0, "train_accuracy_recipient_token_merge_total": 1, "train_accuracy_recipient_token_recommend": 0.75, "train_accuracy_recipient_token_recommend_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.12450592885375494, "grad_norm": 2.2451847636665194, "learning_rate": 4.8892699168362626e-06, "loss": 0.2985, "step": 42 }, { "epoch": 0.12450592885375494, "step": 42, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compile": 1.0, "train_accuracy_recipient_token_compile_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.12450592885375494, "step": 42, "train_accuracy_first_token_arguments": 0.8, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 2, "train_total_number_recipient_token": 16 }, { "epoch": 0.12450592885375494, "step": 42, "train_accuracy_first_token_arguments": 0.8823529411764706, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.1274703557312253, "grad_norm": 2.1597197573051443, "learning_rate": 4.8820684174394935e-06, "loss": 0.3257, "step": 43 }, { "epoch": 0.1274703557312253, "step": 43, "train_accuracy_first_token_arguments": 0.6428571428571429, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 16 }, { "epoch": 0.1274703557312253, "step": 43, "train_accuracy_first_token_arguments": 0.75, "train_accuracy_first_token_arguments_total": 8, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.8461538461538461, "train_accuracy_recipient_token_all_total": 13, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.1274703557312253, "step": 43, "train_accuracy_first_token_arguments": 0.8333333333333334, "train_accuracy_first_token_arguments_total": 6, "train_accuracy_recipient_token": 0.8666666666666667, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 15 }, { "epoch": 0.13043478260869565, "grad_norm": 2.4009249785875033, "learning_rate": 4.874645702613152e-06, "loss": 0.4131, "step": 44 }, { "epoch": 0.13043478260869565, "step": 44, "train_accuracy_first_token_arguments": 0.7727272727272727, "train_accuracy_first_token_arguments_total": 22, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.13043478260869565, "step": 44, "train_accuracy_first_token_arguments": 0.8095238095238095, "train_accuracy_first_token_arguments_total": 21, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_artifact": 1.0, "train_accuracy_recipient_token_artifact_total": 3, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 20 }, { "epoch": 0.13043478260869565, "step": 44, "train_accuracy_first_token_arguments": 0.6, "train_accuracy_first_token_arguments_total": 35, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_build": 1.0, "train_accuracy_recipient_token_build_total": 1, "train_accuracy_recipient_token_calculate": 0.75, "train_accuracy_recipient_token_calculate_total": 4, "train_accuracy_recipient_token_fire": 1.0, "train_accuracy_recipient_token_fire_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.13339920948616601, "grad_norm": 2.2178791639658186, "learning_rate": 4.867002461682129e-06, "loss": 0.401, "step": 45 }, { "epoch": 0.13339920948616601, "step": 45, "train_accuracy_first_token_arguments": 0.8235294117647058, "train_accuracy_first_token_arguments_total": 34, "train_accuracy_recipient_token": 0.7368421052631579, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 0.5, "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_find": 0.4, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.13339920948616601, "step": 45, "train_accuracy_first_token_arguments": 0.9333333333333333, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_ex": 0.3333333333333333, "train_accuracy_recipient_token_ex_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.13339920948616601, "step": 45, "train_accuracy_first_token_arguments": 0.8709677419354839, "train_accuracy_first_token_arguments_total": 31, "train_accuracy_recipient_token": 0.85, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_log": 1.0, "train_accuracy_recipient_token_log_total": 1, "train_accuracy_recipient_token_review": 0.5, "train_accuracy_recipient_token_review_total": 2, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.13636363636363635, "grad_norm": 1.7428907811945011, "learning_rate": 4.8591394044508985e-06, "loss": 0.2065, "step": 46 }, { "epoch": 0.13636363636363635, "step": 46, "train_accuracy_first_token_arguments": 0.9, "train_accuracy_first_token_arguments_total": 200, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_filter": 0.75, "train_accuracy_recipient_token_filter_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.13636363636363635, "step": 46, "train_accuracy_first_token_arguments": 0.8235294117647058, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.85, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.13636363636363635, "step": 46, "train_accuracy_first_token_arguments": 0.9259259259259259, "train_accuracy_first_token_arguments_total": 27, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_city": 0.75, "train_accuracy_recipient_token_city_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.13932806324110672, "grad_norm": 2.3699963492078684, "learning_rate": 4.851057261137608e-06, "loss": 0.2718, "step": 47 }, { "epoch": 0.13932806324110672, "step": 47, "train_accuracy_first_token_arguments": 0.8, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { "epoch": 0.13932806324110672, "step": 47, "train_accuracy_first_token_arguments": 0.59375, "train_accuracy_first_token_arguments_total": 32, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 2, "train_accuracy_recipient_token_detect": 1.0, "train_accuracy_recipient_token_detect_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.13932806324110672, "step": 47, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 22, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.8333333333333334, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_calculate": 0.5, "train_accuracy_recipient_token_calculate_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.1422924901185771, "grad_norm": 2.553448245564224, "learning_rate": 4.842756782306261e-06, "loss": 0.3375, "step": 48 }, { "epoch": 0.1422924901185771, "step": 48, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.6470588235294118, "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_organ": 0.0, "train_accuracy_recipient_token_organ_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.1422924901185771, "step": 48, "train_accuracy_first_token_arguments": 0.5384615384615384, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 2, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_d": 1.0, "train_accuracy_recipient_token_d_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.1422924901185771, "step": 48, "train_accuracy_first_token_arguments": 0.7894736842105263, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_c": 0.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_total_number_recipient_token": 20 }, { "epoch": 0.14525691699604742, "grad_norm": 2.047932501451832, "learning_rate": 4.8342387387970105e-06, "loss": 0.2978, "step": 49 }, { "epoch": 0.14525691699604742, "step": 49, "train_accuracy_first_token_arguments": 0.9090909090909091, "train_accuracy_first_token_arguments_total": 33, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.875, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 20 }, { "epoch": 0.14525691699604742, "step": 49, "train_accuracy_first_token_arguments": 0.85, "train_accuracy_first_token_arguments_total": 40, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 8, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_setup": 1.0, "train_accuracy_recipient_token_setup_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.14525691699604742, "step": 49, "train_accuracy_first_token_arguments": 0.8181818181818182, "train_accuracy_first_token_arguments_total": 55, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_g": 0.6666666666666666, "train_accuracy_recipient_token_g_total": 3, "train_accuracy_recipient_token_track": 1.0, "train_accuracy_recipient_token_track_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.1482213438735178, "grad_norm": 2.251584762112145, "learning_rate": 4.825503921654582e-06, "loss": 0.1679, "step": 50 }, { "epoch": 0.1482213438735178, "step": 50, "train_accuracy_first_token_arguments": 0.9090909090909091, "train_accuracy_first_token_arguments_total": 11, "train_accuracy_recipient_token": 0.8125, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_report": 1.0, "train_accuracy_recipient_token_report_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.1482213438735178, "step": 50, "train_accuracy_first_token_arguments": 0.6923076923076923, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 16 }, { "epoch": 0.1482213438735178, "step": 50, "train_accuracy_first_token_arguments": 0.8, "train_accuracy_first_token_arguments_total": 10, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 0.5, "train_accuracy_recipient_token_analyze_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_predict": 1.0, "train_accuracy_recipient_token_predict_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.15118577075098813, "grad_norm": 2.326983001481526, "learning_rate": 4.816553142054806e-06, "loss": 0.2279, "step": 51 }, { "epoch": 0.15118577075098813, "step": 51, "train_accuracy_first_token_arguments": 0.7692307692307693, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_plan": 1.0, "train_accuracy_recipient_token_plan_total": 1, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.15118577075098813, "step": 51, "train_accuracy_first_token_arguments": 0.875, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_max": 1.0, "train_accuracy_recipient_token_max_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.15118577075098813, "step": 51, "train_accuracy_first_token_arguments": 0.8421052631578947, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.1541501976284585, "grad_norm": 2.104029483073964, "learning_rate": 4.807387231229287e-06, "loss": 0.3356, "step": 52 }, { "epoch": 0.1541501976284585, "step": 52, "train_accuracy_first_token_arguments": 0.7, "train_accuracy_first_token_arguments_total": 10, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.1541501976284585, "step": 52, "train_accuracy_first_token_arguments": 0.8620689655172413, "train_accuracy_first_token_arguments_total": 29, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_evaluate": 1.0, "train_accuracy_recipient_token_evaluate_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_im": 1.0, "train_accuracy_recipient_token_im_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.1541501976284585, "step": 52, "train_accuracy_first_token_arguments": 0.782608695652174, "train_accuracy_first_token_arguments_total": 23, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_display": 1.0, "train_accuracy_recipient_token_display_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_h": 1.0, "train_accuracy_recipient_token_h_total": 3, "train_accuracy_recipient_token_suggest": 1.0, "train_accuracy_recipient_token_suggest_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.15711462450592886, "grad_norm": 2.1920664784328907, "learning_rate": 4.798007040388212e-06, "loss": 0.2247, "step": 53 }, { "epoch": 0.15711462450592886, "step": 53, "train_accuracy_first_token_arguments": 0.9428571428571428, "train_accuracy_first_token_arguments_total": 35, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_calculate": 0.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.15711462450592886, "step": 53, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 11, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.15711462450592886, "step": 53, "train_accuracy_first_token_arguments": 0.6857142857142857, "train_accuracy_first_token_arguments_total": 35, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.1600790513833992, "grad_norm": 1.6828766971614149, "learning_rate": 4.788413440641297e-06, "loss": 0.2468, "step": 54 }, { "epoch": 0.1600790513833992, "step": 54, "train_accuracy_first_token_arguments": 0.9130434782608695, "train_accuracy_first_token_arguments_total": 23, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_flag": 0.6666666666666666, "train_accuracy_recipient_token_flag_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 18 }, { "epoch": 0.1600790513833992, "step": 54, "train_accuracy_first_token_arguments": 0.8461538461538461, "train_accuracy_first_token_arguments_total": 26, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { "epoch": 0.1600790513833992, "step": 54, "train_accuracy_first_token_arguments": 0.5476190476190477, "train_accuracy_first_token_arguments_total": 42, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_log": 1.0, "train_accuracy_recipient_token_log_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.16304347826086957, "grad_norm": 2.3713370104618146, "learning_rate": 4.778607322916896e-06, "loss": 0.4323, "step": 55 }, { "epoch": 0.16304347826086957, "step": 55, "train_accuracy_first_token_arguments": 0.8627450980392157, "train_accuracy_first_token_arguments_total": 51, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 1, "train_accuracy_recipient_token_draw": 1.0, "train_accuracy_recipient_token_draw_total": 1, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.16304347826086957, "step": 55, "train_accuracy_first_token_arguments": 0.9473684210526315, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_build": 0.5, "train_accuracy_recipient_token_build_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_paint": 1.0, "train_accuracy_recipient_token_paint_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.16304347826086957, "step": 55, "train_accuracy_first_token_arguments": 0.7368421052631579, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_fetch": 1.0, "train_accuracy_recipient_token_fetch_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { "epoch": 0.16600790513833993, "grad_norm": 1.892415917915294, "learning_rate": 4.7685895978792564e-06, "loss": 0.2455, "step": 56 }, { "epoch": 0.16600790513833993, "step": 56, "train_accuracy_first_token_arguments": 0.68, "train_accuracy_first_token_arguments_total": 25, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_monitor": 0.0, "train_accuracy_recipient_token_monitor_total": 1, "train_accuracy_recipient_token_remove": 1.0, "train_accuracy_recipient_token_remove_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.16600790513833993, "step": 56, "train_accuracy_first_token_arguments": 0.9761904761904762, "train_accuracy_first_token_arguments_total": 42, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 1, "train_accuracy_recipient_token_explain": 1.0, "train_accuracy_recipient_token_explain_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_number": 1.0, "train_accuracy_recipient_token_number_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.16600790513833993, "step": 56, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 20 }, { "epoch": 0.16897233201581027, "grad_norm": 1.7624413054210974, "learning_rate": 4.7583611958439514e-06, "loss": 0.2711, "step": 57 }, { "epoch": 0.16897233201581027, "step": 57, "train_accuracy_first_token_arguments": 0.8205128205128205, "train_accuracy_first_token_arguments_total": 39, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_hire": 1.0, "train_accuracy_recipient_token_hire_total": 1, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.16897233201581027, "step": 57, "train_accuracy_first_token_arguments": 0.8717948717948718, "train_accuracy_first_token_arguments_total": 39, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.16897233201581027, "step": 57, "train_accuracy_first_token_arguments": 0.7735849056603774, "train_accuracy_first_token_arguments_total": 106, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_ex": 1.0, "train_accuracy_recipient_token_ex_total": 3, "train_accuracy_recipient_token_format": 0.0, "train_accuracy_recipient_token_format_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.17193675889328064, "grad_norm": 2.096786172099922, "learning_rate": 4.747923066691487e-06, "loss": 0.2883, "step": 58 }, { "epoch": 0.17193675889328064, "step": 58, "train_accuracy_first_token_arguments": 0.9090909090909091, "train_accuracy_first_token_arguments_total": 11, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 20 }, { "epoch": 0.17193675889328064, "step": 58, "train_accuracy_first_token_arguments": 0.58, "train_accuracy_first_token_arguments_total": 50, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_f": 1.0, "train_accuracy_recipient_token_f_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 1, "train_accuracy_recipient_token_wall": 1.0, "train_accuracy_recipient_token_wall_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.17193675889328064, "step": 58, "train_accuracy_first_token_arguments": 0.5, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_arch": 1.0, "train_accuracy_recipient_token_arch_total": 1, "train_accuracy_recipient_token_building": 1.0, "train_accuracy_recipient_token_building_total": 1, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 18 }, { "epoch": 0.17490118577075098, "grad_norm": 1.6122336767210732, "learning_rate": 4.7372761797790836e-06, "loss": 0.2393, "step": 59 }, { "epoch": 0.17490118577075098, "step": 59, "train_accuracy_first_token_arguments": 0.6153846153846154, "train_accuracy_first_token_arguments_total": 26, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_detect": 1.0, "train_accuracy_recipient_token_detect_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_truck": 1.0, "train_accuracy_recipient_token_truck_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.17490118577075098, "step": 59, "train_accuracy_first_token_arguments": 0.6428571428571429, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.17490118577075098, "step": 59, "train_accuracy_first_token_arguments": 0.75, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 20 }, { "epoch": 0.17786561264822134, "grad_norm": 1.7027092503573524, "learning_rate": 4.726421523850662e-06, "loss": 0.1559, "step": 60 }, { "epoch": 0.17786561264822134, "step": 60, "train_accuracy_first_token_arguments": 0.7647058823529411, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_door": 1.0, "train_accuracy_recipient_token_door_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_invest": 1.0, "train_accuracy_recipient_token_invest_total": 1, "train_accuracy_recipient_token_trans": 1.0, "train_accuracy_recipient_token_trans_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.17786561264822134, "step": 60, "train_accuracy_first_token_arguments": 0.8947368421052632, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.8125, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_generate": 0.0, "train_accuracy_recipient_token_generate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 16 }, { "epoch": 0.17786561264822134, "step": 60, "train_accuracy_first_token_arguments": 0.7692307692307693, "train_accuracy_first_token_arguments_total": 26, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 2, "train_total_number_recipient_token": 20 }, { "epoch": 0.1808300395256917, "grad_norm": 1.640939742247205, "learning_rate": 4.715360106945015e-06, "loss": 0.2414, "step": 61 }, { "epoch": 0.1808300395256917, "step": 61, "train_accuracy_first_token_arguments": 0.875, "train_accuracy_first_token_arguments_total": 24, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_animal": 1.0, "train_accuracy_recipient_token_animal_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_store": 1.0, "train_accuracy_recipient_token_store_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.1808300395256917, "step": 61, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_plant": 0.75, "train_accuracy_recipient_token_plant_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.1808300395256917, "step": 61, "train_accuracy_first_token_arguments": 0.8947368421052632, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 17 }, { "epoch": 0.18379446640316205, "grad_norm": 2.3050792009276817, "learning_rate": 4.7040929563021975e-06, "loss": 0.2552, "step": 62 }, { "epoch": 0.18379446640316205, "step": 62, "train_accuracy_first_token_arguments": 0.9411764705882353, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.18379446640316205, "step": 62, "train_accuracy_first_token_arguments": 0.8, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_par": 1.0, "train_accuracy_recipient_token_par_total": 2, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.18379446640316205, "step": 62, "train_accuracy_first_token_arguments": 0.8695652173913043, "train_accuracy_first_token_arguments_total": 23, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_flower": 1.0, "train_accuracy_recipient_token_flower_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.18675889328063242, "grad_norm": 2.559937679792714, "learning_rate": 4.6926211182681295e-06, "loss": 0.4525, "step": 63 }, { "epoch": 0.18675889328063242, "step": 63, "train_accuracy_first_token_arguments": 0.9565217391304348, "train_accuracy_first_token_arguments_total": 46, "train_accuracy_recipient_token": 0.7058823529411765, "train_accuracy_recipient_token_all": 0.6666666666666666, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_compare": 0.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 2, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.18675889328063242, "step": 63, "train_accuracy_first_token_arguments": 0.6190476190476191, "train_accuracy_first_token_arguments_total": 21, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_emergency": 1.0, "train_accuracy_recipient_token_emergency_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_first": 0.0, "train_accuracy_recipient_token_first_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 19 }, { "epoch": 0.18675889328063242, "step": 63, "train_accuracy_first_token_arguments": 0.6129032258064516, "train_accuracy_first_token_arguments_total": 31, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_explain": 1.0, "train_accuracy_recipient_token_explain_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.18972332015810275, "grad_norm": 2.2063521775652393, "learning_rate": 4.680945658197425e-06, "loss": 0.2713, "step": 64 }, { "epoch": 0.18972332015810275, "step": 64, "train_accuracy_first_token_arguments": 0.9230769230769231, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { "epoch": 0.18972332015810275, "step": 64, "train_accuracy_first_token_arguments": 0.75, "train_accuracy_first_token_arguments_total": 8, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.18972332015810275, "step": 64, "train_accuracy_first_token_arguments": 0.95, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { "epoch": 0.19268774703557312, "grad_norm": 2.1111555785819514, "learning_rate": 4.669067660354456e-06, "loss": 0.2806, "step": 65 }, { "epoch": 0.19268774703557312, "step": 65, "train_accuracy_first_token_arguments": 0.7272727272727273, "train_accuracy_first_token_arguments_total": 11, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_gr": 0.8, "train_accuracy_recipient_token_gr_total": 5, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.19268774703557312, "step": 65, "train_accuracy_first_token_arguments": 0.7692307692307693, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.8666666666666667, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_license": 1.0, "train_accuracy_recipient_token_license_total": 1, "train_total_number_recipient_token": 15 }, { "epoch": 0.19268774703557312, "step": 65, "train_accuracy_first_token_arguments": 0.9230769230769231, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.7777777777777778, "train_accuracy_recipient_token_all": 0.6666666666666666, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_compose": 1.0, "train_accuracy_recipient_token_compose_total": 4, "train_accuracy_recipient_token_crop": 1.0, "train_accuracy_recipient_token_crop_total": 1, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.1956521739130435, "grad_norm": 2.438873695871041, "learning_rate": 4.656988227812658e-06, "loss": 0.3234, "step": 66 }, { "epoch": 0.1956521739130435, "step": 66, "train_accuracy_first_token_arguments": 0.9534883720930233, "train_accuracy_first_token_arguments_total": 43, "train_accuracy_recipient_token": 0.7058823529411765, "train_accuracy_recipient_token_all": 0.6923076923076923, "train_accuracy_recipient_token_all_total": 13, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_schedule": 0.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.1956521739130435, "step": 66, "train_accuracy_first_token_arguments": 0.55, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 0.7, "train_accuracy_recipient_token_add": 0.0, "train_accuracy_recipient_token_add_total": 2, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 0.3333333333333333, "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_multiply": 0.0, "train_accuracy_recipient_token_multiply_total": 2, "train_total_number_recipient_token": 20 }, { "epoch": 0.1956521739130435, "step": 66, "train_accuracy_first_token_arguments": 0.9090909090909091, "train_accuracy_first_token_arguments_total": 11, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { "epoch": 0.19861660079051383, "grad_norm": 2.7417774764849683, "learning_rate": 4.644708482352093e-06, "loss": 0.2789, "step": 67 }, { "epoch": 0.19861660079051383, "step": 67, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_get": 0.875, "train_accuracy_recipient_token_get_total": 8, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.19861660079051383, "step": 67, "train_accuracy_first_token_arguments": 0.9047619047619048, "train_accuracy_first_token_arguments_total": 21, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.19861660079051383, "step": 67, "train_accuracy_first_token_arguments": 0.8846153846153846, "train_accuracy_first_token_arguments_total": 26, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_build": 1.0, "train_accuracy_recipient_token_build_total": 1, "train_accuracy_recipient_token_conduct": 0.0, "train_accuracy_recipient_token_conduct_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_max": 1.0, "train_accuracy_recipient_token_max_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.2015810276679842, "grad_norm": 2.3471565473735976, "learning_rate": 4.632229564355275e-06, "loss": 0.1919, "step": 68 }, { "epoch": 0.2015810276679842, "step": 68, "train_accuracy_first_token_arguments": 0.8888888888888888, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 0.75, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_be": 0.0, "train_accuracy_recipient_token_be_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.2015810276679842, "step": 68, "train_accuracy_first_token_arguments": 0.6551724137931034, "train_accuracy_first_token_arguments_total": 29, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_identify": 0.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.2015810276679842, "step": 68, "train_accuracy_first_token_arguments": 0.6, "train_accuracy_first_token_arguments_total": 25, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_choose": 0.0, "train_accuracy_recipient_token_choose_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_order": 0.0, "train_accuracy_recipient_token_order_total": 1, "train_accuracy_recipient_token_plan": 0.0, "train_accuracy_recipient_token_plan_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.20454545454545456, "grad_norm": 2.845262992508057, "learning_rate": 4.619552632701263e-06, "loss": 0.2993, "step": 69 }, { "epoch": 0.20454545454545456, "step": 69, "train_accuracy_first_token_arguments": 0.9166666666666666, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_armor": 1.0, "train_accuracy_recipient_token_armor_total": 4, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.20454545454545456, "step": 69, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 36, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.8461538461538461, "train_accuracy_recipient_token_all_total": 13, "train_accuracy_recipient_token_customize": 1.0, "train_accuracy_recipient_token_customize_total": 2, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.20454545454545456, "step": 69, "train_accuracy_first_token_arguments": 0.9166666666666666, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 0.7142857142857143, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { "epoch": 0.2075098814229249, "grad_norm": 2.2300979588264487, "learning_rate": 4.606678864658039e-06, "loss": 0.2752, "step": 70 }, { "epoch": 0.2075098814229249, "step": 70, "train_accuracy_first_token_arguments": 0.6666666666666666, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 2, "train_accuracy_recipient_token_convert": 1.0, "train_accuracy_recipient_token_convert_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_render": 0.0, "train_accuracy_recipient_token_render_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.2075098814229249, "step": 70, "train_accuracy_first_token_arguments": 0.9827586206896551, "train_accuracy_first_token_arguments_total": 58, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 0.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_mon": 0.75, "train_accuracy_recipient_token_mon_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.2075098814229249, "step": 70, "train_accuracy_first_token_arguments": 0.9047619047619048, "train_accuracy_first_token_arguments_total": 21, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.21047430830039526, "grad_norm": 2.1107565687083176, "learning_rate": 4.5936094557731815e-06, "loss": 0.228, "step": 71 }, { "epoch": 0.21047430830039526, "step": 71, "train_accuracy_first_token_arguments": 0.7307692307692307, "train_accuracy_first_token_arguments_total": 26, "train_accuracy_recipient_token": 0.7894736842105263, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.42857142857142855, "train_accuracy_recipient_token_get_total": 7, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.21047430830039526, "step": 71, "train_accuracy_first_token_arguments": 0.7777777777777778, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_flower": 0.75, "train_accuracy_recipient_token_flower_total": 4, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.21047430830039526, "step": 71, "train_accuracy_first_token_arguments": 0.7333333333333333, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.2134387351778656, "grad_norm": 2.1086876066685165, "learning_rate": 4.5803456197628374e-06, "loss": 0.2305, "step": 72 }, { "epoch": 0.2134387351778656, "step": 72, "train_accuracy_first_token_arguments": 0.8823529411764706, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 0.8888888888888888, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 20 }, { "epoch": 0.2134387351778656, "step": 72, "train_accuracy_first_token_arguments": 0.5333333333333333, "train_accuracy_first_token_arguments_total": 30, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_draw": 1.0, "train_accuracy_recipient_token_draw_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.6, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.2134387351778656, "step": 72, "train_accuracy_first_token_arguments": 0.9, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 0.8, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_prior": 0.0, "train_accuracy_recipient_token_prior_total": 1, "train_accuracy_recipient_token_report": 0.0, "train_accuracy_recipient_token_report_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_accuracy_recipient_token_store": 0.5, "train_accuracy_recipient_token_store_total": 2, "train_total_number_recipient_token": 15 }, { "epoch": 0.21640316205533597, "grad_norm": 2.406206005805186, "learning_rate": 4.566888588399007e-06, "loss": 0.3295, "step": 73 }, { "epoch": 0.21640316205533597, "step": 73, "train_accuracy_first_token_arguments": 0.8846153846153846, "train_accuracy_first_token_arguments_total": 26, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8888888888888888, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.21640316205533597, "step": 73, "train_accuracy_first_token_arguments": 0.875, "train_accuracy_first_token_arguments_total": 24, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_log": 0.5, "train_accuracy_recipient_token_log_total": 2, "train_total_number_recipient_token": 16 }, { "epoch": 0.21640316205533597, "step": 73, "train_accuracy_first_token_arguments": 0.9285714285714286, "train_accuracy_first_token_arguments_total": 28, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_hire": 1.0, "train_accuracy_recipient_token_hire_total": 1, "train_accuracy_recipient_token_track": 1.0, "train_accuracy_recipient_token_track_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.21936758893280633, "grad_norm": 2.1304599088968823, "learning_rate": 4.553239611395156e-06, "loss": 0.2566, "step": 74 }, { "epoch": 0.21936758893280633, "step": 74, "train_accuracy_first_token_arguments": 0.8947368421052632, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_project": 1.0, "train_accuracy_recipient_token_project_total": 1, "train_accuracy_recipient_token_stat": 1.0, "train_accuracy_recipient_token_stat_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.21936758893280633, "step": 74, "train_accuracy_first_token_arguments": 0.5517241379310345, "train_accuracy_first_token_arguments_total": 29, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_accuracy_recipient_token_suggest": 1.0, "train_accuracy_recipient_token_suggest_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.21936758893280633, "step": 74, "train_accuracy_first_token_arguments": 0.75, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.7058823529411765, "train_accuracy_recipient_token_all": 0.6363636363636364, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_church": 1.0, "train_accuracy_recipient_token_church_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 17 }, { "epoch": 0.22233201581027667, "grad_norm": 2.03869187550964, "learning_rate": 4.539399956290152e-06, "loss": 0.3017, "step": 75 }, { "epoch": 0.22233201581027667, "step": 75, "train_accuracy_first_token_arguments": 0.7391304347826086, "train_accuracy_first_token_arguments_total": 46, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_f": 0.0, "train_accuracy_recipient_token_f_total": 1, "train_accuracy_recipient_token_plant": 1.0, "train_accuracy_recipient_token_plant_total": 1, "train_accuracy_recipient_token_se": 0.0, "train_accuracy_recipient_token_se_total": 1, "train_accuracy_recipient_token_trim": 1.0, "train_accuracy_recipient_token_trim_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.22233201581027667, "step": 75, "train_accuracy_first_token_arguments": 0.9285714285714286, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.7857142857142857, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_f": 1.0, "train_accuracy_recipient_token_f_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_restore": 0.0, "train_accuracy_recipient_token_restore_total": 1, "train_total_number_recipient_token": 14 }, { "epoch": 0.22233201581027667, "step": 75, "train_accuracy_first_token_arguments": 0.9411764705882353, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_gr": 0.6666666666666666, "train_accuracy_recipient_token_gr_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.22529644268774704, "grad_norm": 2.1715977940434494, "learning_rate": 4.525370908330564e-06, "loss": 0.2781, "step": 76 }, { "epoch": 0.22529644268774704, "step": 76, "train_accuracy_first_token_arguments": 0.8421052631578947, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.7647058823529411, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 0.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.22529644268774704, "step": 76, "train_accuracy_first_token_arguments": 0.71875, "train_accuracy_first_token_arguments_total": 64, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.22529644268774704, "step": 76, "train_accuracy_first_token_arguments": 0.9523809523809523, "train_accuracy_first_token_arguments_total": 21, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.22826086956521738, "grad_norm": 2.6029973414283223, "learning_rate": 4.511153770351288e-06, "loss": 0.3502, "step": 77 }, { "epoch": 0.22826086956521738, "step": 77, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { "epoch": 0.22826086956521738, "step": 77, "train_accuracy_first_token_arguments": 0.6153846153846154, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_flower": 1.0, "train_accuracy_recipient_token_flower_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.22826086956521738, "step": 77, "train_accuracy_first_token_arguments": 0.7049180327868853, "train_accuracy_first_token_arguments_total": 61, "train_accuracy_recipient_token": 0.9333333333333333, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 15 }, { "epoch": 0.23122529644268774, "grad_norm": 2.67404592050148, "learning_rate": 4.496749862654574e-06, "loss": 0.3309, "step": 78 }, { "epoch": 0.23122529644268774, "step": 78, "train_accuracy_first_token_arguments": 0.9, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_add": 1.0, "train_accuracy_recipient_token_add_total": 1, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 19 }, { "epoch": 0.23122529644268774, "step": 78, "train_accuracy_first_token_arguments": 0.95, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 20 }, { "epoch": 0.23122529644268774, "step": 78, "train_accuracy_first_token_arguments": 0.84, "train_accuracy_first_token_arguments_total": 25, "train_accuracy_recipient_token": 0.9523809523809523, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_extract": 1.0, "train_accuracy_recipient_token_extract_total": 1, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 2, "train_accuracy_recipient_token_sum": 1.0, "train_accuracy_recipient_token_sum_total": 1, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 2, "train_total_number_recipient_token": 21 }, { "epoch": 0.2341897233201581, "grad_norm": 1.8475653464555015, "learning_rate": 4.482160522887404e-06, "loss": 0.228, "step": 79 }, { "epoch": 0.2341897233201581, "step": 79, "train_accuracy_first_token_arguments": 0.6153846153846154, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.2341897233201581, "step": 79, "train_accuracy_first_token_arguments": 0.75, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_arc": 1.0, "train_accuracy_recipient_token_arc_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.2341897233201581, "step": 79, "train_accuracy_first_token_arguments": 0.4444444444444444, "train_accuracy_first_token_arguments_total": 45, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 20 }, { "epoch": 0.23715415019762845, "grad_norm": 2.189256435693527, "learning_rate": 4.467387105917269e-06, "loss": 0.3682, "step": 80 }, { "epoch": 0.23715415019762845, "step": 80, "train_accuracy_first_token_arguments": 0.8275862068965517, "train_accuracy_first_token_arguments_total": 29, "train_accuracy_recipient_token": 0.8125, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_property": 1.0, "train_accuracy_recipient_token_property_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.23715415019762845, "step": 80, "train_accuracy_first_token_arguments": 0.7333333333333333, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.23715415019762845, "step": 80, "train_accuracy_first_token_arguments": 0.8571428571428571, "train_accuracy_first_token_arguments_total": 21, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.24011857707509882, "grad_norm": 1.6646080394168112, "learning_rate": 4.452430983706351e-06, "loss": 0.1579, "step": 81 }, { "epoch": 0.24011857707509882, "step": 81, "train_accuracy_first_token_arguments": 0.875, "train_accuracy_first_token_arguments_total": 8, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 2, "train_accuracy_recipient_token_explain": 1.0, "train_accuracy_recipient_token_explain_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_sc": 1.0, "train_accuracy_recipient_token_sc_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.24011857707509882, "step": 81, "train_accuracy_first_token_arguments": 0.8, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.8125, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.24011857707509882, "step": 81, "train_accuracy_first_token_arguments": 0.75, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_box": 1.0, "train_accuracy_recipient_token_box_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.24308300395256918, "grad_norm": 1.9011227117633136, "learning_rate": 4.437293545184111e-06, "loss": 0.2544, "step": 82 }, { "epoch": 0.24308300395256918, "step": 82, "train_accuracy_first_token_arguments": 0.9090909090909091, "train_accuracy_first_token_arguments_total": 11, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.24308300395256918, "step": 82, "train_accuracy_first_token_arguments": 0.631578947368421, "train_accuracy_first_token_arguments_total": 38, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_inspect": 1.0, "train_accuracy_recipient_token_inspect_total": 1, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.24308300395256918, "step": 82, "train_accuracy_first_token_arguments": 0.8846153846153846, "train_accuracy_first_token_arguments_total": 26, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { "epoch": 0.24604743083003952, "grad_norm": 1.9670272993707267, "learning_rate": 4.421976196118297e-06, "loss": 0.2366, "step": 83 }, { "epoch": 0.24604743083003952, "step": 83, "train_accuracy_first_token_arguments": 0.9090909090909091, "train_accuracy_first_token_arguments_total": 22, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.24604743083003952, "step": 83, "train_accuracy_first_token_arguments": 0.7368421052631579, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.7777777777777778, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 0.5, "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.24604743083003952, "step": 83, "train_accuracy_first_token_arguments": 0.7, "train_accuracy_first_token_arguments_total": 10, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compile": 1.0, "train_accuracy_recipient_token_compile_total": 2, "train_accuracy_recipient_token_compress": 1.0, "train_accuracy_recipient_token_compress_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.2490118577075099, "grad_norm": 2.214361868557108, "learning_rate": 4.40648035898441e-06, "loss": 0.3157, "step": 84 }, { "epoch": 0.2490118577075099, "step": 84, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.2490118577075099, "step": 84, "train_accuracy_first_token_arguments": 0.9285714285714286, "train_accuracy_first_token_arguments_total": 28, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_research": 1.0, "train_accuracy_recipient_token_research_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.2490118577075099, "step": 84, "train_accuracy_first_token_arguments": 0.8888888888888888, "train_accuracy_first_token_arguments_total": 9, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compose": 1.0, "train_accuracy_recipient_token_compose_total": 4, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.2519762845849802, "grad_norm": 1.79425216853278, "learning_rate": 4.390807472833585e-06, "loss": 0.2629, "step": 85 }, { "epoch": 0.2519762845849802, "step": 85, "train_accuracy_first_token_arguments": 0.9375, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_write": 1.0, "train_accuracy_recipient_token_write_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.2519762845849802, "step": 85, "train_accuracy_first_token_arguments": 0.8461538461538461, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.7777777777777778, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.2519762845849802, "step": 85, "train_accuracy_first_token_arguments": 0.7272727272727273, "train_accuracy_first_token_arguments_total": 11, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_explain": 1.0, "train_accuracy_recipient_token_explain_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.2549407114624506, "grad_norm": 2.256598889106997, "learning_rate": 4.374958993158965e-06, "loss": 0.2383, "step": 86 }, { "epoch": 0.2549407114624506, "step": 86, "train_accuracy_first_token_arguments": 0.7368421052631579, "train_accuracy_first_token_arguments_total": 38, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_flag": 1.0, "train_accuracy_recipient_token_flag_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.2549407114624506, "step": 86, "train_accuracy_first_token_arguments": 0.8695652173913043, "train_accuracy_first_token_arguments_total": 23, "train_accuracy_recipient_token": 0.7894736842105263, "train_accuracy_recipient_token_all": 0.8461538461538461, "train_accuracy_recipient_token_all_total": 13, "train_accuracy_recipient_token_compare": 0.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_describe": 1.0, "train_accuracy_recipient_token_describe_total": 4, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.2549407114624506, "step": 86, "train_accuracy_first_token_arguments": 0.8378378378378378, "train_accuracy_first_token_arguments_total": 37, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_ass": 0.5, "train_accuracy_recipient_token_ass_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_plan": 1.0, "train_accuracy_recipient_token_plan_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.25790513833992096, "grad_norm": 2.49197628899743, "learning_rate": 4.358936391760524e-06, "loss": 0.4685, "step": 87 }, { "epoch": 0.25790513833992096, "step": 87, "train_accuracy_first_token_arguments": 0.9285714285714286, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_accuracy_recipient_token_identify": 0.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.25790513833992096, "step": 87, "train_accuracy_first_token_arguments": 0.9583333333333334, "train_accuracy_first_token_arguments_total": 24, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9166666666666666, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_cross": 1.0, "train_accuracy_recipient_token_cross_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.25790513833992096, "step": 87, "train_accuracy_first_token_arguments": 0.9047619047619048, "train_accuracy_first_token_arguments_total": 21, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 18 }, { "epoch": 0.2608695652173913, "grad_norm": 2.280644257269206, "learning_rate": 4.342741156608392e-06, "loss": 0.2713, "step": 88 }, { "epoch": 0.2608695652173913, "step": 88, "train_accuracy_first_token_arguments": 0.9444444444444444, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 16 }, { "epoch": 0.2608695652173913, "step": 88, "train_accuracy_first_token_arguments": 0.7407407407407407, "train_accuracy_first_token_arguments_total": 27, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_track": 1.0, "train_accuracy_recipient_token_track_total": 1, "train_accuracy_recipient_token_write": 0.0, "train_accuracy_recipient_token_write_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.2608695652173913, "step": 88, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 1, "train_accuracy_recipient_token_rent": 1.0, "train_accuracy_recipient_token_rent_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.26383399209486164, "grad_norm": 2.337688603854322, "learning_rate": 4.32637479170467e-06, "loss": 0.3134, "step": 89 }, { "epoch": 0.26383399209486164, "step": 89, "train_accuracy_first_token_arguments": 0.6666666666666666, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_explain": 0.0, "train_accuracy_recipient_token_explain_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_track": 1.0, "train_accuracy_recipient_token_track_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.26383399209486164, "step": 89, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_evaluate": 1.0, "train_accuracy_recipient_token_evaluate_total": 1, "train_accuracy_recipient_token_fetch": 1.0, "train_accuracy_recipient_token_fetch_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_sport": 1.0, "train_accuracy_recipient_token_sport_total": 3, "train_total_number_recipient_token": 20 }, { "epoch": 0.26383399209486164, "step": 89, "train_accuracy_first_token_arguments": 0.9047619047619048, "train_accuracy_first_token_arguments_total": 21, "train_accuracy_recipient_token": 0.8125, "train_accuracy_recipient_token_add": 1.0, "train_accuracy_recipient_token_add_total": 1, "train_accuracy_recipient_token_all": 0.7692307692307693, "train_accuracy_recipient_token_all_total": 13, "train_accuracy_recipient_token_apply": 1.0, "train_accuracy_recipient_token_apply_total": 1, "train_accuracy_recipient_token_crop": 1.0, "train_accuracy_recipient_token_crop_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.26679841897233203, "grad_norm": 2.596535079636875, "learning_rate": 4.309838816943755e-06, "loss": 0.2432, "step": 90 }, { "epoch": 0.26679841897233203, "step": 90, "train_accuracy_first_token_arguments": 0.76, "train_accuracy_first_token_arguments_total": 25, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.26679841897233203, "step": 90, "train_accuracy_first_token_arguments": 0.9333333333333333, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_m": 1.0, "train_accuracy_recipient_token_m_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.26679841897233203, "step": 90, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.75, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 16 }, { "epoch": 0.26976284584980237, "grad_norm": 2.2037749517640015, "learning_rate": 4.293134767971193e-06, "loss": 0.2113, "step": 91 }, { "epoch": 0.26976284584980237, "step": 91, "train_accuracy_first_token_arguments": 0.7, "train_accuracy_first_token_arguments_total": 10, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_ele": 1.0, "train_accuracy_recipient_token_ele_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 17 }, { "epoch": 0.26976284584980237, "step": 91, "train_accuracy_first_token_arguments": 0.6363636363636364, "train_accuracy_first_token_arguments_total": 11, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 16 }, { "epoch": 0.26976284584980237, "step": 91, "train_accuracy_first_token_arguments": 0.8103448275862069, "train_accuracy_first_token_arguments_total": 58, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_accuracy_recipient_token_object": 1.0, "train_accuracy_recipient_token_object_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.2727272727272727, "grad_norm": 2.1267599638494628, "learning_rate": 4.276264196041074e-06, "loss": 0.3654, "step": 92 }, { "epoch": 0.2727272727272727, "step": 92, "train_accuracy_first_token_arguments": 0.8292682926829268, "train_accuracy_first_token_arguments_total": 41, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_deploy": 1.0, "train_accuracy_recipient_token_deploy_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_plan": 1.0, "train_accuracy_recipient_token_plan_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.2727272727272727, "step": 92, "train_accuracy_first_token_arguments": 0.7105263157894737, "train_accuracy_first_token_arguments_total": 38, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_convert": 1.0, "train_accuracy_recipient_token_convert_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.2727272727272727, "step": 92, "train_accuracy_first_token_arguments": 0.8421052631578947, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_detect": 0.0, "train_accuracy_recipient_token_detect_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.2756916996047431, "grad_norm": 2.2199589796264414, "learning_rate": 4.259228667871963e-06, "loss": 0.3508, "step": 93 }, { "epoch": 0.2756916996047431, "step": 93, "train_accuracy_first_token_arguments": 0.9444444444444444, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 0.7894736842105263, "train_accuracy_recipient_token_add": 1.0, "train_accuracy_recipient_token_add_total": 1, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_calculate": 0.5, "train_accuracy_recipient_token_calculate_total": 2, "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 4, "train_accuracy_recipient_token_subtract": 0.0, "train_accuracy_recipient_token_subtract_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.2756916996047431, "step": 93, "train_accuracy_first_token_arguments": 0.7407407407407407, "train_accuracy_first_token_arguments_total": 27, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.2756916996047431, "step": 93, "train_accuracy_first_token_arguments": 0.875, "train_accuracy_first_token_arguments_total": 24, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_study": 1.0, "train_accuracy_recipient_token_study_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.27865612648221344, "grad_norm": 2.6263899594688755, "learning_rate": 4.242029765501411e-06, "loss": 0.2873, "step": 94 }, { "epoch": 0.27865612648221344, "step": 94, "train_accuracy_first_token_arguments": 0.9473684210526315, "train_accuracy_first_token_arguments_total": 76, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_d": 0.5, "train_accuracy_recipient_token_d_total": 2, "train_accuracy_recipient_token_fly": 1.0, "train_accuracy_recipient_token_fly_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.27865612648221344, "step": 94, "train_accuracy_first_token_arguments": 0.64, "train_accuracy_first_token_arguments_total": 25, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.27865612648221344, "step": 94, "train_accuracy_first_token_arguments": 0.8125, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_horse": 1.0, "train_accuracy_recipient_token_horse_total": 2, "train_accuracy_recipient_token_invest": 1.0, "train_accuracy_recipient_token_invest_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.2816205533596838, "grad_norm": 1.8705277085481686, "learning_rate": 4.22466908613903e-06, "loss": 0.26, "step": 95 }, { "epoch": 0.2816205533596838, "step": 95, "train_accuracy_first_token_arguments": 0.8076923076923077, "train_accuracy_first_token_arguments_total": 26, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_arr": 1.0, "train_accuracy_recipient_token_arr_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_accuracy_recipient_token_suggest": 1.0, "train_accuracy_recipient_token_suggest_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.2816205533596838, "step": 95, "train_accuracy_first_token_arguments": 0.5510204081632653, "train_accuracy_first_token_arguments_total": 49, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_f": 1.0, "train_accuracy_recipient_token_f_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_im": 1.0, "train_accuracy_recipient_token_im_total": 1, "train_accuracy_recipient_token_schedule": 0.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.2816205533596838, "step": 95, "train_accuracy_first_token_arguments": 0.8947368421052632, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_sh": 1.0, "train_accuracy_recipient_token_sh_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.2845849802371542, "grad_norm": 2.0123868841853065, "learning_rate": 4.20714824201817e-06, "loss": 0.3846, "step": 96 }, { "epoch": 0.2845849802371542, "step": 96, "train_accuracy_first_token_arguments": 0.8928571428571429, "train_accuracy_first_token_arguments_total": 28, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 18 }, { "epoch": 0.2845849802371542, "step": 96, "train_accuracy_first_token_arguments": 0.88, "train_accuracy_first_token_arguments_total": 25, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_accuracy_recipient_token_verify": 1.0, "train_accuracy_recipient_token_verify_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.2845849802371542, "step": 96, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 26, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8888888888888888, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.2875494071146245, "grad_norm": 1.6695567253827632, "learning_rate": 4.189468860246192e-06, "loss": 0.1603, "step": 97 }, { "epoch": 0.2875494071146245, "step": 97, "train_accuracy_first_token_arguments": 0.75, "train_accuracy_first_token_arguments_total": 36, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_enh": 1.0, "train_accuracy_recipient_token_enh_total": 1, "train_accuracy_recipient_token_explain": 1.0, "train_accuracy_recipient_token_explain_total": 1, "train_accuracy_recipient_token_public": 1.0, "train_accuracy_recipient_token_public_total": 1, "train_accuracy_recipient_token_street": 1.0, "train_accuracy_recipient_token_street_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.2875494071146245, "step": 97, "train_accuracy_first_token_arguments": 0.9090909090909091, "train_accuracy_first_token_arguments_total": 44, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.75, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 17 }, { "epoch": 0.2875494071146245, "step": 97, "train_accuracy_first_token_arguments": 0.8947368421052632, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 17 }, { "epoch": 0.29051383399209485, "grad_norm": 2.3366857423927834, "learning_rate": 4.171632582653368e-06, "loss": 0.3712, "step": 98 }, { "epoch": 0.29051383399209485, "step": 98, "train_accuracy_first_token_arguments": 0.8636363636363636, "train_accuracy_first_token_arguments_total": 22, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 0.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_c": 0.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.29051383399209485, "step": 98, "train_accuracy_first_token_arguments": 0.8260869565217391, "train_accuracy_first_token_arguments_total": 23, "train_accuracy_recipient_token": 0.8125, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 0.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_assign": 1.0, "train_accuracy_recipient_token_assign_total": 1, "train_accuracy_recipient_token_configure": 1.0, "train_accuracy_recipient_token_configure_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 16 }, { "epoch": 0.29051383399209485, "step": 98, "train_accuracy_first_token_arguments": 0.6153846153846154, "train_accuracy_first_token_arguments_total": 52, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.29347826086956524, "grad_norm": 2.2368586683604175, "learning_rate": 4.153641065640402e-06, "loss": 0.238, "step": 99 }, { "epoch": 0.29347826086956524, "step": 99, "train_accuracy_first_token_arguments": 0.9375, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.8, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 20 }, { "epoch": 0.29347826086956524, "step": 99, "train_accuracy_first_token_arguments": 0.8, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_clean": 1.0, "train_accuracy_recipient_token_clean_total": 1, "train_accuracy_recipient_token_install": 1.0, "train_accuracy_recipient_token_install_total": 1, "train_accuracy_recipient_token_remove": 0.0, "train_accuracy_recipient_token_remove_total": 1, "train_accuracy_recipient_token_replace": 0.0, "train_accuracy_recipient_token_replace_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.29347826086956524, "step": 99, "train_accuracy_first_token_arguments": 0.9411764705882353, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 4, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.2964426877470356, "grad_norm": 2.6698723558857944, "learning_rate": 4.1354959800246155e-06, "loss": 0.2791, "step": 100 }, { "epoch": 0.2964426877470356, "step": 100, "train_accuracy_first_token_arguments": 0.6666666666666666, "train_accuracy_first_token_arguments_total": 21, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_facility": 1.0, "train_accuracy_recipient_token_facility_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.2964426877470356, "step": 100, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 0.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_mount": 1.0, "train_accuracy_recipient_token_mount_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.2964426877470356, "step": 100, "train_accuracy_first_token_arguments": 0.71875, "train_accuracy_first_token_arguments_total": 32, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 0.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_park": 0.6666666666666666, "train_accuracy_recipient_token_park_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.2994071146245059, "grad_norm": 1.5724579692892722, "learning_rate": 4.1171990108847705e-06, "loss": 0.1851, "step": 101 }, { "epoch": 0.2994071146245059, "step": 101, "train_accuracy_first_token_arguments": 0.9069767441860465, "train_accuracy_first_token_arguments_total": 43, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_accuracy_recipient_token_schedule": 0.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.2994071146245059, "step": 101, "train_accuracy_first_token_arguments": 0.5909090909090909, "train_accuracy_first_token_arguments_total": 22, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_detect": 1.0, "train_accuracy_recipient_token_detect_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.2994071146245059, "step": 101, "train_accuracy_first_token_arguments": 0.7692307692307693, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 18 }, { "epoch": 0.30237154150197626, "grad_norm": 1.3808933034142918, "learning_rate": 4.098751857404595e-06, "loss": 0.1429, "step": 102 }, { "epoch": 0.30237154150197626, "step": 102, "train_accuracy_first_token_arguments": 0.9166666666666666, "train_accuracy_first_token_arguments_total": 24, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_detect": 1.0, "train_accuracy_recipient_token_detect_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_sum": 1.0, "train_accuracy_recipient_token_sum_total": 1, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.30237154150197626, "step": 102, "train_accuracy_first_token_arguments": 0.5833333333333334, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_city": 1.0, "train_accuracy_recipient_token_city_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.30237154150197626, "step": 102, "train_accuracy_first_token_arguments": 0.8421052631578947, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_event": 1.0, "train_accuracy_recipient_token_event_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.30533596837944665, "grad_norm": 2.4623439109068372, "learning_rate": 4.080156232714976e-06, "loss": 0.4204, "step": 103 }, { "epoch": 0.30533596837944665, "step": 103, "train_accuracy_first_token_arguments": 0.95, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_bike": 1.0, "train_accuracy_recipient_token_bike_total": 2, "train_accuracy_recipient_token_calculate": 0.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.30533596837944665, "step": 103, "train_accuracy_first_token_arguments": 0.9, "train_accuracy_first_token_arguments_total": 10, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9166666666666666, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_disc": 1.0, "train_accuracy_recipient_token_disc_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 16 }, { "epoch": 0.30533596837944665, "step": 103, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_convert": 1.0, "train_accuracy_recipient_token_convert_total": 2, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 21 }, { "epoch": 0.308300395256917, "grad_norm": 1.9389616807058327, "learning_rate": 4.061413863734869e-06, "loss": 0.169, "step": 104 }, { "epoch": 0.308300395256917, "step": 104, "train_accuracy_first_token_arguments": 0.7, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_accuracy_recipient_token_interpret": 1.0, "train_accuracy_recipient_token_interpret_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.308300395256917, "step": 104, "train_accuracy_first_token_arguments": 0.9523809523809523, "train_accuracy_first_token_arguments_total": 21, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 5, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 20 }, { "epoch": 0.308300395256917, "step": 104, "train_accuracy_first_token_arguments": 0.8421052631578947, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 1, "train_accuracy_recipient_token_villa": 1.0, "train_accuracy_recipient_token_villa_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.31126482213438733, "grad_norm": 1.819003861346643, "learning_rate": 4.0425264910109245e-06, "loss": 0.1843, "step": 105 }, { "epoch": 0.31126482213438733, "step": 105, "train_accuracy_first_token_arguments": 0.8333333333333334, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_m": 1.0, "train_accuracy_recipient_token_m_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.31126482213438733, "step": 105, "train_accuracy_first_token_arguments": 0.875, "train_accuracy_first_token_arguments_total": 8, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 1, "train_accuracy_recipient_token_sum": 1.0, "train_accuracy_recipient_token_sum_total": 1, "train_accuracy_recipient_token_verify": 1.0, "train_accuracy_recipient_token_verify_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.31126482213438733, "step": 105, "train_accuracy_first_token_arguments": 0.8888888888888888, "train_accuracy_first_token_arguments_total": 9, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.3142292490118577, "grad_norm": 2.4539744307883553, "learning_rate": 4.023495868555848e-06, "loss": 0.4031, "step": 106 }, { "epoch": 0.3142292490118577, "eval_accuracy_first_token_arguments": 0.9266727772685609, "eval_accuracy_first_token_arguments_total": 1091, "eval_accuracy_recipient_token": 0.9235033259423503, "eval_accuracy_recipient_token_all": 0.9044834307992202, "eval_accuracy_recipient_token_all_total": 513, "eval_accuracy_recipient_token_calculate": 1.0, "eval_accuracy_recipient_token_calculate_total": 13, "eval_accuracy_recipient_token_find": 0.9649122807017544, "eval_accuracy_recipient_token_find_total": 57, "eval_accuracy_recipient_token_get": 0.9744897959183674, "eval_accuracy_recipient_token_get_total": 196, "eval_accuracy_recipient_token_identify": 0.8333333333333334, "eval_accuracy_recipient_token_identify_total": 6, "eval_loss": 0.25004321336746216, "eval_perplexity": 1.0396230889094866, "eval_runtime": 119.1349, "eval_samples_per_second": 0.848, "eval_steps_per_second": 0.428, "eval_total_number_recipient_token": 902, "step": 106 }, { "epoch": 0.3142292490118577, "step": 106, "train_accuracy_first_token_arguments": 0.8, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.3142292490118577, "step": 106, "train_accuracy_first_token_arguments": 0.88, "train_accuracy_first_token_arguments_total": 25, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_report": 1.0, "train_accuracy_recipient_token_report_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.3142292490118577, "step": 106, "train_accuracy_first_token_arguments": 0.4166666666666667, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_interpret": 1.0, "train_accuracy_recipient_token_interpret_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.31719367588932806, "grad_norm": 2.6219196171389454, "learning_rate": 4.004323763685511e-06, "loss": 0.2683, "step": 107 }, { "epoch": 0.31719367588932806, "step": 107, "train_accuracy_first_token_arguments": 0.9130434782608695, "train_accuracy_first_token_arguments_total": 23, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_d": 1.0, "train_accuracy_recipient_token_d_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.31719367588932806, "step": 107, "train_accuracy_first_token_arguments": 0.918918918918919, "train_accuracy_first_token_arguments_total": 37, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_audit": 1.0, "train_accuracy_recipient_token_audit_total": 1, "train_accuracy_recipient_token_environment": 1.0, "train_accuracy_recipient_token_environment_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_public": 1.0, "train_accuracy_recipient_token_public_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.31719367588932806, "step": 107, "train_accuracy_first_token_arguments": 0.8333333333333334, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_f": 1.0, "train_accuracy_recipient_token_f_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_search": 0.6666666666666666, "train_accuracy_recipient_token_search_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.3201581027667984, "grad_norm": 2.3979253771216524, "learning_rate": 3.985011956854826e-06, "loss": 0.3371, "step": 108 }, { "epoch": 0.3201581027667984, "step": 108, "train_accuracy_first_token_arguments": 0.9285714285714286, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.3201581027667984, "step": 108, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_histor": 1.0, "train_accuracy_recipient_token_histor_total": 2, "train_total_number_recipient_token": 16 }, { "epoch": 0.3201581027667984, "step": 108, "train_accuracy_first_token_arguments": 0.8571428571428571, "train_accuracy_first_token_arguments_total": 28, "train_accuracy_recipient_token": 0.5, "train_accuracy_recipient_token_all": 0.5454545454545454, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_detect": 0.0, "train_accuracy_recipient_token_detect_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_translate": 0.5, "train_accuracy_recipient_token_translate_total": 2, "train_total_number_recipient_token": 16 }, { "epoch": 0.3231225296442688, "grad_norm": 2.9561525888835276, "learning_rate": 3.965562241492401e-06, "loss": 0.3818, "step": 109 }, { "epoch": 0.3231225296442688, "step": 109, "train_accuracy_first_token_arguments": 0.7142857142857143, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_plan": 1.0, "train_accuracy_recipient_token_plan_total": 1, "train_accuracy_recipient_token_suggest": 1.0, "train_accuracy_recipient_token_suggest_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.3231225296442688, "step": 109, "train_accuracy_first_token_arguments": 0.9230769230769231, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 17 }, { "epoch": 0.3231225296442688, "step": 109, "train_accuracy_first_token_arguments": 0.96875, "train_accuracy_first_token_arguments_total": 32, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.32608695652173914, "grad_norm": 1.9549696262976763, "learning_rate": 3.945976423833987e-06, "loss": 0.1918, "step": 110 }, { "epoch": 0.32608695652173914, "step": 110, "train_accuracy_first_token_arguments": 0.9333333333333333, "train_accuracy_first_token_arguments_total": 30, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.7692307692307693, "train_accuracy_recipient_token_all_total": 13, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_detect": 1.0, "train_accuracy_recipient_token_detect_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.32608695652173914, "step": 110, "train_accuracy_first_token_arguments": 0.7894736842105263, "train_accuracy_first_token_arguments_total": 38, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_race": 1.0, "train_accuracy_recipient_token_race_total": 2, "train_total_number_recipient_token": 16 }, { "epoch": 0.32608695652173914, "step": 110, "train_accuracy_first_token_arguments": 0.9333333333333333, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9166666666666666, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_evaluate": 1.0, "train_accuracy_recipient_token_evaluate_total": 1, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 20 }, { "epoch": 0.3290513833992095, "grad_norm": 2.295435621606061, "learning_rate": 3.92625632275474e-06, "loss": 0.378, "step": 111 }, { "epoch": 0.3290513833992095, "step": 111, "train_accuracy_first_token_arguments": 0.8888888888888888, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_add": 1.0, "train_accuracy_recipient_token_add_total": 1, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_change": 1.0, "train_accuracy_recipient_token_change_total": 1, "train_accuracy_recipient_token_highlight": 0.5, "train_accuracy_recipient_token_highlight_total": 2, "train_accuracy_recipient_token_remove": 1.0, "train_accuracy_recipient_token_remove_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.3290513833992095, "step": 111, "train_accuracy_first_token_arguments": 0.896551724137931, "train_accuracy_first_token_arguments_total": 29, "train_accuracy_recipient_token": 0.7647058823529411, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_configure": 1.0, "train_accuracy_recipient_token_configure_total": 1, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.3290513833992095, "step": 111, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_total_number_recipient_token": 20 }, { "epoch": 0.33201581027667987, "grad_norm": 2.4185983751815114, "learning_rate": 3.906403769600311e-06, "loss": 0.2139, "step": 112 }, { "epoch": 0.33201581027667987, "step": 112, "train_accuracy_first_token_arguments": 0.9090909090909091, "train_accuracy_first_token_arguments_total": 22, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_dispose": 1.0, "train_accuracy_recipient_token_dispose_total": 1, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_up": 1.0, "train_accuracy_recipient_token_up_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.33201581027667987, "step": 112, "train_accuracy_first_token_arguments": 0.9230769230769231, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 17 }, { "epoch": 0.33201581027667987, "step": 112, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_accuracy_recipient_token_verify": 1.0, "train_accuracy_recipient_token_verify_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.3349802371541502, "grad_norm": 2.321966400855487, "learning_rate": 3.886420608016767e-06, "loss": 0.2771, "step": 113 }, { "epoch": 0.3349802371541502, "step": 113, "train_accuracy_first_token_arguments": 0.9666666666666667, "train_accuracy_first_token_arguments_total": 30, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_add": 1.0, "train_accuracy_recipient_token_add_total": 1, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_change": 1.0, "train_accuracy_recipient_token_change_total": 1, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_replace": 0.5, "train_accuracy_recipient_token_replace_total": 2, "train_total_number_recipient_token": 20 }, { "epoch": 0.3349802371541502, "step": 113, "train_accuracy_first_token_arguments": 0.6, "train_accuracy_first_token_arguments_total": 10, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 2, "train_accuracy_recipient_token_bird": 1.0, "train_accuracy_recipient_token_bird_total": 4, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.3349802371541502, "step": 113, "train_accuracy_first_token_arguments": 0.5833333333333334, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.7777777777777778, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.33794466403162055, "grad_norm": 2.028221423810436, "learning_rate": 3.8663086937793845e-06, "loss": 0.2435, "step": 114 }, { "epoch": 0.33794466403162055, "step": 114, "train_accuracy_first_token_arguments": 0.9473684210526315, "train_accuracy_first_token_arguments_total": 114, "train_accuracy_recipient_token": 0.7368421052631579, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_arch": 0.0, "train_accuracy_recipient_token_arch_total": 1, "train_accuracy_recipient_token_building": 1.0, "train_accuracy_recipient_token_building_total": 1, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_histor": 0.0, "train_accuracy_recipient_token_histor_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.33794466403162055, "step": 114, "train_accuracy_first_token_arguments": 0.5833333333333334, "train_accuracy_first_token_arguments_total": 36, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_im": 1.0, "train_accuracy_recipient_token_im_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.33794466403162055, "step": 114, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 11, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { "epoch": 0.3409090909090909, "grad_norm": 1.8641084259637979, "learning_rate": 3.846069894620306e-06, "loss": 0.2402, "step": 115 }, { "epoch": 0.3409090909090909, "step": 115, "train_accuracy_first_token_arguments": 0.75, "train_accuracy_first_token_arguments_total": 24, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 0.6, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_organ": 1.0, "train_accuracy_recipient_token_organ_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.3409090909090909, "step": 115, "train_accuracy_first_token_arguments": 0.8181818181818182, "train_accuracy_first_token_arguments_total": 22, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 0.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.3409090909090909, "step": 115, "train_accuracy_first_token_arguments": 0.84, "train_accuracy_first_token_arguments_total": 25, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_bookmark": 1.0, "train_accuracy_recipient_token_bookmark_total": 1, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_refresh": 1.0, "train_accuracy_recipient_token_refresh_total": 1, "train_accuracy_recipient_token_switch": 0.5, "train_accuracy_recipient_token_switch_total": 2, "train_total_number_recipient_token": 20 }, { "epoch": 0.3438735177865613, "grad_norm": 2.0312557408429788, "learning_rate": 3.825706090055088e-06, "loss": 0.2456, "step": 116 }, { "epoch": 0.3438735177865613, "step": 116, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_remove": 1.0, "train_accuracy_recipient_token_remove_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.3438735177865613, "step": 116, "train_accuracy_first_token_arguments": 0.7391304347826086, "train_accuracy_first_token_arguments_total": 23, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_suggest": 1.0, "train_accuracy_recipient_token_suggest_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.3438735177865613, "step": 116, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 2, "train_accuracy_recipient_token_design": 1.0, "train_accuracy_recipient_token_design_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.3468379446640316, "grad_norm": 2.9747145067980383, "learning_rate": 3.80521917120816e-06, "loss": 0.3586, "step": 117 }, { "epoch": 0.3468379446640316, "step": 117, "train_accuracy_first_token_arguments": 0.8043478260869565, "train_accuracy_first_token_arguments_total": 46, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 16 }, { "epoch": 0.3468379446640316, "step": 117, "train_accuracy_first_token_arguments": 0.8888888888888888, "train_accuracy_first_token_arguments_total": 27, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.3468379446640316, "step": 117, "train_accuracy_first_token_arguments": 0.7894736842105263, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_setup": 1.0, "train_accuracy_recipient_token_setup_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.34980237154150196, "grad_norm": 2.619218782176094, "learning_rate": 3.784611040637198e-06, "loss": 0.386, "step": 118 }, { "epoch": 0.34980237154150196, "step": 118, "train_accuracy_first_token_arguments": 0.9285714285714286, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.34980237154150196, "step": 118, "train_accuracy_first_token_arguments": 0.9, "train_accuracy_first_token_arguments_total": 10, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_project": 1.0, "train_accuracy_recipient_token_project_total": 1, "train_accuracy_recipient_token_sc": 1.0, "train_accuracy_recipient_token_sc_total": 2, "train_accuracy_recipient_token_verify": 1.0, "train_accuracy_recipient_token_verify_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.34980237154150196, "step": 118, "train_accuracy_first_token_arguments": 0.76, "train_accuracy_first_token_arguments_total": 25, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.35276679841897235, "grad_norm": 1.783725464528477, "learning_rate": 3.7638836121564414e-06, "loss": 0.3088, "step": 119 }, { "epoch": 0.35276679841897235, "step": 119, "train_accuracy_first_token_arguments": 0.8695652173913043, "train_accuracy_first_token_arguments_total": 23, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.9, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 20 }, { "epoch": 0.35276679841897235, "step": 119, "train_accuracy_first_token_arguments": 0.9090909090909091, "train_accuracy_first_token_arguments_total": 11, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_financial": 1.0, "train_accuracy_recipient_token_financial_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_invest": 1.0, "train_accuracy_recipient_token_invest_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.35276679841897235, "step": 119, "train_accuracy_first_token_arguments": 0.967741935483871, "train_accuracy_first_token_arguments_total": 31, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { "epoch": 0.3557312252964427, "grad_norm": 1.880678138487034, "learning_rate": 3.7430388106589632e-06, "loss": 0.2326, "step": 120 }, { "epoch": 0.3557312252964427, "step": 120, "train_accuracy_first_token_arguments": 0.7894736842105263, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_detect": 1.0, "train_accuracy_recipient_token_detect_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.3557312252964427, "step": 120, "train_accuracy_first_token_arguments": 0.7916666666666666, "train_accuracy_first_token_arguments_total": 24, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 2, "train_accuracy_recipient_token_buy": 1.0, "train_accuracy_recipient_token_buy_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_suggest": 1.0, "train_accuracy_recipient_token_suggest_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.3557312252964427, "step": 120, "train_accuracy_first_token_arguments": 0.8947368421052632, "train_accuracy_first_token_arguments_total": 57, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.358695652173913, "grad_norm": 2.3660230978435117, "learning_rate": 3.72207857193791e-06, "loss": 0.2103, "step": 121 }, { "epoch": 0.358695652173913, "step": 121, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 9, "train_accuracy_recipient_token": 0.7647058823529411, "train_accuracy_recipient_token_all": 0.6, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.358695652173913, "step": 121, "train_accuracy_first_token_arguments": 0.7575757575757576, "train_accuracy_first_token_arguments_total": 33, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_cl": 1.0, "train_accuracy_recipient_token_cl_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_im": 1.0, "train_accuracy_recipient_token_im_total": 1, "train_accuracy_recipient_token_search": 0.5, "train_accuracy_recipient_token_search_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.358695652173913, "step": 121, "train_accuracy_first_token_arguments": 0.8260869565217391, "train_accuracy_first_token_arguments_total": 23, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.3616600790513834, "grad_norm": 1.9095006331871722, "learning_rate": 3.7010048425067317e-06, "loss": 0.2754, "step": 122 }, { "epoch": 0.3616600790513834, "step": 122, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.3616600790513834, "step": 122, "train_accuracy_first_token_arguments": 0.8571428571428571, "train_accuracy_first_token_arguments_total": 28, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 3, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.3616600790513834, "step": 122, "train_accuracy_first_token_arguments": 0.8666666666666667, "train_accuracy_first_token_arguments_total": 45, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.36462450592885376, "grad_norm": 1.6047057072988866, "learning_rate": 3.679819579418414e-06, "loss": 0.1743, "step": 123 }, { "epoch": 0.36462450592885376, "step": 123, "train_accuracy_first_token_arguments": 0.71875, "train_accuracy_first_token_arguments_total": 32, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_arch": 1.0, "train_accuracy_recipient_token_arch_total": 2, "train_accuracy_recipient_token_build": 1.0, "train_accuracy_recipient_token_build_total": 3, "train_accuracy_recipient_token_histor": 0.0, "train_accuracy_recipient_token_histor_total": 1, "train_accuracy_recipient_token_suggest": 1.0, "train_accuracy_recipient_token_suggest_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.36462450592885376, "step": 123, "train_accuracy_first_token_arguments": 0.8, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9166666666666666, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.36462450592885376, "step": 123, "train_accuracy_first_token_arguments": 0.9375, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_land": 1.0, "train_accuracy_recipient_token_land_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.3675889328063241, "grad_norm": 1.6871352500543018, "learning_rate": 3.658524750083733e-06, "loss": 0.3106, "step": 124 }, { "epoch": 0.3675889328063241, "step": 124, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 7, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 16 }, { "epoch": 0.3675889328063241, "step": 124, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { "epoch": 0.3675889328063241, "step": 124, "train_accuracy_first_token_arguments": 0.9090909090909091, "train_accuracy_first_token_arguments_total": 22, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_evaluate": 1.0, "train_accuracy_recipient_token_evaluate_total": 1, "train_accuracy_recipient_token_phrase": 1.0, "train_accuracy_recipient_token_phrase_total": 4, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 1, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.3705533596837945, "grad_norm": 2.016083076720157, "learning_rate": 3.6371223320885492e-06, "loss": 0.3026, "step": 125 }, { "epoch": 0.3705533596837945, "step": 125, "train_accuracy_first_token_arguments": 0.6764705882352942, "train_accuracy_first_token_arguments_total": 34, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 1, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_plan": 0.0, "train_accuracy_recipient_token_plan_total": 1, "train_accuracy_recipient_token_suggest": 1.0, "train_accuracy_recipient_token_suggest_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.3705533596837945, "step": 125, "train_accuracy_first_token_arguments": 0.9090909090909091, "train_accuracy_first_token_arguments_total": 11, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.3705533596837945, "step": 125, "train_accuracy_first_token_arguments": 0.9523809523809523, "train_accuracy_first_token_arguments_total": 21, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_suggest": 1.0, "train_accuracy_recipient_token_suggest_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.37351778656126483, "grad_norm": 2.1413931479434973, "learning_rate": 3.6156143130101516e-06, "loss": 0.3208, "step": 126 }, { "epoch": 0.37351778656126483, "step": 126, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.8666666666666667, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_organ": 1.0, "train_accuracy_recipient_token_organ_total": 1, "train_total_number_recipient_token": 15 }, { "epoch": 0.37351778656126483, "step": 126, "train_accuracy_first_token_arguments": 0.9705882352941176, "train_accuracy_first_token_arguments_total": 34, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.37351778656126483, "step": 126, "train_accuracy_first_token_arguments": 0.8, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.37648221343873517, "grad_norm": 2.3478599036765084, "learning_rate": 3.5940026902326825e-06, "loss": 0.2732, "step": 127 }, { "epoch": 0.37648221343873517, "step": 127, "train_accuracy_first_token_arguments": 0.782608695652174, "train_accuracy_first_token_arguments_total": 69, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 18 }, { "epoch": 0.37648221343873517, "step": 127, "train_accuracy_first_token_arguments": 0.7631578947368421, "train_accuracy_first_token_arguments_total": 38, "train_accuracy_recipient_token": 0.7222222222222222, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 0.5, "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.37648221343873517, "step": 127, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8888888888888888, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.3794466403162055, "grad_norm": 1.8233289275067917, "learning_rate": 3.5722894707616417e-06, "loss": 0.2708, "step": 128 }, { "epoch": 0.3794466403162055, "step": 128, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.3794466403162055, "step": 128, "train_accuracy_first_token_arguments": 0.8095238095238095, "train_accuracy_first_token_arguments_total": 21, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_clean": 1.0, "train_accuracy_recipient_token_clean_total": 1, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 3, "train_accuracy_recipient_token_repair": 1.0, "train_accuracy_recipient_token_repair_total": 2, "train_accuracy_recipient_token_restore": 1.0, "train_accuracy_recipient_token_restore_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.3794466403162055, "step": 128, "train_accuracy_first_token_arguments": 0.9722222222222222, "train_accuracy_first_token_arguments_total": 36, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 0.5, "train_accuracy_recipient_token_book_total": 2, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_send": 1.0, "train_accuracy_recipient_token_send_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.3824110671936759, "grad_norm": 2.7033960410551843, "learning_rate": 3.550476671037505e-06, "loss": 0.3598, "step": 129 }, { "epoch": 0.3824110671936759, "step": 129, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 22, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.3824110671936759, "step": 129, "train_accuracy_first_token_arguments": 0.803921568627451, "train_accuracy_first_token_arguments_total": 51, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.3824110671936759, "step": 129, "train_accuracy_first_token_arguments": 0.7142857142857143, "train_accuracy_first_token_arguments_total": 7, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_arch": 1.0, "train_accuracy_recipient_token_arch_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 17 }, { "epoch": 0.38537549407114624, "grad_norm": 1.817103444485219, "learning_rate": 3.528566316748462e-06, "loss": 0.2691, "step": 130 }, { "epoch": 0.38537549407114624, "step": 130, "train_accuracy_first_token_arguments": 0.875, "train_accuracy_first_token_arguments_total": 8, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 17 }, { "epoch": 0.38537549407114624, "step": 130, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_log": 1.0, "train_accuracy_recipient_token_log_total": 4, "train_total_number_recipient_token": 17 }, { "epoch": 0.38537549407114624, "step": 130, "train_accuracy_first_token_arguments": 0.9411764705882353, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 0.5, "train_accuracy_recipient_token_book_total": 2, "train_accuracy_recipient_token_find": 0.75, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.3883399209486166, "grad_norm": 2.3178394841243724, "learning_rate": 3.5065604426422995e-06, "loss": 0.2196, "step": 131 }, { "epoch": 0.3883399209486166, "step": 131, "train_accuracy_first_token_arguments": 0.9166666666666666, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 20 }, { "epoch": 0.3883399209486166, "step": 131, "train_accuracy_first_token_arguments": 0.7580645161290323, "train_accuracy_first_token_arguments_total": 62, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 0.3333333333333333, "train_accuracy_recipient_token_analyze_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_log": 1.0, "train_accuracy_recipient_token_log_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.3883399209486166, "step": 131, "train_accuracy_first_token_arguments": 0.7368421052631579, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 1, "train_accuracy_recipient_token_recommend": 0.5, "train_accuracy_recipient_token_recommend_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.391304347826087, "grad_norm": 1.7888078327053374, "learning_rate": 3.484461092337434e-06, "loss": 0.2452, "step": 132 }, { "epoch": 0.391304347826087, "step": 132, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_trans": 1.0, "train_accuracy_recipient_token_trans_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.391304347826087, "step": 132, "train_accuracy_first_token_arguments": 0.5897435897435898, "train_accuracy_first_token_arguments_total": 39, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 0.6666666666666666, "train_accuracy_recipient_token_analyze_total": 3, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_histor": 1.0, "train_accuracy_recipient_token_histor_total": 3, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.391304347826087, "step": 132, "train_accuracy_first_token_arguments": 0.8947368421052632, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.8571428571428571, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.7777777777777778, "train_accuracy_recipient_token_get_total": 9, "train_accuracy_recipient_token_order": 0.0, "train_accuracy_recipient_token_order_total": 1, "train_total_number_recipient_token": 21 }, { "epoch": 0.3942687747035573, "grad_norm": 2.3453355040982307, "learning_rate": 3.462270318133136e-06, "loss": 0.273, "step": 133 }, { "epoch": 0.3942687747035573, "step": 133, "train_accuracy_first_token_arguments": 0.75, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_identify": 0.5, "train_accuracy_recipient_token_identify_total": 2, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.3942687747035573, "step": 133, "train_accuracy_first_token_arguments": 0.7894736842105263, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_add": 1.0, "train_accuracy_recipient_token_add_total": 4, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_construct": 1.0, "train_accuracy_recipient_token_construct_total": 1, "train_accuracy_recipient_token_remove": 1.0, "train_accuracy_recipient_token_remove_total": 1, "train_accuracy_recipient_token_select": 1.0, "train_accuracy_recipient_token_select_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.3942687747035573, "step": 133, "train_accuracy_first_token_arguments": 0.8333333333333334, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.39723320158102765, "grad_norm": 2.500149372884034, "learning_rate": 3.4399901808189327e-06, "loss": 0.2709, "step": 134 }, { "epoch": 0.39723320158102765, "step": 134, "train_accuracy_first_token_arguments": 0.5555555555555556, "train_accuracy_first_token_arguments_total": 9, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.39723320158102765, "step": 134, "train_accuracy_first_token_arguments": 0.8181818181818182, "train_accuracy_first_token_arguments_total": 11, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compose": 1.0, "train_accuracy_recipient_token_compose_total": 1, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 1, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.39723320158102765, "step": 134, "train_accuracy_first_token_arguments": 0.75, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.40019762845849804, "grad_norm": 2.056275874523913, "learning_rate": 3.4176227494832305e-06, "loss": 0.3526, "step": 135 }, { "epoch": 0.40019762845849804, "step": 135, "train_accuracy_first_token_arguments": 0.9230769230769231, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.85, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 0.7777777777777778, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 20 }, { "epoch": 0.40019762845849804, "step": 135, "train_accuracy_first_token_arguments": 0.8888888888888888, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 20 }, { "epoch": 0.40019762845849804, "step": 135, "train_accuracy_first_token_arguments": 0.9090909090909091, "train_accuracy_first_token_arguments_total": 11, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_convert": 1.0, "train_accuracy_recipient_token_convert_total": 1, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 17 }, { "epoch": 0.4031620553359684, "grad_norm": 1.9432341730217852, "learning_rate": 3.3951701013211665e-06, "loss": 0.214, "step": 136 }, { "epoch": 0.4031620553359684, "step": 136, "train_accuracy_first_token_arguments": 0.8, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_locate": 1.0, "train_accuracy_recipient_token_locate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.4031620553359684, "step": 136, "train_accuracy_first_token_arguments": 0.9130434782608695, "train_accuracy_first_token_arguments_total": 23, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_invest": 1.0, "train_accuracy_recipient_token_invest_total": 1, "train_accuracy_recipient_token_tractor": 1.0, "train_accuracy_recipient_token_tractor_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.4031620553359684, "step": 136, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 28, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9166666666666666, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_detect": 1.0, "train_accuracy_recipient_token_detect_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 20 }, { "epoch": 0.4061264822134387, "grad_norm": 1.6348659932120928, "learning_rate": 3.3726343214417023e-06, "loss": 0.1342, "step": 137 }, { "epoch": 0.4061264822134387, "step": 137, "train_accuracy_first_token_arguments": 0.8333333333333334, "train_accuracy_first_token_arguments_total": 24, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_pay": 0.5, "train_accuracy_recipient_token_pay_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.4061264822134387, "step": 137, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_make": 1.0, "train_accuracy_recipient_token_make_total": 4, "train_total_number_recipient_token": 17 }, { "epoch": 0.4061264822134387, "step": 137, "train_accuracy_first_token_arguments": 0.9230769230769231, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { "epoch": 0.4090909090909091, "grad_norm": 2.0536484514852344, "learning_rate": 3.3500175026739916e-06, "loss": 0.279, "step": 138 }, { "epoch": 0.4090909090909091, "step": 138, "train_accuracy_first_token_arguments": 0.7142857142857143, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_bird": 1.0, "train_accuracy_recipient_token_bird_total": 4, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_in": 1.0, "train_accuracy_recipient_token_in_total": 3, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.4090909090909091, "step": 138, "train_accuracy_first_token_arguments": 0.8048780487804879, "train_accuracy_first_token_arguments_total": 41, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 18 }, { "epoch": 0.4090909090909091, "step": 138, "train_accuracy_first_token_arguments": 0.9411764705882353, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 17 }, { "epoch": 0.41205533596837945, "grad_norm": 2.2289267853293144, "learning_rate": 3.327321745373021e-06, "loss": 0.3159, "step": 139 }, { "epoch": 0.41205533596837945, "step": 139, "train_accuracy_first_token_arguments": 0.9333333333333333, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_pest": 1.0, "train_accuracy_recipient_token_pest_total": 1, "train_accuracy_recipient_token_sun": 1.0, "train_accuracy_recipient_token_sun_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.41205533596837945, "step": 139, "train_accuracy_first_token_arguments": 0.8888888888888888, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_evaluate": 1.0, "train_accuracy_recipient_token_evaluate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_provide": 1.0, "train_accuracy_recipient_token_provide_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.41205533596837945, "step": 139, "train_accuracy_first_token_arguments": 0.8461538461538461, "train_accuracy_first_token_arguments_total": 26, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 20 }, { "epoch": 0.4150197628458498, "grad_norm": 2.1319285672179005, "learning_rate": 3.304549157224558e-06, "loss": 0.3213, "step": 140 }, { "epoch": 0.4150197628458498, "step": 140, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 9, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.4150197628458498, "step": 140, "train_accuracy_first_token_arguments": 0.8333333333333334, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 20 }, { "epoch": 0.4150197628458498, "step": 140, "train_accuracy_first_token_arguments": 0.8, "train_accuracy_first_token_arguments_total": 60, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_explain": 0.5, "train_accuracy_recipient_token_explain_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_invest": 1.0, "train_accuracy_recipient_token_invest_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.41798418972332013, "grad_norm": 2.0722211722052575, "learning_rate": 3.2817018530494164e-06, "loss": 0.2898, "step": 141 }, { "epoch": 0.41798418972332013, "step": 141, "train_accuracy_first_token_arguments": 0.7368421052631579, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_locate": 1.0, "train_accuracy_recipient_token_locate_total": 1, "train_accuracy_recipient_token_p": 1.0, "train_accuracy_recipient_token_p_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.41798418972332013, "step": 141, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.9523809523809523, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 2, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 21 }, { "epoch": 0.41798418972332013, "step": 141, "train_accuracy_first_token_arguments": 0.8656716417910447, "train_accuracy_first_token_arguments_total": 67, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 3, "train_accuracy_recipient_token_verify": 1.0, "train_accuracy_recipient_token_verify_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.4209486166007905, "grad_norm": 2.1191079314907766, "learning_rate": 3.2587819546070596e-06, "loss": 0.255, "step": 142 }, { "epoch": 0.4209486166007905, "step": 142, "train_accuracy_first_token_arguments": 0.7692307692307693, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.4209486166007905, "step": 142, "train_accuracy_first_token_arguments": 0.8333333333333334, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_l": 1.0, "train_accuracy_recipient_token_l_total": 3, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.4209486166007905, "step": 142, "train_accuracy_first_token_arguments": 0.9, "train_accuracy_first_token_arguments_total": 10, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_event": 0.6666666666666666, "train_accuracy_recipient_token_event_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.42391304347826086, "grad_norm": 2.5497967401591204, "learning_rate": 3.2357915903985605e-06, "loss": 0.3137, "step": 143 }, { "epoch": 0.42391304347826086, "step": 143, "train_accuracy_first_token_arguments": 0.95, "train_accuracy_first_token_arguments_total": 40, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_clean": 1.0, "train_accuracy_recipient_token_clean_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_paint": 1.0, "train_accuracy_recipient_token_paint_total": 2, "train_accuracy_recipient_token_remove": 1.0, "train_accuracy_recipient_token_remove_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.42391304347826086, "step": 143, "train_accuracy_first_token_arguments": 0.7619047619047619, "train_accuracy_first_token_arguments_total": 21, "train_accuracy_recipient_token": 0.8, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 2, "train_accuracy_recipient_token_device": 1.0, "train_accuracy_recipient_token_device_total": 3, "train_accuracy_recipient_token_extract": 0.0, "train_accuracy_recipient_token_extract_total": 1, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_total_number_recipient_token": 20 }, { "epoch": 0.42391304347826086, "step": 143, "train_accuracy_first_token_arguments": 0.7924528301886793, "train_accuracy_first_token_arguments_total": 53, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.4268774703557312, "grad_norm": 1.9145256420639876, "learning_rate": 3.2127328954689307e-06, "loss": 0.192, "step": 144 }, { "epoch": 0.4268774703557312, "step": 144, "train_accuracy_first_token_arguments": 0.7391304347826086, "train_accuracy_first_token_arguments_total": 46, "train_accuracy_recipient_token": 0.7894736842105263, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_accuracy_recipient_token_translate": 0.75, "train_accuracy_recipient_token_translate_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.4268774703557312, "step": 144, "train_accuracy_first_token_arguments": 0.875, "train_accuracy_first_token_arguments_total": 104, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_optimize": 1.0, "train_accuracy_recipient_token_optimize_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.4268774703557312, "step": 144, "train_accuracy_first_token_arguments": 0.9090909090909091, "train_accuracy_first_token_arguments_total": 22, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.4298418972332016, "grad_norm": 1.6421584423874946, "learning_rate": 3.1896080112088477e-06, "loss": 0.2441, "step": 145 }, { "epoch": 0.4298418972332016, "step": 145, "train_accuracy_first_token_arguments": 0.8846153846153846, "train_accuracy_first_token_arguments_total": 26, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_deploy": 1.0, "train_accuracy_recipient_token_deploy_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 17 }, { "epoch": 0.4298418972332016, "step": 145, "train_accuracy_first_token_arguments": 0.8947368421052632, "train_accuracy_first_token_arguments_total": 38, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_car": 1.0, "train_accuracy_recipient_token_car_total": 1, "train_accuracy_recipient_token_evaluate": 1.0, "train_accuracy_recipient_token_evaluate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.4298418972332016, "step": 145, "train_accuracy_first_token_arguments": 0.8235294117647058, "train_accuracy_first_token_arguments_total": 34, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_accuracy_recipient_token_schedule": 0.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.43280632411067194, "grad_norm": 1.9573866581692538, "learning_rate": 3.166419085155793e-06, "loss": 0.2593, "step": 146 }, { "epoch": 0.43280632411067194, "step": 146, "train_accuracy_first_token_arguments": 0.7857142857142857, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 1, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 1, "train_accuracy_recipient_token_stat": 1.0, "train_accuracy_recipient_token_stat_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.43280632411067194, "step": 146, "train_accuracy_first_token_arguments": 0.84, "train_accuracy_first_token_arguments_total": 25, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.9, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 20 }, { "epoch": 0.43280632411067194, "step": 146, "train_accuracy_first_token_arguments": 0.7142857142857143, "train_accuracy_first_token_arguments_total": 7, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.4357707509881423, "grad_norm": 2.277891831921691, "learning_rate": 3.143168270794612e-06, "loss": 0.4509, "step": 147 }, { "epoch": 0.4357707509881423, "step": 147, "train_accuracy_first_token_arguments": 0.8888888888888888, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_explain": 1.0, "train_accuracy_recipient_token_explain_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_traffic": 1.0, "train_accuracy_recipient_token_traffic_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.4357707509881423, "step": 147, "train_accuracy_first_token_arguments": 0.85, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 0.9047619047619048, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 0.6666666666666666, "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 21 }, { "epoch": 0.4357707509881423, "step": 147, "train_accuracy_first_token_arguments": 0.9125, "train_accuracy_first_token_arguments_total": 80, "train_accuracy_recipient_token": 0.7777777777777778, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_arr": 1.0, "train_accuracy_recipient_token_arr_total": 1, "train_accuracy_recipient_token_book": 0.5, "train_accuracy_recipient_token_book_total": 2, "train_accuracy_recipient_token_hire": 1.0, "train_accuracy_recipient_token_hire_total": 1, "train_accuracy_recipient_token_plan": 1.0, "train_accuracy_recipient_token_plan_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.43873517786561267, "grad_norm": 2.085474389214059, "learning_rate": 3.119857727357527e-06, "loss": 0.2157, "step": 148 }, { "epoch": 0.43873517786561267, "step": 148, "train_accuracy_first_token_arguments": 0.782608695652174, "train_accuracy_first_token_arguments_total": 23, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 0.5, "train_accuracy_recipient_token_calculate_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.43873517786561267, "step": 148, "train_accuracy_first_token_arguments": 0.75, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_stamp": 1.0, "train_accuracy_recipient_token_stamp_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.43873517786561267, "step": 148, "train_accuracy_first_token_arguments": 0.9090909090909091, "train_accuracy_first_token_arguments_total": 11, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_display": 1.0, "train_accuracy_recipient_token_display_total": 1, "train_accuracy_recipient_token_find": 0.75, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.441699604743083, "grad_norm": 1.8828482742789217, "learning_rate": 3.0964896196236217e-06, "loss": 0.2592, "step": 149 }, { "epoch": 0.441699604743083, "step": 149, "train_accuracy_first_token_arguments": 0.8095238095238095, "train_accuracy_first_token_arguments_total": 21, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.441699604743083, "step": 149, "train_accuracy_first_token_arguments": 0.6666666666666666, "train_accuracy_first_token_arguments_total": 21, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_suggest": 1.0, "train_accuracy_recipient_token_suggest_total": 2, "train_total_number_recipient_token": 20 }, { "epoch": 0.441699604743083, "step": 149, "train_accuracy_first_token_arguments": 0.9285714285714286, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.44466403162055335, "grad_norm": 1.6792776522615755, "learning_rate": 3.0730661177177957e-06, "loss": 0.2082, "step": 150 }, { "epoch": 0.44466403162055335, "step": 150, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 10, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_discover": 1.0, "train_accuracy_recipient_token_discover_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 18 }, { "epoch": 0.44466403162055335, "step": 150, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.44466403162055335, "step": 150, "train_accuracy_first_token_arguments": 0.9473684210526315, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_stamp": 1.0, "train_accuracy_recipient_token_stamp_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.44762845849802374, "grad_norm": 2.0438916010331187, "learning_rate": 3.0495893969092395e-06, "loss": 0.341, "step": 151 }, { "epoch": 0.44762845849802374, "step": 151, "train_accuracy_first_token_arguments": 0.8888888888888888, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { "epoch": 0.44762845849802374, "step": 151, "train_accuracy_first_token_arguments": 0.9166666666666666, "train_accuracy_first_token_arguments_total": 24, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_locate": 1.0, "train_accuracy_recipient_token_locate_total": 1, "train_accuracy_recipient_token_provide": 1.0, "train_accuracy_recipient_token_provide_total": 1, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.44762845849802374, "step": 151, "train_accuracy_first_token_arguments": 0.8333333333333334, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.8125, "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.4505928853754941, "grad_norm": 1.870929622393286, "learning_rate": 3.0260616374094208e-06, "loss": 0.2028, "step": 152 }, { "epoch": 0.4505928853754941, "step": 152, "train_accuracy_first_token_arguments": 0.7894736842105263, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_configure": 0.0, "train_accuracy_recipient_token_configure_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_track": 0.0, "train_accuracy_recipient_token_track_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.4505928853754941, "step": 152, "train_accuracy_first_token_arguments": 0.8636363636363636, "train_accuracy_first_token_arguments_total": 44, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.4505928853754941, "step": 152, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_download": 1.0, "train_accuracy_recipient_token_download_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_rec": 1.0, "train_accuracy_recipient_token_rec_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.4535573122529644, "grad_norm": 2.1374043090533816, "learning_rate": 3.0024850241696128e-06, "loss": 0.1932, "step": 153 }, { "epoch": 0.4535573122529644, "step": 153, "train_accuracy_first_token_arguments": 0.7916666666666666, "train_accuracy_first_token_arguments_total": 24, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.4535573122529644, "step": 153, "train_accuracy_first_token_arguments": 0.8, "train_accuracy_first_token_arguments_total": 35, "train_accuracy_recipient_token": 0.7894736842105263, "train_accuracy_recipient_token_all": 0.8333333333333334, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_describe": 0.0, "train_accuracy_recipient_token_describe_total": 1, "train_accuracy_recipient_token_find": 0.75, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_organ": 1.0, "train_accuracy_recipient_token_organ_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.4535573122529644, "step": 153, "train_accuracy_first_token_arguments": 0.9555555555555556, "train_accuracy_first_token_arguments_total": 90, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 20 }, { "epoch": 0.45652173913043476, "grad_norm": 1.934815353317823, "learning_rate": 2.9788617466779884e-06, "loss": 0.2336, "step": 154 }, { "epoch": 0.45652173913043476, "step": 154, "train_accuracy_first_token_arguments": 0.8205128205128205, "train_accuracy_first_token_arguments_total": 39, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_f": 1.0, "train_accuracy_recipient_token_f_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.45652173913043476, "step": 154, "train_accuracy_first_token_arguments": 0.8095238095238095, "train_accuracy_first_token_arguments_total": 21, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_multip": 0.5, "train_accuracy_recipient_token_multip_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.45652173913043476, "step": 154, "train_accuracy_first_token_arguments": 0.9130434782608695, "train_accuracy_first_token_arguments_total": 23, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.875, "train_accuracy_recipient_token_all_total": 8, "train_accuracy_recipient_token_check": 0.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_histor": 1.0, "train_accuracy_recipient_token_histor_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.45948616600790515, "grad_norm": 2.3253564073005224, "learning_rate": 2.9551939987562866e-06, "loss": 0.3042, "step": 155 }, { "epoch": 0.45948616600790515, "step": 155, "train_accuracy_first_token_arguments": 0.95, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9166666666666666, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.45948616600790515, "step": 155, "train_accuracy_first_token_arguments": 0.84, "train_accuracy_first_token_arguments_total": 75, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_estimate": 0.5, "train_accuracy_recipient_token_estimate_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.45948616600790515, "step": 155, "train_accuracy_first_token_arguments": 0.8, "train_accuracy_first_token_arguments_total": 10, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_translate": 0.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.4624505928853755, "grad_norm": 2.2711295887254184, "learning_rate": 2.93148397835608e-06, "loss": 0.2997, "step": 156 }, { "epoch": 0.4624505928853755, "step": 156, "train_accuracy_first_token_arguments": 0.5, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.4624505928853755, "step": 156, "train_accuracy_first_token_arguments": 0.8333333333333334, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 19 }, { "epoch": 0.4624505928853755, "step": 156, "train_accuracy_first_token_arguments": 0.8333333333333334, "train_accuracy_first_token_arguments_total": 30, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 20 }, { "epoch": 0.4654150197628458, "grad_norm": 2.0895540217432282, "learning_rate": 2.907733887354657e-06, "loss": 0.2419, "step": 157 }, { "epoch": 0.4654150197628458, "step": 157, "train_accuracy_first_token_arguments": 0.7837837837837838, "train_accuracy_first_token_arguments_total": 37, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_sum": 0.0, "train_accuracy_recipient_token_sum_total": 1, "train_accuracy_recipient_token_translate": 0.5, "train_accuracy_recipient_token_translate_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.4654150197628458, "step": 157, "train_accuracy_first_token_arguments": 0.7804878048780488, "train_accuracy_first_token_arguments_total": 41, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_plan": 1.0, "train_accuracy_recipient_token_plan_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.4654150197628458, "step": 157, "train_accuracy_first_token_arguments": 0.6896551724137931, "train_accuracy_first_token_arguments_total": 58, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.3333333333333333, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_prior": 1.0, "train_accuracy_recipient_token_prior_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.4683794466403162, "grad_norm": 2.203763069427427, "learning_rate": 2.8839459313505407e-06, "loss": 0.2426, "step": 158 }, { "epoch": 0.4683794466403162, "step": 158, "train_accuracy_first_token_arguments": 0.9285714285714286, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_recommend": 0.6666666666666666, "train_accuracy_recipient_token_recommend_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.4683794466403162, "step": 158, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_shop": 1.0, "train_accuracy_recipient_token_shop_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.4683794466403162, "step": 158, "train_accuracy_first_token_arguments": 0.9411764705882353, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.47134387351778656, "grad_norm": 1.9347965319589504, "learning_rate": 2.8601223194586613e-06, "loss": 0.2107, "step": 159 }, { "epoch": 0.47134387351778656, "step": 159, "train_accuracy_first_token_arguments": 0.7777777777777778, "train_accuracy_first_token_arguments_total": 27, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.47134387351778656, "step": 159, "train_accuracy_first_token_arguments": 0.9375, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_invest": 1.0, "train_accuracy_recipient_token_invest_total": 1, "train_accuracy_recipient_token_vehicle": 1.0, "train_accuracy_recipient_token_vehicle_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.47134387351778656, "step": 159, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 19 }, { "epoch": 0.4743083003952569, "grad_norm": 1.9897698233644936, "learning_rate": 2.8362652641052024e-06, "loss": 0.1806, "step": 160 }, { "epoch": 0.4743083003952569, "step": 160, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.4743083003952569, "step": 160, "train_accuracy_first_token_arguments": 0.875, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 3, "train_accuracy_recipient_token_artist": 0.0, "train_accuracy_recipient_token_artist_total": 1, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 1, "train_accuracy_recipient_token_water": 1.0, "train_accuracy_recipient_token_water_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.4743083003952569, "step": 160, "train_accuracy_first_token_arguments": 0.9130434782608695, "train_accuracy_first_token_arguments_total": 23, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_schedule": 0.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.4772727272727273, "grad_norm": 1.869456628968754, "learning_rate": 2.8123769808221407e-06, "loss": 0.1735, "step": 161 }, { "epoch": 0.4772727272727273, "step": 161, "train_accuracy_first_token_arguments": 0.6521739130434783, "train_accuracy_first_token_arguments_total": 23, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.4772727272727273, "step": 161, "train_accuracy_first_token_arguments": 0.9166666666666666, "train_accuracy_first_token_arguments_total": 36, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.4772727272727273, "step": 161, "train_accuracy_first_token_arguments": 0.8333333333333334, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_total_number_recipient_token": 16 }, { "epoch": 0.48023715415019763, "grad_norm": 2.502634515773549, "learning_rate": 2.788459688041495e-06, "loss": 0.2973, "step": 162 }, { "epoch": 0.48023715415019763, "step": 162, "train_accuracy_first_token_arguments": 0.9375, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 0.8571428571428571, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_detect": 1.0, "train_accuracy_recipient_token_detect_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_translate": 0.5, "train_accuracy_recipient_token_translate_total": 2, "train_total_number_recipient_token": 21 }, { "epoch": 0.48023715415019763, "step": 162, "train_accuracy_first_token_arguments": 0.9473684210526315, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 0.5, "train_accuracy_recipient_token_calculate_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.48023715415019763, "step": 162, "train_accuracy_first_token_arguments": 0.9285714285714286, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 5, "train_total_number_recipient_token": 18 }, { "epoch": 0.48320158102766797, "grad_norm": 2.341420807222338, "learning_rate": 2.7645156068893075e-06, "loss": 0.2014, "step": 163 }, { "epoch": 0.48320158102766797, "step": 163, "train_accuracy_first_token_arguments": 0.8333333333333334, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 3, "train_total_number_recipient_token": 20 }, { "epoch": 0.48320158102766797, "step": 163, "train_accuracy_first_token_arguments": 0.4782608695652174, "train_accuracy_first_token_arguments_total": 46, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_color": 1.0, "train_accuracy_recipient_token_color_total": 1, "train_accuracy_recipient_token_draw": 1.0, "train_accuracy_recipient_token_draw_total": 1, "train_accuracy_recipient_token_object": 1.0, "train_accuracy_recipient_token_object_total": 1, "train_accuracy_recipient_token_scene": 1.0, "train_accuracy_recipient_token_scene_total": 2, "train_total_number_recipient_token": 20 }, { "epoch": 0.48320158102766797, "step": 163, "train_accuracy_first_token_arguments": 0.9090909090909091, "train_accuracy_first_token_arguments_total": 11, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.48616600790513836, "grad_norm": 1.6883872990934619, "learning_rate": 2.7405469609793746e-06, "loss": 0.2554, "step": 164 }, { "epoch": 0.48616600790513836, "step": 164, "train_accuracy_first_token_arguments": 0.7407407407407407, "train_accuracy_first_token_arguments_total": 81, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_inspect": 1.0, "train_accuracy_recipient_token_inspect_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.48616600790513836, "step": 164, "train_accuracy_first_token_arguments": 0.8947368421052632, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_review": 1.0, "train_accuracy_recipient_token_review_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.48616600790513836, "step": 164, "train_accuracy_first_token_arguments": 0.7142857142857143, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_learn": 1.0, "train_accuracy_recipient_token_learn_total": 1, "train_accuracy_recipient_token_research": 1.0, "train_accuracy_recipient_token_research_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.4891304347826087, "grad_norm": 1.7224600166430244, "learning_rate": 2.716555976206748e-06, "loss": 0.2725, "step": 165 }, { "epoch": 0.4891304347826087, "step": 165, "train_accuracy_first_token_arguments": 0.6, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 0.9333333333333333, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_stat": 1.0, "train_accuracy_recipient_token_stat_total": 1, "train_accuracy_recipient_token_suggest": 1.0, "train_accuracy_recipient_token_suggest_total": 1, "train_total_number_recipient_token": 15 }, { "epoch": 0.4891304347826087, "step": 165, "train_accuracy_first_token_arguments": 0.9166666666666666, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_st": 1.0, "train_accuracy_recipient_token_st_total": 3, "train_total_number_recipient_token": 20 }, { "epoch": 0.4891304347826087, "step": 165, "train_accuracy_first_token_arguments": 0.6029411764705882, "train_accuracy_first_token_arguments_total": 68, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 2, "train_accuracy_recipient_token_flag": 1.0, "train_accuracy_recipient_token_flag_total": 1, "train_accuracy_recipient_token_pl": 1.0, "train_accuracy_recipient_token_pl_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.49209486166007904, "grad_norm": 2.488668771406232, "learning_rate": 2.6925448805410197e-06, "loss": 0.4772, "step": 166 }, { "epoch": 0.49209486166007904, "step": 166, "train_accuracy_first_token_arguments": 0.6153846153846154, "train_accuracy_first_token_arguments_total": 26, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.49209486166007904, "step": 166, "train_accuracy_first_token_arguments": 0.96875, "train_accuracy_first_token_arguments_total": 32, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 2, "train_accuracy_recipient_token_design": 1.0, "train_accuracy_recipient_token_design_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.49209486166007904, "step": 166, "train_accuracy_first_token_arguments": 0.8333333333333334, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.4950592885375494, "grad_norm": 2.6791883381737107, "learning_rate": 2.6685159038194202e-06, "loss": 0.2803, "step": 167 }, { "epoch": 0.4950592885375494, "step": 167, "train_accuracy_first_token_arguments": 0.8333333333333334, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 18 }, { "epoch": 0.4950592885375494, "step": 167, "train_accuracy_first_token_arguments": 0.7083333333333334, "train_accuracy_first_token_arguments_total": 24, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_compile": 1.0, "train_accuracy_recipient_token_compile_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.4950592885375494, "step": 167, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 9, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.4980237154150198, "grad_norm": 2.068138976829196, "learning_rate": 2.6444712775397397e-06, "loss": 0.3013, "step": 168 }, { "epoch": 0.4980237154150198, "step": 168, "train_accuracy_first_token_arguments": 0.6559139784946236, "train_accuracy_first_token_arguments_total": 93, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.4980237154150198, "step": 168, "train_accuracy_first_token_arguments": 0.9782608695652174, "train_accuracy_first_token_arguments_total": 92, "train_accuracy_recipient_token": 0.75, "train_accuracy_recipient_token_all": 0.6666666666666666, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 12 }, { "epoch": 0.4980237154150198, "step": 168, "train_accuracy_first_token_arguments": 0.75, "train_accuracy_first_token_arguments_total": 8, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.5009881422924901, "grad_norm": 2.097242382871384, "learning_rate": 2.6204132346530936e-06, "loss": 0.3079, "step": 169 }, { "epoch": 0.5009881422924901, "step": 169, "train_accuracy_first_token_arguments": 0.875, "train_accuracy_first_token_arguments_total": 32, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.5009881422924901, "step": 169, "train_accuracy_first_token_arguments": 0.6764705882352942, "train_accuracy_first_token_arguments_total": 34, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_camera": 1.0, "train_accuracy_recipient_token_camera_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_invest": 1.0, "train_accuracy_recipient_token_invest_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.5009881422924901, "step": 169, "train_accuracy_first_token_arguments": 0.7692307692307693, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.5039525691699605, "grad_norm": 1.961726304062012, "learning_rate": 2.5963440093565567e-06, "loss": 0.3133, "step": 170 }, { "epoch": 0.5039525691699605, "step": 170, "train_accuracy_first_token_arguments": 0.8461538461538461, "train_accuracy_first_token_arguments_total": 26, "train_accuracy_recipient_token": 0.6666666666666666, "train_accuracy_recipient_token_all": 0.6, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 0.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_multiply": 0.75, "train_accuracy_recipient_token_multiply_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.5039525691699605, "step": 170, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { "epoch": 0.5039525691699605, "step": 170, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 11, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_extract": 1.0, "train_accuracy_recipient_token_extract_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_recogn": 1.0, "train_accuracy_recipient_token_recogn_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.5069169960474308, "grad_norm": 2.0506442694884104, "learning_rate": 2.572265836885682e-06, "loss": 0.2265, "step": 171 }, { "epoch": 0.5069169960474308, "step": 171, "train_accuracy_first_token_arguments": 0.9, "train_accuracy_first_token_arguments_total": 30, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 21 }, { "epoch": 0.5069169960474308, "step": 171, "train_accuracy_first_token_arguments": 0.8125, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.5069169960474308, "step": 171, "train_accuracy_first_token_arguments": 0.8888888888888888, "train_accuracy_first_token_arguments_total": 9, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 5, "train_total_number_recipient_token": 18 }, { "epoch": 0.5098814229249012, "grad_norm": 2.4101209365627443, "learning_rate": 2.548180953306918e-06, "loss": 0.2901, "step": 172 }, { "epoch": 0.5098814229249012, "step": 172, "train_accuracy_first_token_arguments": 0.9, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 0.8666666666666667, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 15 }, { "epoch": 0.5098814229249012, "step": 172, "train_accuracy_first_token_arguments": 0.7678571428571429, "train_accuracy_first_token_arguments_total": 56, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_histor": 0.5, "train_accuracy_recipient_token_histor_total": 2, "train_accuracy_recipient_token_organ": 1.0, "train_accuracy_recipient_token_organ_total": 1, "train_accuracy_recipient_token_public": 1.0, "train_accuracy_recipient_token_public_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.5098814229249012, "step": 172, "train_accuracy_first_token_arguments": 0.5714285714285714, "train_accuracy_first_token_arguments_total": 21, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 1, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.5128458498023716, "grad_norm": 2.403125205150021, "learning_rate": 2.524091595309952e-06, "loss": 0.2281, "step": 173 }, { "epoch": 0.5128458498023716, "step": 173, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 11, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_plant": 1.0, "train_accuracy_recipient_token_plant_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.5128458498023716, "step": 173, "train_accuracy_first_token_arguments": 0.9285714285714286, "train_accuracy_first_token_arguments_total": 28, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_evaluate": 1.0, "train_accuracy_recipient_token_evaluate_total": 1, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_plan": 1.0, "train_accuracy_recipient_token_plan_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.5128458498023716, "step": 173, "train_accuracy_first_token_arguments": 0.8947368421052632, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_update": 1.0, "train_accuracy_recipient_token_update_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.5158102766798419, "grad_norm": 1.7135004510273728, "learning_rate": 2.5e-06, "loss": 0.1725, "step": 174 }, { "epoch": 0.5158102766798419, "step": 174, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 7, "train_accuracy_recipient_token": 0.9333333333333333, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_int": 1.0, "train_accuracy_recipient_token_int_total": 1, "train_total_number_recipient_token": 15 }, { "epoch": 0.5158102766798419, "step": 174, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.5158102766798419, "step": 174, "train_accuracy_first_token_arguments": 0.75, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.5187747035573123, "grad_norm": 2.193026807831911, "learning_rate": 2.475908404690049e-06, "loss": 0.4577, "step": 175 }, { "epoch": 0.5187747035573123, "step": 175, "train_accuracy_first_token_arguments": 0.9090909090909091, "train_accuracy_first_token_arguments_total": 22, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_ca": 1.0, "train_accuracy_recipient_token_ca_total": 4, "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 20 }, { "epoch": 0.5187747035573123, "step": 175, "train_accuracy_first_token_arguments": 0.7058823529411765, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 2, "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 20 }, { "epoch": 0.5187747035573123, "step": 175, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 24, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 16 }, { "epoch": 0.5217391304347826, "grad_norm": 2.046983156954581, "learning_rate": 2.4518190466930837e-06, "loss": 0.4614, "step": 176 }, { "epoch": 0.5217391304347826, "step": 176, "train_accuracy_first_token_arguments": 0.8235294117647058, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 0.8571428571428571, "train_accuracy_recipient_token_calculate_total": 7, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.5217391304347826, "step": 176, "train_accuracy_first_token_arguments": 0.8666666666666667, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 1, "train_accuracy_recipient_token_draw": 1.0, "train_accuracy_recipient_token_draw_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 17 }, { "epoch": 0.5217391304347826, "step": 176, "train_accuracy_first_token_arguments": 0.9444444444444444, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_phone": 1.0, "train_accuracy_recipient_token_phone_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.5247035573122529, "grad_norm": 2.7763963997279046, "learning_rate": 2.427734163114319e-06, "loss": 0.3222, "step": 177 }, { "epoch": 0.5247035573122529, "step": 177, "train_accuracy_first_token_arguments": 0.9473684210526315, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.8333333333333334, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_generate": 0.5, "train_accuracy_recipient_token_generate_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.5247035573122529, "step": 177, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 0.7647058823529411, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 0.5, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_sum": 1.0, "train_accuracy_recipient_token_sum_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.5247035573122529, "step": 177, "train_accuracy_first_token_arguments": 0.8823529411764706, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compress": 1.0, "train_accuracy_recipient_token_compress_total": 1, "train_accuracy_recipient_token_crop": 1.0, "train_accuracy_recipient_token_crop_total": 1, "train_accuracy_recipient_token_d": 1.0, "train_accuracy_recipient_token_d_total": 2, "train_accuracy_recipient_token_enh": 0.0, "train_accuracy_recipient_token_enh_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.5276679841897233, "grad_norm": 2.245421899066183, "learning_rate": 2.403655990643444e-06, "loss": 0.3072, "step": 178 }, { "epoch": 0.5276679841897233, "step": 178, "train_accuracy_first_token_arguments": 0.875, "train_accuracy_first_token_arguments_total": 24, "train_accuracy_recipient_token": 0.7894736842105263, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_customize": 0.5, "train_accuracy_recipient_token_customize_total": 4, "train_accuracy_recipient_token_get": 0.6, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.5276679841897233, "step": 178, "train_accuracy_first_token_arguments": 0.8333333333333334, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.7777777777777778, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 0.5714285714285714, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { "epoch": 0.5276679841897233, "step": 178, "train_accuracy_first_token_arguments": 0.8823529411764706, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.5306324110671937, "grad_norm": 2.405264876175068, "learning_rate": 2.3795867653469072e-06, "loss": 0.344, "step": 179 }, { "epoch": 0.5306324110671937, "step": 179, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 9, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 5, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 10 }, { "epoch": 0.5306324110671937, "step": 179, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 0.7142857142857143, "train_accuracy_recipient_token_get_total": 7, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.5306324110671937, "step": 179, "train_accuracy_first_token_arguments": 0.7954545454545454, "train_accuracy_first_token_arguments_total": 44, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_describe": 1.0, "train_accuracy_recipient_token_describe_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.5335968379446641, "grad_norm": 5.342437189948287, "learning_rate": 2.355528722460261e-06, "loss": 0.7124, "step": 180 }, { "epoch": 0.5335968379446641, "step": 180, "train_accuracy_first_token_arguments": 0.8666666666666667, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { "epoch": 0.5335968379446641, "step": 180, "train_accuracy_first_token_arguments": 0.8461538461538461, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_explain": 1.0, "train_accuracy_recipient_token_explain_total": 3, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.5335968379446641, "step": 180, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 22, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_fish": 1.0, "train_accuracy_recipient_token_fish_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_purchase": 1.0, "train_accuracy_recipient_token_purchase_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.5365612648221344, "grad_norm": 2.1777035417905934, "learning_rate": 2.3314840961805806e-06, "loss": 0.2801, "step": 181 }, { "epoch": 0.5365612648221344, "step": 181, "train_accuracy_first_token_arguments": 0.75, "train_accuracy_first_token_arguments_total": 36, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 20 }, { "epoch": 0.5365612648221344, "step": 181, "train_accuracy_first_token_arguments": 0.8, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.5365612648221344, "step": 181, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 8, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 2, "train_accuracy_recipient_token_extract": 1.0, "train_accuracy_recipient_token_extract_total": 1, "train_accuracy_recipient_token_monitor": 0.0, "train_accuracy_recipient_token_monitor_total": 1, "train_accuracy_recipient_token_recogn": 0.0, "train_accuracy_recipient_token_recogn_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.5395256916996047, "grad_norm": 1.9844826167678842, "learning_rate": 2.3074551194589816e-06, "loss": 0.3629, "step": 182 }, { "epoch": 0.5395256916996047, "step": 182, "train_accuracy_first_token_arguments": 0.9019607843137255, "train_accuracy_first_token_arguments_total": 51, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_notify": 0.0, "train_accuracy_recipient_token_notify_total": 1, "train_accuracy_recipient_token_set": 1.0, "train_accuracy_recipient_token_set_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.5395256916996047, "step": 182, "train_accuracy_first_token_arguments": 0.9090909090909091, "train_accuracy_first_token_arguments_total": 22, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_sponsor": 1.0, "train_accuracy_recipient_token_sponsor_total": 1, "train_accuracy_recipient_token_vol": 1.0, "train_accuracy_recipient_token_vol_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.5395256916996047, "step": 182, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 10, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_book": 0.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.5424901185770751, "grad_norm": 2.429551580616199, "learning_rate": 2.2834440237932537e-06, "loss": 0.3087, "step": 183 }, { "epoch": 0.5424901185770751, "step": 183, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_order": 0.0, "train_accuracy_recipient_token_order_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.5424901185770751, "step": 183, "train_accuracy_first_token_arguments": 0.9354838709677419, "train_accuracy_first_token_arguments_total": 31, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.5424901185770751, "step": 183, "train_accuracy_first_token_arguments": 0.875, "train_accuracy_first_token_arguments_total": 8, "train_accuracy_recipient_token": 0.8, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 15 }, { "epoch": 0.5454545454545454, "grad_norm": 2.062437887240769, "learning_rate": 2.259453039020626e-06, "loss": 0.2448, "step": 184 }, { "epoch": 0.5454545454545454, "step": 184, "train_accuracy_first_token_arguments": 0.6976744186046512, "train_accuracy_first_token_arguments_total": 43, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_log": 1.0, "train_accuracy_recipient_token_log_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.5454545454545454, "step": 184, "train_accuracy_first_token_arguments": 0.6153846153846154, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_interpret": 1.0, "train_accuracy_recipient_token_interpret_total": 1, "train_accuracy_recipient_token_stage": 1.0, "train_accuracy_recipient_token_stage_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.5454545454545454, "step": 184, "train_accuracy_first_token_arguments": 0.8846153846153846, "train_accuracy_first_token_arguments_total": 26, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.5484189723320159, "grad_norm": 1.7733764406317996, "learning_rate": 2.2354843931106933e-06, "loss": 0.2303, "step": 185 }, { "epoch": 0.5484189723320159, "step": 185, "train_accuracy_first_token_arguments": 0.8918918918918919, "train_accuracy_first_token_arguments_total": 37, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9166666666666666, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 3, "train_accuracy_recipient_token_estimate": 0.5, "train_accuracy_recipient_token_estimate_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.5484189723320159, "step": 185, "train_accuracy_first_token_arguments": 0.6, "train_accuracy_first_token_arguments_total": 55, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 0.5, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_update": 1.0, "train_accuracy_recipient_token_update_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.5484189723320159, "step": 185, "train_accuracy_first_token_arguments": 0.7352941176470589, "train_accuracy_first_token_arguments_total": 34, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.5513833992094862, "grad_norm": 1.7231295422156472, "learning_rate": 2.211540311958506e-06, "loss": 0.2471, "step": 186 }, { "epoch": 0.5513833992094862, "step": 186, "train_accuracy_first_token_arguments": 0.8571428571428571, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.5513833992094862, "step": 186, "train_accuracy_first_token_arguments": 0.9230769230769231, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.5513833992094862, "step": 186, "train_accuracy_first_token_arguments": 0.6712328767123288, "train_accuracy_first_token_arguments_total": 73, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_describe": 1.0, "train_accuracy_recipient_token_describe_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.5543478260869565, "grad_norm": 1.9589966620887675, "learning_rate": 2.1876230191778598e-06, "loss": 0.3246, "step": 187 }, { "epoch": 0.5543478260869565, "step": 187, "train_accuracy_first_token_arguments": 0.8260869565217391, "train_accuracy_first_token_arguments_total": 23, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.5543478260869565, "step": 187, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 32, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_convert": 1.0, "train_accuracy_recipient_token_convert_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 17 }, { "epoch": 0.5543478260869565, "step": 187, "train_accuracy_first_token_arguments": 0.75, "train_accuracy_first_token_arguments_total": 8, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { "epoch": 0.5573122529644269, "grad_norm": 2.2200460526334824, "learning_rate": 2.1637347358947984e-06, "loss": 0.216, "step": 188 }, { "epoch": 0.5573122529644269, "step": 188, "train_accuracy_first_token_arguments": 0.95, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_explain": 1.0, "train_accuracy_recipient_token_explain_total": 1, "train_accuracy_recipient_token_gesture": 1.0, "train_accuracy_recipient_token_gesture_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 20 }, { "epoch": 0.5573122529644269, "step": 188, "train_accuracy_first_token_arguments": 0.9473684210526315, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8461538461538461, "train_accuracy_recipient_token_all_total": 13, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 18 }, { "epoch": 0.5573122529644269, "step": 188, "train_accuracy_first_token_arguments": 0.75, "train_accuracy_first_token_arguments_total": 24, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_add": 1.0, "train_accuracy_recipient_token_add_total": 1, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 1, "train_accuracy_recipient_token_update": 1.0, "train_accuracy_recipient_token_update_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.5602766798418972, "grad_norm": 2.440285593927014, "learning_rate": 2.13987768054134e-06, "loss": 0.3342, "step": 189 }, { "epoch": 0.5602766798418972, "step": 189, "train_accuracy_first_token_arguments": 0.9130434782608695, "train_accuracy_first_token_arguments_total": 23, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_product": 0.75, "train_accuracy_recipient_token_product_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.5602766798418972, "step": 189, "train_accuracy_first_token_arguments": 0.9886363636363636, "train_accuracy_first_token_arguments_total": 88, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 4, "train_accuracy_recipient_token_hang": 1.0, "train_accuracy_recipient_token_hang_total": 1, "train_accuracy_recipient_token_re": 0.0, "train_accuracy_recipient_token_re_total": 1, "train_accuracy_recipient_token_remove": 1.0, "train_accuracy_recipient_token_remove_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.5602766798418972, "step": 189, "train_accuracy_first_token_arguments": 0.7391304347826086, "train_accuracy_first_token_arguments_total": 23, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.5632411067193676, "grad_norm": 2.2639594853684795, "learning_rate": 2.1160540686494597e-06, "loss": 0.2142, "step": 190 }, { "epoch": 0.5632411067193676, "step": 190, "train_accuracy_first_token_arguments": 0.7857142857142857, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.5632411067193676, "step": 190, "train_accuracy_first_token_arguments": 0.9166666666666666, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.5632411067193676, "step": 190, "train_accuracy_first_token_arguments": 0.7777777777777778, "train_accuracy_first_token_arguments_total": 36, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.5662055335968379, "grad_norm": 3.2075927432234157, "learning_rate": 2.0922661126453436e-06, "loss": 0.4145, "step": 191 }, { "epoch": 0.5662055335968379, "step": 191, "train_accuracy_first_token_arguments": 0.7166666666666667, "train_accuracy_first_token_arguments_total": 60, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compose": 1.0, "train_accuracy_recipient_token_compose_total": 1, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 2, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 1, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.5662055335968379, "step": 191, "train_accuracy_first_token_arguments": 0.6385542168674698, "train_accuracy_first_token_arguments_total": 83, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 2, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.5662055335968379, "step": 191, "train_accuracy_first_token_arguments": 0.7121212121212122, "train_accuracy_first_token_arguments_total": 66, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 2, "train_accuracy_recipient_token_draw": 1.0, "train_accuracy_recipient_token_draw_total": 4, "train_accuracy_recipient_token_suggest": 1.0, "train_accuracy_recipient_token_suggest_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.5691699604743083, "grad_norm": 1.9745905408889657, "learning_rate": 2.0685160216439205e-06, "loss": 0.2894, "step": 192 }, { "epoch": 0.5691699604743083, "step": 192, "train_accuracy_first_token_arguments": 0.7647058823529411, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 16 }, { "epoch": 0.5691699604743083, "step": 192, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 53, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 5, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.5691699604743083, "step": 192, "train_accuracy_first_token_arguments": 0.9545454545454546, "train_accuracy_first_token_arguments_total": 22, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8333333333333334, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 17 }, { "epoch": 0.5721343873517787, "grad_norm": 2.517506728021728, "learning_rate": 2.044806001243714e-06, "loss": 0.2701, "step": 193 }, { "epoch": 0.5721343873517787, "step": 193, "train_accuracy_first_token_arguments": 0.6296296296296297, "train_accuracy_first_token_arguments_total": 27, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_g": 1.0, "train_accuracy_recipient_token_g_total": 3, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.5721343873517787, "step": 193, "train_accuracy_first_token_arguments": 0.75, "train_accuracy_first_token_arguments_total": 8, "train_accuracy_recipient_token": 0.9333333333333333, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 1, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 15 }, { "epoch": 0.5721343873517787, "step": 193, "train_accuracy_first_token_arguments": 0.5471698113207547, "train_accuracy_first_token_arguments_total": 53, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_energy": 1.0, "train_accuracy_recipient_token_energy_total": 1, "train_accuracy_recipient_token_landscape": 1.0, "train_accuracy_recipient_token_landscape_total": 1, "train_accuracy_recipient_token_suggest": 1.0, "train_accuracy_recipient_token_suggest_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.575098814229249, "grad_norm": 2.477613864495461, "learning_rate": 2.021138253322012e-06, "loss": 0.3818, "step": 194 }, { "epoch": 0.575098814229249, "step": 194, "train_accuracy_first_token_arguments": 0.967741935483871, "train_accuracy_first_token_arguments_total": 31, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_accuracy_recipient_token_sort": 1.0, "train_accuracy_recipient_token_sort_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.575098814229249, "step": 194, "train_accuracy_first_token_arguments": 0.9705882352941176, "train_accuracy_first_token_arguments_total": 34, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.575098814229249, "step": 194, "train_accuracy_first_token_arguments": 0.7368421052631579, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.5780632411067194, "grad_norm": 1.7625337382807924, "learning_rate": 1.9975149758303885e-06, "loss": 0.148, "step": 195 }, { "epoch": 0.5780632411067194, "step": 195, "train_accuracy_first_token_arguments": 0.8055555555555556, "train_accuracy_first_token_arguments_total": 36, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 19 }, { "epoch": 0.5780632411067194, "step": 195, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 23, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { "epoch": 0.5780632411067194, "step": 195, "train_accuracy_first_token_arguments": 0.9523809523809523, "train_accuracy_first_token_arguments_total": 21, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 1, "train_accuracy_recipient_token_stat": 1.0, "train_accuracy_recipient_token_stat_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.5810276679841897, "grad_norm": 1.8753430945988403, "learning_rate": 1.97393836259058e-06, "loss": 0.2221, "step": 196 }, { "epoch": 0.5810276679841897, "step": 196, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 28, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.5810276679841897, "step": 196, "train_accuracy_first_token_arguments": 0.8367346938775511, "train_accuracy_first_token_arguments_total": 49, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.5810276679841897, "step": 196, "train_accuracy_first_token_arguments": 0.9230769230769231, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.8571428571428571, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_monitor": 1.0, "train_accuracy_recipient_token_monitor_total": 1, "train_total_number_recipient_token": 14 }, { "epoch": 0.58399209486166, "grad_norm": 2.1633104244151853, "learning_rate": 1.9504106030907605e-06, "loss": 0.2133, "step": 197 }, { "epoch": 0.58399209486166, "step": 197, "train_accuracy_first_token_arguments": 0.9090909090909091, "train_accuracy_first_token_arguments_total": 22, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_construction": 1.0, "train_accuracy_recipient_token_construction_total": 2, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.58399209486166, "step": 197, "train_accuracy_first_token_arguments": 0.9411764705882353, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 17 }, { "epoch": 0.58399209486166, "step": 197, "train_accuracy_first_token_arguments": 0.6527777777777778, "train_accuracy_first_token_arguments_total": 72, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.5869565217391305, "grad_norm": 2.0288710387741813, "learning_rate": 1.9269338822822047e-06, "loss": 0.1518, "step": 198 }, { "epoch": 0.5869565217391305, "step": 198, "train_accuracy_first_token_arguments": 0.8235294117647058, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_make": 1.0, "train_accuracy_recipient_token_make_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.5869565217391305, "step": 198, "train_accuracy_first_token_arguments": 0.7777777777777778, "train_accuracy_first_token_arguments_total": 27, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.5869565217391305, "step": 198, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 0.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_compile": 1.0, "train_accuracy_recipient_token_compile_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 20 }, { "epoch": 0.5899209486166008, "grad_norm": 2.090710210779781, "learning_rate": 1.9035103803763793e-06, "loss": 0.1552, "step": 199 }, { "epoch": 0.5899209486166008, "step": 199, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 21, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.5899209486166008, "step": 199, "train_accuracy_first_token_arguments": 0.8571428571428571, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 1, "train_accuracy_recipient_token_suggest": 1.0, "train_accuracy_recipient_token_suggest_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.5899209486166008, "step": 199, "train_accuracy_first_token_arguments": 0.8125, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_building": 1.0, "train_accuracy_recipient_token_building_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.5928853754940712, "grad_norm": 1.7471771012162318, "learning_rate": 1.8801422726424735e-06, "loss": 0.1951, "step": 200 }, { "epoch": 0.5928853754940712, "step": 200, "train_accuracy_first_token_arguments": 0.9230769230769231, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_air": 1.0, "train_accuracy_recipient_token_air_total": 1, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 16 }, { "epoch": 0.5928853754940712, "step": 200, "train_accuracy_first_token_arguments": 0.8888888888888888, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_average": 1.0, "train_accuracy_recipient_token_average_total": 1, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.5928853754940712, "step": 200, "train_accuracy_first_token_arguments": 0.8823529411764706, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_attach": 1.0, "train_accuracy_recipient_token_attach_total": 1, "train_accuracy_recipient_token_clean": 1.0, "train_accuracy_recipient_token_clean_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_remove": 1.0, "train_accuracy_recipient_token_remove_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.5958498023715415, "grad_norm": 2.1505922703967726, "learning_rate": 1.8568317292053894e-06, "loss": 0.2902, "step": 201 }, { "epoch": 0.5958498023715415, "step": 201, "train_accuracy_first_token_arguments": 0.8235294117647058, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.5958498023715415, "step": 201, "train_accuracy_first_token_arguments": 0.5925925925925926, "train_accuracy_first_token_arguments_total": 27, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_ext": 1.0, "train_accuracy_recipient_token_ext_total": 1, "train_accuracy_recipient_token_fire": 1.0, "train_accuracy_recipient_token_fire_total": 4, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 20 }, { "epoch": 0.5958498023715415, "step": 201, "train_accuracy_first_token_arguments": 0.9032258064516129, "train_accuracy_first_token_arguments_total": 31, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_d": 1.0, "train_accuracy_recipient_token_d_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.5988142292490118, "grad_norm": 1.8633933589741076, "learning_rate": 1.8335809148442074e-06, "loss": 0.2595, "step": 202 }, { "epoch": 0.5988142292490118, "step": 202, "train_accuracy_first_token_arguments": 0.8888888888888888, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_accuracy_recipient_token_plant": 1.0, "train_accuracy_recipient_token_plant_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.5988142292490118, "step": 202, "train_accuracy_first_token_arguments": 0.8214285714285714, "train_accuracy_first_token_arguments_total": 28, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_convert": 1.0, "train_accuracy_recipient_token_convert_total": 1, "train_accuracy_recipient_token_create": 0.0, "train_accuracy_recipient_token_create_total": 1, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.5988142292490118, "step": 202, "train_accuracy_first_token_arguments": 0.52, "train_accuracy_first_token_arguments_total": 25, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_calculate": 0.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_measure": 1.0, "train_accuracy_recipient_token_measure_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.6017786561264822, "grad_norm": 2.32947277265756, "learning_rate": 1.8103919887911525e-06, "loss": 0.3225, "step": 203 }, { "epoch": 0.6017786561264822, "step": 203, "train_accuracy_first_token_arguments": 0.8888888888888888, "train_accuracy_first_token_arguments_total": 9, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.6017786561264822, "step": 203, "train_accuracy_first_token_arguments": 0.9, "train_accuracy_first_token_arguments_total": 10, "train_accuracy_recipient_token": 0.8125, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_d": 1.0, "train_accuracy_recipient_token_d_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 16 }, { "epoch": 0.6017786561264822, "step": 203, "train_accuracy_first_token_arguments": 0.9411764705882353, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9166666666666666, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_bridge": 1.0, "train_accuracy_recipient_token_bridge_total": 3, "train_accuracy_recipient_token_find": 0.75, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_invest": 1.0, "train_accuracy_recipient_token_invest_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.6047430830039525, "grad_norm": 1.8953594708836015, "learning_rate": 1.7872671045310703e-06, "loss": 0.33, "step": 204 }, { "epoch": 0.6047430830039525, "step": 204, "train_accuracy_first_token_arguments": 0.9, "train_accuracy_first_token_arguments_total": 10, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 3, "train_accuracy_recipient_token_track": 1.0, "train_accuracy_recipient_token_track_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.6047430830039525, "step": 204, "train_accuracy_first_token_arguments": 0.8235294117647058, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.6047430830039525, "step": 204, "train_accuracy_first_token_arguments": 0.8, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_book": 0.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.607707509881423, "grad_norm": 1.4151900097248626, "learning_rate": 1.7642084096014405e-06, "loss": 0.1614, "step": 205 }, { "epoch": 0.607707509881423, "step": 205, "train_accuracy_first_token_arguments": 0.9705882352941176, "train_accuracy_first_token_arguments_total": 34, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_organ": 0.0, "train_accuracy_recipient_token_organ_total": 1, "train_accuracy_recipient_token_review": 0.5, "train_accuracy_recipient_token_review_total": 2, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.607707509881423, "step": 205, "train_accuracy_first_token_arguments": 0.9166666666666666, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_parse": 0.5, "train_accuracy_recipient_token_parse_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.607707509881423, "step": 205, "train_accuracy_first_token_arguments": 0.9230769230769231, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_convert": 1.0, "train_accuracy_recipient_token_convert_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.6106719367588933, "grad_norm": 2.1839121635772756, "learning_rate": 1.7412180453929412e-06, "loss": 0.2944, "step": 206 }, { "epoch": 0.6106719367588933, "step": 206, "train_accuracy_first_token_arguments": 0.8387096774193549, "train_accuracy_first_token_arguments_total": 31, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 0.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.6106719367588933, "step": 206, "train_accuracy_first_token_arguments": 0.9180327868852459, "train_accuracy_first_token_arguments_total": 61, "train_accuracy_recipient_token": 0.8571428571428571, "train_accuracy_recipient_token_all": 0.9285714285714286, "train_accuracy_recipient_token_all_total": 14, "train_accuracy_recipient_token_calculate": 0.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 21 }, { "epoch": 0.6106719367588933, "step": 206, "train_accuracy_first_token_arguments": 0.8695652173913043, "train_accuracy_first_token_arguments_total": 23, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_average": 1.0, "train_accuracy_recipient_token_average_total": 1, "train_accuracy_recipient_token_cat": 0.0, "train_accuracy_recipient_token_cat_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.6136363636363636, "grad_norm": 1.7289424673499263, "learning_rate": 1.718298146950585e-06, "loss": 0.1488, "step": 207 }, { "epoch": 0.6136363636363636, "step": 207, "train_accuracy_first_token_arguments": 0.7272727272727273, "train_accuracy_first_token_arguments_total": 11, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8571428571428571, "train_accuracy_recipient_token_all_total": 14, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.6136363636363636, "step": 207, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.8666666666666667, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 15 }, { "epoch": 0.6136363636363636, "step": 207, "train_accuracy_first_token_arguments": 0.8181818181818182, "train_accuracy_first_token_arguments_total": 11, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_design": 1.0, "train_accuracy_recipient_token_design_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_monitor": 1.0, "train_accuracy_recipient_token_monitor_total": 1, "train_accuracy_recipient_token_order": 0.0, "train_accuracy_recipient_token_order_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.616600790513834, "grad_norm": 2.1828846473287093, "learning_rate": 1.6954508427754435e-06, "loss": 0.2959, "step": 208 }, { "epoch": 0.616600790513834, "step": 208, "train_accuracy_first_token_arguments": 0.7608695652173914, "train_accuracy_first_token_arguments_total": 46, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 0.5, "train_accuracy_recipient_token_calculate_total": 2, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.616600790513834, "step": 208, "train_accuracy_first_token_arguments": 0.8888888888888888, "train_accuracy_first_token_arguments_total": 9, "train_accuracy_recipient_token": 0.75, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 16 }, { "epoch": 0.616600790513834, "step": 208, "train_accuracy_first_token_arguments": 0.95, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 0.7894736842105263, "train_accuracy_recipient_token_all": 0.7333333333333333, "train_accuracy_recipient_token_all_total": 15, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.6195652173913043, "grad_norm": 2.3743297354062536, "learning_rate": 1.6726782546269793e-06, "loss": 0.253, "step": 209 }, { "epoch": 0.6195652173913043, "step": 209, "train_accuracy_first_token_arguments": 0.8571428571428571, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_p": 1.0, "train_accuracy_recipient_token_p_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.6195652173913043, "step": 209, "train_accuracy_first_token_arguments": 0.8571428571428571, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.7142857142857143, "train_accuracy_recipient_token_get_total": 7, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.6195652173913043, "step": 209, "train_accuracy_first_token_arguments": 0.8, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_locate": 1.0, "train_accuracy_recipient_token_locate_total": 1, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.6225296442687747, "grad_norm": 1.6892603326816198, "learning_rate": 1.6499824973260086e-06, "loss": 0.2491, "step": 210 }, { "epoch": 0.6225296442687747, "step": 210, "train_accuracy_first_token_arguments": 0.9230769230769231, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.6225296442687747, "step": 210, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 26, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.6225296442687747, "step": 210, "train_accuracy_first_token_arguments": 0.75, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.6254940711462451, "grad_norm": 2.3949353891643197, "learning_rate": 1.6273656785582986e-06, "loss": 0.3576, "step": 211 }, { "epoch": 0.6254940711462451, "step": 211, "train_accuracy_first_token_arguments": 0.6111111111111112, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.6254940711462451, "step": 211, "train_accuracy_first_token_arguments": 0.9166666666666666, "train_accuracy_first_token_arguments_total": 24, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_check": 0.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_interpret": 1.0, "train_accuracy_recipient_token_interpret_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.6254940711462451, "step": 211, "train_accuracy_first_token_arguments": 0.75, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 1, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.6284584980237155, "grad_norm": 2.3847289223183123, "learning_rate": 1.6048298986788345e-06, "loss": 0.2644, "step": 212 }, { "epoch": 0.6284584980237155, "eval_accuracy_first_token_arguments": 0.923923006416132, "eval_accuracy_first_token_arguments_total": 1091, "eval_accuracy_recipient_token": 0.9356984478935698, "eval_accuracy_recipient_token_all": 0.9551656920077972, "eval_accuracy_recipient_token_all_total": 513, "eval_accuracy_recipient_token_calculate": 1.0, "eval_accuracy_recipient_token_calculate_total": 13, "eval_accuracy_recipient_token_find": 0.9473684210526315, "eval_accuracy_recipient_token_find_total": 57, "eval_accuracy_recipient_token_get": 0.9387755102040817, "eval_accuracy_recipient_token_get_total": 196, "eval_accuracy_recipient_token_identify": 0.6666666666666666, "eval_accuracy_recipient_token_identify_total": 6, "eval_loss": 0.22125016152858734, "eval_perplexity": 1.034604144934006, "eval_runtime": 113.1216, "eval_samples_per_second": 0.893, "eval_steps_per_second": 0.451, "eval_total_number_recipient_token": 902, "step": 212 }, { "epoch": 0.6284584980237155, "step": 212, "train_accuracy_first_token_arguments": 0.9473684210526315, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.6284584980237155, "step": 212, "train_accuracy_first_token_arguments": 0.9411764705882353, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 19 }, { "epoch": 0.6284584980237155, "step": 212, "train_accuracy_first_token_arguments": 0.8333333333333334, "train_accuracy_first_token_arguments_total": 42, "train_accuracy_recipient_token": 0.7894736842105263, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 0.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 0.75, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_locate": 0.0, "train_accuracy_recipient_token_locate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.6314229249011858, "grad_norm": 1.4624962550695582, "learning_rate": 1.58237725051677e-06, "loss": 0.1802, "step": 213 }, { "epoch": 0.6314229249011858, "step": 213, "train_accuracy_first_token_arguments": 0.8, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_sum": 1.0, "train_accuracy_recipient_token_sum_total": 2, "train_accuracy_recipient_token_translate": 0.5, "train_accuracy_recipient_token_translate_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.6314229249011858, "step": 213, "train_accuracy_first_token_arguments": 0.9230769230769231, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_register": 1.0, "train_accuracy_recipient_token_register_total": 1, "train_accuracy_recipient_token_verify": 1.0, "train_accuracy_recipient_token_verify_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.6314229249011858, "step": 213, "train_accuracy_first_token_arguments": 0.8918918918918919, "train_accuracy_first_token_arguments_total": 37, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_filter": 1.0, "train_accuracy_recipient_token_filter_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_accuracy_recipient_token_identify": 0.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.6343873517786561, "grad_norm": 1.6957397755168004, "learning_rate": 1.5600098191810682e-06, "loss": 0.275, "step": 214 }, { "epoch": 0.6343873517786561, "step": 214, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 28, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.6343873517786561, "step": 214, "train_accuracy_first_token_arguments": 0.8873239436619719, "train_accuracy_first_token_arguments_total": 71, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_boat": 1.0, "train_accuracy_recipient_token_boat_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_suggest": 1.0, "train_accuracy_recipient_token_suggest_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.6343873517786561, "step": 214, "train_accuracy_first_token_arguments": 0.8214285714285714, "train_accuracy_first_token_arguments_total": 28, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_hare": 0.6666666666666666, "train_accuracy_recipient_token_hare_total": 3, "train_accuracy_recipient_token_invest": 1.0, "train_accuracy_recipient_token_invest_total": 1, "train_accuracy_recipient_token_plane": 1.0, "train_accuracy_recipient_token_plane_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.6373517786561265, "grad_norm": 1.64437830735148, "learning_rate": 1.5377296818668638e-06, "loss": 0.1791, "step": 215 }, { "epoch": 0.6373517786561265, "step": 215, "train_accuracy_first_token_arguments": 0.7407407407407407, "train_accuracy_first_token_arguments_total": 27, "train_accuracy_recipient_token": 0.8125, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_update": 0.0, "train_accuracy_recipient_token_update_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.6373517786561265, "step": 215, "train_accuracy_first_token_arguments": 0.6764705882352942, "train_accuracy_first_token_arguments_total": 34, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_evaluate": 0.0, "train_accuracy_recipient_token_evaluate_total": 1, "train_accuracy_recipient_token_find": 0.3333333333333333, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_pack": 1.0, "train_accuracy_recipient_token_pack_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.6373517786561265, "step": 215, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_ind": 1.0, "train_accuracy_recipient_token_ind_total": 4, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.6403162055335968, "grad_norm": 2.485352785980747, "learning_rate": 1.5155389076625663e-06, "loss": 0.293, "step": 216 }, { "epoch": 0.6403162055335968, "step": 216, "train_accuracy_first_token_arguments": 0.9473684210526315, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_architecture": 1.0, "train_accuracy_recipient_token_architecture_total": 1, "train_accuracy_recipient_token_compare": 0.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 0.7142857142857143, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { "epoch": 0.6403162055335968, "step": 216, "train_accuracy_first_token_arguments": 0.967741935483871, "train_accuracy_first_token_arguments_total": 31, "train_accuracy_recipient_token": 0.85, "train_accuracy_recipient_token_all": 0.8461538461538461, "train_accuracy_recipient_token_all_total": 13, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 20 }, { "epoch": 0.6403162055335968, "step": 216, "train_accuracy_first_token_arguments": 0.8421052631578947, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 20 }, { "epoch": 0.6432806324110671, "grad_norm": 1.78457120796898, "learning_rate": 1.4934395573577016e-06, "loss": 0.3178, "step": 217 }, { "epoch": 0.6432806324110671, "step": 217, "train_accuracy_first_token_arguments": 0.9565217391304348, "train_accuracy_first_token_arguments_total": 23, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_compare": 0.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.6432806324110671, "step": 217, "train_accuracy_first_token_arguments": 0.7692307692307693, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.6432806324110671, "step": 217, "train_accuracy_first_token_arguments": 0.9333333333333333, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_inspect": 1.0, "train_accuracy_recipient_token_inspect_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.6462450592885376, "grad_norm": 2.027667840038845, "learning_rate": 1.4714336832515386e-06, "loss": 0.167, "step": 218 }, { "epoch": 0.6462450592885376, "step": 218, "train_accuracy_first_token_arguments": 0.9285714285714286, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_language": 1.0, "train_accuracy_recipient_token_language_total": 1, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.6462450592885376, "step": 218, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.6462450592885376, "step": 218, "train_accuracy_first_token_arguments": 0.85, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.6492094861660079, "grad_norm": 1.4967382429323417, "learning_rate": 1.449523328962496e-06, "loss": 0.2031, "step": 219 }, { "epoch": 0.6492094861660079, "step": 219, "train_accuracy_first_token_arguments": 0.7692307692307693, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.3333333333333333, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.6492094861660079, "step": 219, "train_accuracy_first_token_arguments": 0.8461538461538461, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_explain": 0.8, "train_accuracy_recipient_token_explain_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 20 }, { "epoch": 0.6492094861660079, "step": 219, "train_accuracy_first_token_arguments": 0.8125, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_b": 1.0, "train_accuracy_recipient_token_b_total": 1, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.6521739130434783, "grad_norm": 1.8670991248523237, "learning_rate": 1.4277105292383594e-06, "loss": 0.1995, "step": 220 }, { "epoch": 0.6521739130434783, "step": 220, "train_accuracy_first_token_arguments": 0.9090909090909091, "train_accuracy_first_token_arguments_total": 11, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_inspect": 1.0, "train_accuracy_recipient_token_inspect_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.6521739130434783, "step": 220, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 10, "train_accuracy_recipient_token": 0.8125, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 16 }, { "epoch": 0.6521739130434783, "step": 220, "train_accuracy_first_token_arguments": 0.8333333333333334, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 19 }, { "epoch": 0.6551383399209486, "grad_norm": 2.550246191316389, "learning_rate": 1.4059973097673187e-06, "loss": 0.2752, "step": 221 }, { "epoch": 0.6551383399209486, "step": 221, "train_accuracy_first_token_arguments": 0.9333333333333333, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_invest": 1.0, "train_accuracy_recipient_token_invest_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.6551383399209486, "step": 221, "train_accuracy_first_token_arguments": 0.625, "train_accuracy_first_token_arguments_total": 24, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_load": 1.0, "train_accuracy_recipient_token_load_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.6551383399209486, "step": 221, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 9, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.658102766798419, "grad_norm": 1.8305223962446249, "learning_rate": 1.3843856869898486e-06, "loss": 0.2692, "step": 222 }, { "epoch": 0.658102766798419, "step": 222, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_device": 1.0, "train_accuracy_recipient_token_device_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.658102766798419, "step": 222, "train_accuracy_first_token_arguments": 0.8823529411764706, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.9, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 20 }, { "epoch": 0.658102766798419, "step": 222, "train_accuracy_first_token_arguments": 0.88, "train_accuracy_first_token_arguments_total": 50, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_park": 1.0, "train_accuracy_recipient_token_park_total": 2, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.6610671936758893, "grad_norm": 1.7979076335597703, "learning_rate": 1.3628776679114516e-06, "loss": 0.1334, "step": 223 }, { "epoch": 0.6610671936758893, "step": 223, "train_accuracy_first_token_arguments": 0.7894736842105263, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.8666666666666667, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 0.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 15 }, { "epoch": 0.6610671936758893, "step": 223, "train_accuracy_first_token_arguments": 0.5, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_explain": 1.0, "train_accuracy_recipient_token_explain_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_translate": 0.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.6610671936758893, "step": 223, "train_accuracy_first_token_arguments": 0.9210526315789473, "train_accuracy_first_token_arguments_total": 38, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_local": 1.0, "train_accuracy_recipient_token_local_total": 1, "train_accuracy_recipient_token_plan": 1.0, "train_accuracy_recipient_token_plan_total": 1, "train_accuracy_recipient_token_restaurant": 1.0, "train_accuracy_recipient_token_restaurant_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.6640316205533597, "grad_norm": 2.7045237044612875, "learning_rate": 1.3414752499162676e-06, "loss": 0.3284, "step": 224 }, { "epoch": 0.6640316205533597, "step": 224, "train_accuracy_first_token_arguments": 0.9310344827586207, "train_accuracy_first_token_arguments_total": 29, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_display": 1.0, "train_accuracy_recipient_token_display_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_suggest": 0.0, "train_accuracy_recipient_token_suggest_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.6640316205533597, "step": 224, "train_accuracy_first_token_arguments": 0.8333333333333334, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_classify": 0.0, "train_accuracy_recipient_token_classify_total": 1, "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 1, "train_accuracy_recipient_token_g": 1.0, "train_accuracy_recipient_token_g_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.6640316205533597, "step": 224, "train_accuracy_first_token_arguments": 0.7692307692307693, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.6669960474308301, "grad_norm": 1.8337644487638454, "learning_rate": 1.3201804205815872e-06, "loss": 0.2451, "step": 225 }, { "epoch": 0.6669960474308301, "step": 225, "train_accuracy_first_token_arguments": 0.9523809523809523, "train_accuracy_first_token_arguments_total": 21, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 19 }, { "epoch": 0.6669960474308301, "step": 225, "train_accuracy_first_token_arguments": 0.9166666666666666, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_christ": 1.0, "train_accuracy_recipient_token_christ_total": 1, "train_accuracy_recipient_token_explain": 0.0, "train_accuracy_recipient_token_explain_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.6669960474308301, "step": 225, "train_accuracy_first_token_arguments": 0.8333333333333334, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compose": 1.0, "train_accuracy_recipient_token_compose_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.6699604743083004, "grad_norm": 1.601543108317484, "learning_rate": 1.2989951574932693e-06, "loss": 0.2814, "step": 226 }, { "epoch": 0.6699604743083004, "step": 226, "train_accuracy_first_token_arguments": 0.967741935483871, "train_accuracy_first_token_arguments_total": 31, "train_accuracy_recipient_token": 0.8, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_configure": 0.0, "train_accuracy_recipient_token_configure_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 15 }, { "epoch": 0.6699604743083004, "step": 226, "train_accuracy_first_token_arguments": 0.8888888888888888, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 0.9047619047619048, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 21 }, { "epoch": 0.6699604743083004, "step": 226, "train_accuracy_first_token_arguments": 0.9333333333333333, "train_accuracy_first_token_arguments_total": 30, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_calculate": 0.75, "train_accuracy_recipient_token_calculate_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 20 }, { "epoch": 0.6729249011857708, "grad_norm": 1.9791094562714513, "learning_rate": 1.277921428062091e-06, "loss": 0.2229, "step": 227 }, { "epoch": 0.6729249011857708, "step": 227, "train_accuracy_first_token_arguments": 0.9629629629629629, "train_accuracy_first_token_arguments_total": 54, "train_accuracy_recipient_token": 0.7777777777777778, "train_accuracy_recipient_token_all": 0.6666666666666666, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.6729249011857708, "step": 227, "train_accuracy_first_token_arguments": 0.7777777777777778, "train_accuracy_first_token_arguments_total": 9, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_d": 1.0, "train_accuracy_recipient_token_d_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 14 }, { "epoch": 0.6729249011857708, "step": 227, "train_accuracy_first_token_arguments": 0.6862745098039216, "train_accuracy_first_token_arguments_total": 51, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_customize": 1.0, "train_accuracy_recipient_token_customize_total": 1, "train_accuracy_recipient_token_draw": 1.0, "train_accuracy_recipient_token_draw_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.6758893280632411, "grad_norm": 2.308708795701363, "learning_rate": 1.2569611893410374e-06, "loss": 0.3605, "step": 228 }, { "epoch": 0.6758893280632411, "step": 228, "train_accuracy_first_token_arguments": 0.6666666666666666, "train_accuracy_first_token_arguments_total": 21, "train_accuracy_recipient_token": 0.7894736842105263, "train_accuracy_recipient_token_all": 0.7142857142857143, "train_accuracy_recipient_token_all_total": 14, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_event": 1.0, "train_accuracy_recipient_token_event_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.6758893280632411, "step": 228, "train_accuracy_first_token_arguments": 0.7619047619047619, "train_accuracy_first_token_arguments_total": 21, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 20 }, { "epoch": 0.6758893280632411, "step": 228, "train_accuracy_first_token_arguments": 0.7, "train_accuracy_first_token_arguments_total": 10, "train_accuracy_recipient_token": 0.8571428571428571, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_customize": 1.0, "train_accuracy_recipient_token_customize_total": 1, "train_accuracy_recipient_token_remove": 1.0, "train_accuracy_recipient_token_remove_total": 1, "train_accuracy_recipient_token_translate": 0.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 14 }, { "epoch": 0.6788537549407114, "grad_norm": 2.4044180995153486, "learning_rate": 1.2361163878435594e-06, "loss": 0.4432, "step": 229 }, { "epoch": 0.6788537549407114, "step": 229, "train_accuracy_first_token_arguments": 0.9615384615384616, "train_accuracy_first_token_arguments_total": 26, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 15 }, { "epoch": 0.6788537549407114, "step": 229, "train_accuracy_first_token_arguments": 0.9230769230769231, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 20 }, { "epoch": 0.6788537549407114, "step": 229, "train_accuracy_first_token_arguments": 0.7666666666666667, "train_accuracy_first_token_arguments_total": 30, "train_accuracy_recipient_token": 0.85, "train_accuracy_recipient_token_all": 0.75, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.6818181818181818, "grad_norm": 2.0846069978627417, "learning_rate": 1.2153889593628032e-06, "loss": 0.2884, "step": 230 }, { "epoch": 0.6818181818181818, "step": 230, "train_accuracy_first_token_arguments": 0.5185185185185185, "train_accuracy_first_token_arguments_total": 27, "train_accuracy_recipient_token": 0.8571428571428571, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 14 }, { "epoch": 0.6818181818181818, "step": 230, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.6818181818181818, "step": 230, "train_accuracy_first_token_arguments": 0.6060606060606061, "train_accuracy_first_token_arguments_total": 33, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_wine": 1.0, "train_accuracy_recipient_token_wine_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.6847826086956522, "grad_norm": 1.903972306272471, "learning_rate": 1.1947808287918406e-06, "loss": 0.233, "step": 231 }, { "epoch": 0.6847826086956522, "step": 231, "train_accuracy_first_token_arguments": 0.75, "train_accuracy_first_token_arguments_total": 24, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.6847826086956522, "step": 231, "train_accuracy_first_token_arguments": 0.9411764705882353, "train_accuracy_first_token_arguments_total": 51, "train_accuracy_recipient_token": 0.7647058823529411, "train_accuracy_recipient_token_all": 0.7333333333333333, "train_accuracy_recipient_token_all_total": 15, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.6847826086956522, "step": 231, "train_accuracy_first_token_arguments": 0.76, "train_accuracy_first_token_arguments_total": 25, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.6877470355731226, "grad_norm": 3.071022026928369, "learning_rate": 1.1742939099449126e-06, "loss": 0.3897, "step": 232 }, { "epoch": 0.6877470355731226, "step": 232, "train_accuracy_first_token_arguments": 0.6440677966101694, "train_accuracy_first_token_arguments_total": 59, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_draw": 1.0, "train_accuracy_recipient_token_draw_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_inspect": 1.0, "train_accuracy_recipient_token_inspect_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.6877470355731226, "step": 232, "train_accuracy_first_token_arguments": 0.5454545454545454, "train_accuracy_first_token_arguments_total": 33, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_dec": 1.0, "train_accuracy_recipient_token_dec_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_verify": 1.0, "train_accuracy_recipient_token_verify_total": 1, "train_total_number_recipient_token": 15 }, { "epoch": 0.6877470355731226, "step": 232, "train_accuracy_first_token_arguments": 0.8571428571428571, "train_accuracy_first_token_arguments_total": 21, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_convert": 1.0, "train_accuracy_recipient_token_convert_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.6907114624505929, "grad_norm": 2.858676916448621, "learning_rate": 1.153930105379695e-06, "loss": 0.2713, "step": 233 }, { "epoch": 0.6907114624505929, "step": 233, "train_accuracy_first_token_arguments": 0.9411764705882353, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.6907114624505929, "step": 233, "train_accuracy_first_token_arguments": 0.9459459459459459, "train_accuracy_first_token_arguments_total": 37, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.8571428571428571, "train_accuracy_recipient_token_find_total": 7, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.6907114624505929, "step": 233, "train_accuracy_first_token_arguments": 0.8260869565217391, "train_accuracy_first_token_arguments_total": 23, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_enh": 1.0, "train_accuracy_recipient_token_enh_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_optimize": 1.0, "train_accuracy_recipient_token_optimize_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.6936758893280632, "grad_norm": 1.4873714467096886, "learning_rate": 1.1336913062206157e-06, "loss": 0.1483, "step": 234 }, { "epoch": 0.6936758893280632, "step": 234, "train_accuracy_first_token_arguments": 0.9, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8571428571428571, "train_accuracy_recipient_token_all_total": 14, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.6936758893280632, "step": 234, "train_accuracy_first_token_arguments": 0.9090909090909091, "train_accuracy_first_token_arguments_total": 11, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.8333333333333334, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { "epoch": 0.6936758893280632, "step": 234, "train_accuracy_first_token_arguments": 0.8333333333333334, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_detect": 1.0, "train_accuracy_recipient_token_detect_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_review": 1.0, "train_accuracy_recipient_token_review_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.6966403162055336, "grad_norm": 1.5877712208219548, "learning_rate": 1.1135793919832336e-06, "loss": 0.2483, "step": 235 }, { "epoch": 0.6966403162055336, "step": 235, "train_accuracy_first_token_arguments": 0.9411764705882353, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_join": 1.0, "train_accuracy_recipient_token_join_total": 1, "train_accuracy_recipient_token_vol": 1.0, "train_accuracy_recipient_token_vol_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.6966403162055336, "step": 235, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 11, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.6966403162055336, "step": 235, "train_accuracy_first_token_arguments": 0.9047619047619048, "train_accuracy_first_token_arguments_total": 21, "train_accuracy_recipient_token": 0.85, "train_accuracy_recipient_token_all": 0.7692307692307693, "train_accuracy_recipient_token_all_total": 13, "train_accuracy_recipient_token_detect": 1.0, "train_accuracy_recipient_token_detect_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.6996047430830039, "grad_norm": 1.392788536484163, "learning_rate": 1.09359623039969e-06, "loss": 0.1911, "step": 236 }, { "epoch": 0.6996047430830039, "step": 236, "train_accuracy_first_token_arguments": 0.9090909090909091, "train_accuracy_first_token_arguments_total": 22, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_optimize": 0.0, "train_accuracy_recipient_token_optimize_total": 1, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.6996047430830039, "step": 236, "train_accuracy_first_token_arguments": 0.8108108108108109, "train_accuracy_first_token_arguments_total": 37, "train_accuracy_recipient_token": 0.7857142857142857, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_design": 1.0, "train_accuracy_recipient_token_design_total": 1, "train_accuracy_recipient_token_im": 1.0, "train_accuracy_recipient_token_im_total": 1, "train_accuracy_recipient_token_install": 0.0, "train_accuracy_recipient_token_install_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 14 }, { "epoch": 0.6996047430830039, "step": 236, "train_accuracy_first_token_arguments": 0.8695652173913043, "train_accuracy_first_token_arguments_total": 23, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.7025691699604744, "grad_norm": 1.9185387350295975, "learning_rate": 1.0737436772452602e-06, "loss": 0.2789, "step": 237 }, { "epoch": 0.7025691699604744, "step": 237, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.7142857142857143, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.7025691699604744, "step": 237, "train_accuracy_first_token_arguments": 0.9285714285714286, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 20 }, { "epoch": 0.7025691699604744, "step": 237, "train_accuracy_first_token_arguments": 0.7435897435897436, "train_accuracy_first_token_arguments_total": 39, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.7055335968379447, "grad_norm": 1.6960089025056184, "learning_rate": 1.054023576166014e-06, "loss": 0.2385, "step": 238 }, { "epoch": 0.7055335968379447, "step": 238, "train_accuracy_first_token_arguments": 0.9, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.7055335968379447, "step": 238, "train_accuracy_first_token_arguments": 0.8947368421052632, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.9333333333333333, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_track": 1.0, "train_accuracy_recipient_token_track_total": 1, "train_total_number_recipient_token": 15 }, { "epoch": 0.7055335968379447, "step": 238, "train_accuracy_first_token_arguments": 0.9090909090909091, "train_accuracy_first_token_arguments_total": 44, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 16 }, { "epoch": 0.708498023715415, "grad_norm": 2.366641831793585, "learning_rate": 1.0344377585076e-06, "loss": 0.4772, "step": 239 }, { "epoch": 0.708498023715415, "step": 239, "train_accuracy_first_token_arguments": 0.9444444444444444, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { "epoch": 0.708498023715415, "step": 239, "train_accuracy_first_token_arguments": 0.75, "train_accuracy_first_token_arguments_total": 28, "train_accuracy_recipient_token": 0.9333333333333333, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_merge": 1.0, "train_accuracy_recipient_token_merge_total": 1, "train_total_number_recipient_token": 15 }, { "epoch": 0.708498023715415, "step": 239, "train_accuracy_first_token_arguments": 0.9411764705882353, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.7114624505928854, "grad_norm": 2.0428633070085302, "learning_rate": 1.0149880431451736e-06, "loss": 0.2845, "step": 240 }, { "epoch": 0.7114624505928854, "step": 240, "train_accuracy_first_token_arguments": 0.9473684210526315, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 0.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.7114624505928854, "step": 240, "train_accuracy_first_token_arguments": 0.8421052631578947, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_inspect": 1.0, "train_accuracy_recipient_token_inspect_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.7114624505928854, "step": 240, "train_accuracy_first_token_arguments": 0.9, "train_accuracy_first_token_arguments_total": 30, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_d": 1.0, "train_accuracy_recipient_token_d_total": 1, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_vehicle": 0.6666666666666666, "train_accuracy_recipient_token_vehicle_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.7144268774703557, "grad_norm": 1.870434102867515, "learning_rate": 9.956762363144892e-07, "loss": 0.1578, "step": 241 }, { "epoch": 0.7144268774703557, "step": 241, "train_accuracy_first_token_arguments": 0.7222222222222222, "train_accuracy_first_token_arguments_total": 36, "train_accuracy_recipient_token": 0.7894736842105263, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.7144268774703557, "step": 241, "train_accuracy_first_token_arguments": 0.8235294117647058, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_recommend": 0.0, "train_accuracy_recipient_token_recommend_total": 1, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.7144268774703557, "step": 241, "train_accuracy_first_token_arguments": 0.7272727272727273, "train_accuracy_first_token_arguments_total": 11, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 0.5, "train_accuracy_recipient_token_analyze_total": 2, "train_accuracy_recipient_token_building": 1.0, "train_accuracy_recipient_token_building_total": 3, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_object": 1.0, "train_accuracy_recipient_token_object_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.717391304347826, "grad_norm": 2.192061804607294, "learning_rate": 9.765041314441529e-07, "loss": 0.3503, "step": 242 }, { "epoch": 0.717391304347826, "step": 242, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8888888888888888, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.717391304347826, "step": 242, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 11, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_join": 1.0, "train_accuracy_recipient_token_join_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.717391304347826, "step": 242, "train_accuracy_first_token_arguments": 0.9090909090909091, "train_accuracy_first_token_arguments_total": 11, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compile": 1.0, "train_accuracy_recipient_token_compile_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 17 }, { "epoch": 0.7203557312252964, "grad_norm": 2.540677501201676, "learning_rate": 9.574735089890765e-07, "loss": 0.3029, "step": 243 }, { "epoch": 0.7203557312252964, "step": 243, "train_accuracy_first_token_arguments": 0.8, "train_accuracy_first_token_arguments_total": 10, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.7203557312252964, "step": 243, "train_accuracy_first_token_arguments": 0.8235294117647058, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.7203557312252964, "step": 243, "train_accuracy_first_token_arguments": 0.6875, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_exp": 1.0, "train_accuracy_recipient_token_exp_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.7233201581027668, "grad_norm": 1.681752257297182, "learning_rate": 9.385861362651322e-07, "loss": 0.2678, "step": 244 }, { "epoch": 0.7233201581027668, "step": 244, "train_accuracy_first_token_arguments": 0.717391304347826, "train_accuracy_first_token_arguments_total": 46, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 1, "train_accuracy_recipient_token_design": 1.0, "train_accuracy_recipient_token_design_total": 2, "train_accuracy_recipient_token_draw": 1.0, "train_accuracy_recipient_token_draw_total": 1, "train_accuracy_recipient_token_evaluate": 0.0, "train_accuracy_recipient_token_evaluate_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.7233201581027668, "step": 244, "train_accuracy_first_token_arguments": 0.9130434782608695, "train_accuracy_first_token_arguments_total": 23, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 3, "train_accuracy_recipient_token_use": 1.0, "train_accuracy_recipient_token_use_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.7233201581027668, "step": 244, "train_accuracy_first_token_arguments": 0.9423076923076923, "train_accuracy_first_token_arguments_total": 52, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_design": 1.0, "train_accuracy_recipient_token_design_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.7262845849802372, "grad_norm": 1.6081692582595575, "learning_rate": 9.198437672850249e-07, "loss": 0.2291, "step": 245 }, { "epoch": 0.7262845849802372, "step": 245, "train_accuracy_first_token_arguments": 0.8, "train_accuracy_first_token_arguments_total": 10, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 2, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.7262845849802372, "step": 245, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { "epoch": 0.7262845849802372, "step": 245, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9166666666666666, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.7292490118577075, "grad_norm": 1.786148716474924, "learning_rate": 9.012481425954053e-07, "loss": 0.2712, "step": 246 }, { "epoch": 0.7292490118577075, "step": 246, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 9, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 17 }, { "epoch": 0.7292490118577075, "step": 246, "train_accuracy_first_token_arguments": 0.9333333333333333, "train_accuracy_first_token_arguments_total": 30, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 2, "train_accuracy_recipient_token_inventory": 1.0, "train_accuracy_recipient_token_inventory_total": 1, "train_accuracy_recipient_token_sales": 1.0, "train_accuracy_recipient_token_sales_total": 1, "train_accuracy_recipient_token_shelf": 1.0, "train_accuracy_recipient_token_shelf_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.7292490118577075, "step": 246, "train_accuracy_first_token_arguments": 0.875, "train_accuracy_first_token_arguments_total": 24, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_buy": 0.0, "train_accuracy_recipient_token_buy_total": 1, "train_accuracy_recipient_token_find": 0.8571428571428571, "train_accuracy_recipient_token_find_total": 7, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 20 }, { "epoch": 0.7322134387351779, "grad_norm": 1.7149609341130112, "learning_rate": 8.828009891152301e-07, "loss": 0.2038, "step": 247 }, { "epoch": 0.7322134387351779, "step": 247, "train_accuracy_first_token_arguments": 0.7857142857142857, "train_accuracy_first_token_arguments_total": 28, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_locate": 1.0, "train_accuracy_recipient_token_locate_total": 1, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.7322134387351779, "step": 247, "train_accuracy_first_token_arguments": 0.8571428571428571, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_m": 1.0, "train_accuracy_recipient_token_m_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.7322134387351779, "step": 247, "train_accuracy_first_token_arguments": 0.9761904761904762, "train_accuracy_first_token_arguments_total": 42, "train_accuracy_recipient_token": 0.7777777777777778, "train_accuracy_recipient_token_all": 0.7857142857142857, "train_accuracy_recipient_token_all_total": 14, "train_accuracy_recipient_token_calculate": 0.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.7351778656126482, "grad_norm": 1.6828278608062666, "learning_rate": 8.64504019975386e-07, "loss": 0.218, "step": 248 }, { "epoch": 0.7351778656126482, "step": 248, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.7351778656126482, "step": 248, "train_accuracy_first_token_arguments": 0.9166666666666666, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.6, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.7351778656126482, "step": 248, "train_accuracy_first_token_arguments": 0.9523809523809523, "train_accuracy_first_token_arguments_total": 21, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9285714285714286, "train_accuracy_recipient_token_all_total": 14, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_translate": 0.0, "train_accuracy_recipient_token_translate_total": 1, "train_accuracy_recipient_token_vehicle": 1.0, "train_accuracy_recipient_token_vehicle_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.7381422924901185, "grad_norm": 2.180566937973925, "learning_rate": 8.463589343595976e-07, "loss": 0.3458, "step": 249 }, { "epoch": 0.7381422924901185, "step": 249, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 27, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9166666666666666, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_construction": 1.0, "train_accuracy_recipient_token_construction_total": 2, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.7381422924901185, "step": 249, "train_accuracy_first_token_arguments": 0.9375, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.7142857142857143, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.7381422924901185, "step": 249, "train_accuracy_first_token_arguments": 0.9285714285714286, "train_accuracy_first_token_arguments_total": 28, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.741106719367589, "grad_norm": 1.7757645324819282, "learning_rate": 8.28367417346633e-07, "loss": 0.2223, "step": 250 }, { "epoch": 0.741106719367589, "step": 250, "train_accuracy_first_token_arguments": 0.75, "train_accuracy_first_token_arguments_total": 8, "train_accuracy_recipient_token": 0.8, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 5, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 10 }, { "epoch": 0.741106719367589, "step": 250, "train_accuracy_first_token_arguments": 0.8571428571428571, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_animal": 1.0, "train_accuracy_recipient_token_animal_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_locate": 0.0, "train_accuracy_recipient_token_locate_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.741106719367589, "step": 250, "train_accuracy_first_token_arguments": 0.95, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 0.7857142857142857, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 14 }, { "epoch": 0.7440711462450593, "grad_norm": 11.280998921270033, "learning_rate": 8.105311397538085e-07, "loss": 0.5545, "step": 251 }, { "epoch": 0.7440711462450593, "step": 251, "train_accuracy_first_token_arguments": 0.7142857142857143, "train_accuracy_first_token_arguments_total": 28, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_compose": 1.0, "train_accuracy_recipient_token_compose_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.7440711462450593, "step": 251, "train_accuracy_first_token_arguments": 0.75, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.7440711462450593, "step": 251, "train_accuracy_first_token_arguments": 0.9090909090909091, "train_accuracy_first_token_arguments_total": 11, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.7470355731225297, "grad_norm": 2.177175537608394, "learning_rate": 7.928517579818312e-07, "loss": 0.2545, "step": 252 }, { "epoch": 0.7470355731225297, "step": 252, "train_accuracy_first_token_arguments": 0.8, "train_accuracy_first_token_arguments_total": 30, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.75, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.7470355731225297, "step": 252, "train_accuracy_first_token_arguments": 0.6666666666666666, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_evaluate": 0.6666666666666666, "train_accuracy_recipient_token_evaluate_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 20 }, { "epoch": 0.7470355731225297, "step": 252, "train_accuracy_first_token_arguments": 0.9473684210526315, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_te": 1.0, "train_accuracy_recipient_token_te_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.75, "grad_norm": 1.527663654975652, "learning_rate": 7.753309138609705e-07, "loss": 0.1884, "step": 253 }, { "epoch": 0.75, "step": 253, "train_accuracy_first_token_arguments": 0.7884615384615384, "train_accuracy_first_token_arguments_total": 52, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_evaluate": 0.0, "train_accuracy_recipient_token_evaluate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_suggest": 1.0, "train_accuracy_recipient_token_suggest_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.75, "step": 253, "train_accuracy_first_token_arguments": 0.6764705882352942, "train_accuracy_first_token_arguments_total": 34, "train_accuracy_recipient_token": 0.7777777777777778, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 0.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_review": 1.0, "train_accuracy_recipient_token_review_total": 1, "train_accuracy_recipient_token_stamp": 0.75, "train_accuracy_recipient_token_stamp_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.75, "step": 253, "train_accuracy_first_token_arguments": 0.8125, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_convert": 1.0, "train_accuracy_recipient_token_convert_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_sum": 1.0, "train_accuracy_recipient_token_sum_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.7529644268774703, "grad_norm": 1.9700572495421322, "learning_rate": 7.579702344985899e-07, "loss": 0.3442, "step": 254 }, { "epoch": 0.7529644268774703, "step": 254, "train_accuracy_first_token_arguments": 0.84, "train_accuracy_first_token_arguments_total": 25, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.7529644268774703, "step": 254, "train_accuracy_first_token_arguments": 0.8125, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_adjust": 1.0, "train_accuracy_recipient_token_adjust_total": 1, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_discover": 1.0, "train_accuracy_recipient_token_discover_total": 1, "train_accuracy_recipient_token_enh": 1.0, "train_accuracy_recipient_token_enh_total": 1, "train_accuracy_recipient_token_event": 1.0, "train_accuracy_recipient_token_event_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.7529644268774703, "step": 254, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_compress": 1.0, "train_accuracy_recipient_token_compress_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.7559288537549407, "grad_norm": 2.171316152952409, "learning_rate": 7.407713321280377e-07, "loss": 0.1791, "step": 255 }, { "epoch": 0.7559288537549407, "step": 255, "train_accuracy_first_token_arguments": 0.8947368421052632, "train_accuracy_first_token_arguments_total": 38, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_customize": 1.0, "train_accuracy_recipient_token_customize_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_style": 1.0, "train_accuracy_recipient_token_style_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.7559288537549407, "step": 255, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 26, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 18 }, { "epoch": 0.7559288537549407, "step": 255, "train_accuracy_first_token_arguments": 0.9285714285714286, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_rec": 1.0, "train_accuracy_recipient_token_rec_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.758893280632411, "grad_norm": 1.7774995166747949, "learning_rate": 7.237358039589271e-07, "loss": 0.2064, "step": 256 }, { "epoch": 0.758893280632411, "step": 256, "train_accuracy_first_token_arguments": 0.8529411764705882, "train_accuracy_first_token_arguments_total": 34, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.758893280632411, "step": 256, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_clean": 1.0, "train_accuracy_recipient_token_clean_total": 1, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_rent": 1.0, "train_accuracy_recipient_token_rent_total": 2, "train_total_number_recipient_token": 16 }, { "epoch": 0.758893280632411, "step": 256, "train_accuracy_first_token_arguments": 0.9090909090909091, "train_accuracy_first_token_arguments_total": 11, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 20 }, { "epoch": 0.7618577075098815, "grad_norm": 1.7945920171198126, "learning_rate": 7.068652320288081e-07, "loss": 0.1652, "step": 257 }, { "epoch": 0.7618577075098815, "step": 257, "train_accuracy_first_token_arguments": 0.9166666666666666, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8888888888888888, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.7618577075098815, "step": 257, "train_accuracy_first_token_arguments": 0.8105263157894737, "train_accuracy_first_token_arguments_total": 95, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.7618577075098815, "step": 257, "train_accuracy_first_token_arguments": 0.8666666666666667, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_calculate": 0.5, "train_accuracy_recipient_token_calculate_total": 2, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.7648221343873518, "grad_norm": 2.033550650864743, "learning_rate": 6.901611830562469e-07, "loss": 0.2407, "step": 258 }, { "epoch": 0.7648221343873518, "step": 258, "train_accuracy_first_token_arguments": 0.9047619047619048, "train_accuracy_first_token_arguments_total": 21, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_ap": 1.0, "train_accuracy_recipient_token_ap_total": 1, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.7648221343873518, "step": 258, "train_accuracy_first_token_arguments": 0.72, "train_accuracy_first_token_arguments_total": 25, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 1, "train_accuracy_recipient_token_discover": 1.0, "train_accuracy_recipient_token_discover_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_interpret": 1.0, "train_accuracy_recipient_token_interpret_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.7648221343873518, "step": 258, "train_accuracy_first_token_arguments": 0.9090909090909091, "train_accuracy_first_token_arguments_total": 22, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_event": 0.6666666666666666, "train_accuracy_recipient_token_event_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_suggest": 1.0, "train_accuracy_recipient_token_suggest_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.7677865612648221, "grad_norm": 1.6980201197958136, "learning_rate": 6.736252082953307e-07, "loss": 0.2026, "step": 259 }, { "epoch": 0.7677865612648221, "step": 259, "train_accuracy_first_token_arguments": 0.9459459459459459, "train_accuracy_first_token_arguments_total": 37, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.7677865612648221, "step": 259, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 22, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.7677865612648221, "step": 259, "train_accuracy_first_token_arguments": 0.92, "train_accuracy_first_token_arguments_total": 50, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 20 }, { "epoch": 0.7707509881422925, "grad_norm": 1.4870369650234745, "learning_rate": 6.572588433916082e-07, "loss": 0.1544, "step": 260 }, { "epoch": 0.7707509881422925, "step": 260, "train_accuracy_first_token_arguments": 0.8064516129032258, "train_accuracy_first_token_arguments_total": 31, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 1, "train_accuracy_recipient_token_restore": 0.0, "train_accuracy_recipient_token_restore_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.7707509881422925, "step": 260, "train_accuracy_first_token_arguments": 0.7413793103448276, "train_accuracy_first_token_arguments_total": 58, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_evaluate": 1.0, "train_accuracy_recipient_token_evaluate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.7707509881422925, "step": 260, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 29, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.7737154150197628, "grad_norm": 1.6094384600411795, "learning_rate": 6.410636082394772e-07, "loss": 0.2592, "step": 261 }, { "epoch": 0.7737154150197628, "step": 261, "train_accuracy_first_token_arguments": 0.7692307692307693, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.8, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_locate": 1.0, "train_accuracy_recipient_token_locate_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.7737154150197628, "step": 261, "train_accuracy_first_token_arguments": 0.7666666666666667, "train_accuracy_first_token_arguments_total": 30, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_ass": 1.0, "train_accuracy_recipient_token_ass_total": 1, "train_accuracy_recipient_token_environment": 1.0, "train_accuracy_recipient_token_environment_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_w": 1.0, "train_accuracy_recipient_token_w_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.7737154150197628, "step": 261, "train_accuracy_first_token_arguments": 0.975, "train_accuracy_first_token_arguments_total": 40, "train_accuracy_recipient_token": 0.9523809523809523, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_f": 1.0, "train_accuracy_recipient_token_f_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_locate": 1.0, "train_accuracy_recipient_token_locate_total": 1, "train_accuracy_recipient_token_recommend": 0.5, "train_accuracy_recipient_token_recommend_total": 2, "train_total_number_recipient_token": 21 }, { "epoch": 0.7766798418972332, "grad_norm": 1.997171944251961, "learning_rate": 6.250410068410367e-07, "loss": 0.2894, "step": 262 }, { "epoch": 0.7766798418972332, "step": 262, "train_accuracy_first_token_arguments": 0.9333333333333333, "train_accuracy_first_token_arguments_total": 30, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 17 }, { "epoch": 0.7766798418972332, "step": 262, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 11, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 16 }, { "epoch": 0.7766798418972332, "step": 262, "train_accuracy_first_token_arguments": 0.6956521739130435, "train_accuracy_first_token_arguments_total": 69, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 2, "train_accuracy_recipient_token_enh": 1.0, "train_accuracy_recipient_token_enh_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.7796442687747036, "grad_norm": 1.5690225372388775, "learning_rate": 6.091925271664156e-07, "loss": 0.1593, "step": 263 }, { "epoch": 0.7796442687747036, "step": 263, "train_accuracy_first_token_arguments": 0.75, "train_accuracy_first_token_arguments_total": 24, "train_accuracy_recipient_token": 0.85, "train_accuracy_recipient_token_all": 0.7692307692307693, "train_accuracy_recipient_token_all_total": 13, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.7796442687747036, "step": 263, "train_accuracy_first_token_arguments": 0.8333333333333334, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_locate": 1.0, "train_accuracy_recipient_token_locate_total": 2, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 2, "train_total_number_recipient_token": 20 }, { "epoch": 0.7796442687747036, "step": 263, "train_accuracy_first_token_arguments": 0.9285714285714286, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 20 }, { "epoch": 0.782608695652174, "grad_norm": 1.769966354853594, "learning_rate": 5.93519641015591e-07, "loss": 0.2617, "step": 264 }, { "epoch": 0.782608695652174, "step": 264, "train_accuracy_first_token_arguments": 0.9, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_stamp": 1.0, "train_accuracy_recipient_token_stamp_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.782608695652174, "step": 264, "train_accuracy_first_token_arguments": 0.6875, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.782608695652174, "step": 264, "train_accuracy_first_token_arguments": 0.8529411764705882, "train_accuracy_first_token_arguments_total": 34, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_city": 1.0, "train_accuracy_recipient_token_city_total": 1, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_urban": 1.0, "train_accuracy_recipient_token_urban_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.7855731225296443, "grad_norm": 2.0500780698786483, "learning_rate": 5.780238038817035e-07, "loss": 0.4104, "step": 265 }, { "epoch": 0.7855731225296443, "step": 265, "train_accuracy_first_token_arguments": 0.9444444444444444, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_gift": 1.0, "train_accuracy_recipient_token_gift_total": 1, "train_accuracy_recipient_token_personal": 1.0, "train_accuracy_recipient_token_personal_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.7855731225296443, "step": 265, "train_accuracy_first_token_arguments": 0.8333333333333334, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_accuracy_recipient_token_initialize": 0.0, "train_accuracy_recipient_token_initialize_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.7855731225296443, "step": 265, "train_accuracy_first_token_arguments": 0.782608695652174, "train_accuracy_first_token_arguments_total": 23, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 0.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_verify": 1.0, "train_accuracy_recipient_token_verify_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.7885375494071146, "grad_norm": 2.3088267661053083, "learning_rate": 5.627064548158903e-07, "loss": 0.215, "step": 266 }, { "epoch": 0.7885375494071146, "step": 266, "train_accuracy_first_token_arguments": 0.7692307692307693, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_building": 1.0, "train_accuracy_recipient_token_building_total": 1, "train_accuracy_recipient_token_flag": 1.0, "train_accuracy_recipient_token_flag_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.7885375494071146, "step": 266, "train_accuracy_first_token_arguments": 0.8333333333333334, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_deb": 1.0, "train_accuracy_recipient_token_deb_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.7885375494071146, "step": 266, "train_accuracy_first_token_arguments": 0.7692307692307693, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_building": 1.0, "train_accuracy_recipient_token_building_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.791501976284585, "grad_norm": 2.3252209612398764, "learning_rate": 5.47569016293649e-07, "loss": 0.2737, "step": 267 }, { "epoch": 0.791501976284585, "step": 267, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_configure": 1.0, "train_accuracy_recipient_token_configure_total": 5, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.791501976284585, "step": 267, "train_accuracy_first_token_arguments": 0.6363636363636364, "train_accuracy_first_token_arguments_total": 22, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.791501976284585, "step": 267, "train_accuracy_first_token_arguments": 0.8666666666666667, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 1, "train_accuracy_recipient_token_rel": 1.0, "train_accuracy_recipient_token_rel_total": 2, "train_accuracy_recipient_token_y": 1.0, "train_accuracy_recipient_token_y_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.7944664031620553, "grad_norm": 1.6931063369189647, "learning_rate": 5.326128940827313e-07, "loss": 0.172, "step": 268 }, { "epoch": 0.7944664031620553, "step": 268, "train_accuracy_first_token_arguments": 0.9166666666666666, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 20 }, { "epoch": 0.7944664031620553, "step": 268, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.7944664031620553, "step": 268, "train_accuracy_first_token_arguments": 0.9375, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 0.9047619047619048, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_find": 0.75, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_locate": 1.0, "train_accuracy_recipient_token_locate_total": 1, "train_total_number_recipient_token": 21 }, { "epoch": 0.7974308300395256, "grad_norm": 1.608131055033889, "learning_rate": 5.178394771125969e-07, "loss": 0.1891, "step": 269 }, { "epoch": 0.7974308300395256, "step": 269, "train_accuracy_first_token_arguments": 0.7058823529411765, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.8333333333333334, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_short": 1.0, "train_accuracy_recipient_token_short_total": 1, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.7974308300395256, "step": 269, "train_accuracy_first_token_arguments": 0.9047619047619048, "train_accuracy_first_token_arguments_total": 21, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_recipe": 1.0, "train_accuracy_recipient_token_recipe_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.7974308300395256, "step": 269, "train_accuracy_first_token_arguments": 0.9047619047619048, "train_accuracy_first_token_arguments_total": 21, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.8003952569169961, "grad_norm": 1.9940648204984408, "learning_rate": 5.032501373454266e-07, "loss": 0.2175, "step": 270 }, { "epoch": 0.8003952569169961, "step": 270, "train_accuracy_first_token_arguments": 0.9354838709677419, "train_accuracy_first_token_arguments_total": 31, "train_accuracy_recipient_token": 0.8125, "train_accuracy_recipient_token_all": 0.8333333333333334, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_crop": 1.0, "train_accuracy_recipient_token_crop_total": 1, "train_accuracy_recipient_token_f": 1.0, "train_accuracy_recipient_token_f_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_har": 0.0, "train_accuracy_recipient_token_har_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.8003952569169961, "step": 270, "train_accuracy_first_token_arguments": 0.9259259259259259, "train_accuracy_first_token_arguments_total": 27, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.8888888888888888, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_land": 1.0, "train_accuracy_recipient_token_land_total": 4, "train_accuracy_recipient_token_mount": 1.0, "train_accuracy_recipient_token_mount_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.8003952569169961, "step": 270, "train_accuracy_first_token_arguments": 0.9090909090909091, "train_accuracy_first_token_arguments_total": 11, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_go": 1.0, "train_accuracy_recipient_token_go_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.8033596837944664, "grad_norm": 1.941454727605692, "learning_rate": 4.888462296487129e-07, "loss": 0.2125, "step": 271 }, { "epoch": 0.8033596837944664, "step": 271, "train_accuracy_first_token_arguments": 0.9411764705882353, "train_accuracy_first_token_arguments_total": 34, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { "epoch": 0.8033596837944664, "step": 271, "train_accuracy_first_token_arguments": 0.7407407407407407, "train_accuracy_first_token_arguments_total": 27, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.8033596837944664, "step": 271, "train_accuracy_first_token_arguments": 0.7272727272727273, "train_accuracy_first_token_arguments_total": 33, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 2, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.8063241106719368, "grad_norm": 1.9052339985594693, "learning_rate": 4.746290916694368e-07, "loss": 0.2996, "step": 272 }, { "epoch": 0.8063241106719368, "step": 272, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 10, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 17 }, { "epoch": 0.8063241106719368, "step": 272, "train_accuracy_first_token_arguments": 0.8333333333333334, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.8063241106719368, "step": 272, "train_accuracy_first_token_arguments": 0.8181818181818182, "train_accuracy_first_token_arguments_total": 11, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.8092885375494071, "grad_norm": 2.933908602104762, "learning_rate": 4.6060004370984763e-07, "loss": 0.3296, "step": 273 }, { "epoch": 0.8092885375494071, "step": 273, "train_accuracy_first_token_arguments": 0.9629629629629629, "train_accuracy_first_token_arguments_total": 27, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_convert": 1.0, "train_accuracy_recipient_token_convert_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 17 }, { "epoch": 0.8092885375494071, "step": 273, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 9, "train_accuracy_recipient_token": 0.7857142857142857, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_data": 1.0, "train_accuracy_recipient_token_data_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 14 }, { "epoch": 0.8092885375494071, "step": 273, "train_accuracy_first_token_arguments": 0.9, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 19 }, { "epoch": 0.8122529644268774, "grad_norm": 2.787651925675771, "learning_rate": 4.467603886048452e-07, "loss": 0.2215, "step": 274 }, { "epoch": 0.8122529644268774, "step": 274, "train_accuracy_first_token_arguments": 0.9583333333333334, "train_accuracy_first_token_arguments_total": 24, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_h": 1.0, "train_accuracy_recipient_token_h_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.8122529644268774, "step": 274, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 0.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { "epoch": 0.8122529644268774, "step": 274, "train_accuracy_first_token_arguments": 0.9473684210526315, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_crop": 1.0, "train_accuracy_recipient_token_crop_total": 1, "train_accuracy_recipient_token_f": 1.0, "train_accuracy_recipient_token_f_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_har": 0.0, "train_accuracy_recipient_token_har_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.8152173913043478, "grad_norm": 2.0435081993694, "learning_rate": 4.331114116009938e-07, "loss": 0.1421, "step": 275 }, { "epoch": 0.8152173913043478, "step": 275, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 10, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 20 }, { "epoch": 0.8152173913043478, "step": 275, "train_accuracy_first_token_arguments": 0.9333333333333333, "train_accuracy_first_token_arguments_total": 30, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 20 }, { "epoch": 0.8152173913043478, "step": 275, "train_accuracy_first_token_arguments": 0.5925925925925926, "train_accuracy_first_token_arguments_total": 27, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_sum": 1.0, "train_accuracy_recipient_token_sum_total": 1, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.8181818181818182, "grad_norm": 2.30706005395235, "learning_rate": 4.196543802371641e-07, "loss": 0.2433, "step": 276 }, { "epoch": 0.8181818181818182, "step": 276, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 17 }, { "epoch": 0.8181818181818182, "step": 276, "train_accuracy_first_token_arguments": 0.7333333333333333, "train_accuracy_first_token_arguments_total": 45, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_rail": 1.0, "train_accuracy_recipient_token_rail_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.8181818181818182, "step": 276, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { "epoch": 0.8211462450592886, "grad_norm": 1.9811813651011425, "learning_rate": 4.063905442268201e-07, "loss": 0.2254, "step": 277 }, { "epoch": 0.8211462450592886, "step": 277, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 28, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 0.75, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.8211462450592886, "step": 277, "train_accuracy_first_token_arguments": 0.8095238095238095, "train_accuracy_first_token_arguments_total": 21, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_motor": 1.0, "train_accuracy_recipient_token_motor_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.8211462450592886, "step": 277, "train_accuracy_first_token_arguments": 0.9130434782608695, "train_accuracy_first_token_arguments_total": 23, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.8241106719367589, "grad_norm": 1.7163941442053223, "learning_rate": 3.9332113534196194e-07, "loss": 0.2561, "step": 278 }, { "epoch": 0.8241106719367589, "step": 278, "train_accuracy_first_token_arguments": 0.9444444444444444, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 0.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 0.875, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.8241106719367589, "step": 278, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 25, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 17 }, { "epoch": 0.8241106719367589, "step": 278, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_be": 0.6666666666666666, "train_accuracy_recipient_token_be_total": 3, "train_accuracy_recipient_token_find": 0.8, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_suggest": 1.0, "train_accuracy_recipient_token_suggest_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.8270750988142292, "grad_norm": 1.6379930568149086, "learning_rate": 3.80447367298738e-07, "loss": 0.1022, "step": 279 }, { "epoch": 0.8270750988142292, "step": 279, "train_accuracy_first_token_arguments": 0.8823529411764706, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_total_number_recipient_token": 18 }, { "epoch": 0.8270750988142292, "step": 279, "train_accuracy_first_token_arguments": 0.9047619047619048, "train_accuracy_first_token_arguments_total": 21, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_gr": 1.0, "train_accuracy_recipient_token_gr_total": 2, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_interpret": 1.0, "train_accuracy_recipient_token_interpret_total": 1, "train_accuracy_recipient_token_remove": 1.0, "train_accuracy_recipient_token_remove_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.8270750988142292, "step": 279, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 22, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.8300395256916996, "grad_norm": 1.6541675532013322, "learning_rate": 3.677704356447254e-07, "loss": 0.1926, "step": 280 }, { "epoch": 0.8300395256916996, "step": 280, "train_accuracy_first_token_arguments": 0.7903225806451613, "train_accuracy_first_token_arguments_total": 62, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_f": 1.0, "train_accuracy_recipient_token_f_total": 2, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_port": 1.0, "train_accuracy_recipient_token_port_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.8300395256916996, "step": 280, "train_accuracy_first_token_arguments": 0.9833333333333333, "train_accuracy_first_token_arguments_total": 120, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.8300395256916996, "step": 280, "train_accuracy_first_token_arguments": 0.8846153846153846, "train_accuracy_first_token_arguments_total": 26, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_convert": 1.0, "train_accuracy_recipient_token_convert_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_set": 1.0, "train_accuracy_recipient_token_set_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.8330039525691699, "grad_norm": 2.430973200540749, "learning_rate": 3.5529151764790715e-07, "loss": 0.282, "step": 281 }, { "epoch": 0.8330039525691699, "step": 281, "train_accuracy_first_token_arguments": 0.9166666666666666, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_aqu": 1.0, "train_accuracy_recipient_token_aqu_total": 1, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_fish": 1.0, "train_accuracy_recipient_token_fish_total": 3, "train_total_number_recipient_token": 16 }, { "epoch": 0.8330039525691699, "step": 281, "train_accuracy_first_token_arguments": 0.9393939393939394, "train_accuracy_first_token_arguments_total": 33, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9166666666666666, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_wild": 1.0, "train_accuracy_recipient_token_wild_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.8330039525691699, "step": 281, "train_accuracy_first_token_arguments": 0.8181818181818182, "train_accuracy_first_token_arguments_total": 11, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 20 }, { "epoch": 0.8359683794466403, "grad_norm": 2.0245615896487577, "learning_rate": 3.43011772187343e-07, "loss": 0.3178, "step": 282 }, { "epoch": 0.8359683794466403, "step": 282, "train_accuracy_first_token_arguments": 0.8947368421052632, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 18 }, { "epoch": 0.8359683794466403, "step": 282, "train_accuracy_first_token_arguments": 0.8571428571428571, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_pill": 1.0, "train_accuracy_recipient_token_pill_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.8359683794466403, "step": 282, "train_accuracy_first_token_arguments": 0.8947368421052632, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_evaluate": 1.0, "train_accuracy_recipient_token_evaluate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_plan": 1.0, "train_accuracy_recipient_token_plan_total": 1, "train_accuracy_recipient_token_suggest": 1.0, "train_accuracy_recipient_token_suggest_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.8389328063241107, "grad_norm": 2.1514655537590763, "learning_rate": 3.3093233964554464e-07, "loss": 0.2761, "step": 283 }, { "epoch": 0.8389328063241107, "step": 283, "train_accuracy_first_token_arguments": 0.9375, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 0.5, "train_accuracy_recipient_token_calculate_total": 2, "train_accuracy_recipient_token_estimate": 0.5, "train_accuracy_recipient_token_estimate_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_optimize": 1.0, "train_accuracy_recipient_token_optimize_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.8389328063241107, "step": 283, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 27, "train_accuracy_recipient_token": 0.75, "train_accuracy_recipient_token_all": 0.6363636363636364, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_conduct": 1.0, "train_accuracy_recipient_token_conduct_total": 1, "train_accuracy_recipient_token_d": 1.0, "train_accuracy_recipient_token_d_total": 1, "train_accuracy_recipient_token_landscape": 1.0, "train_accuracy_recipient_token_landscape_total": 1, "train_accuracy_recipient_token_organ": 1.0, "train_accuracy_recipient_token_organ_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.8389328063241107, "step": 283, "train_accuracy_first_token_arguments": 0.8421052631578947, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8888888888888888, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.841897233201581, "grad_norm": 2.2087100686535135, "learning_rate": 3.190543418025749e-07, "loss": 0.2386, "step": 284 }, { "epoch": 0.841897233201581, "step": 284, "train_accuracy_first_token_arguments": 0.8666666666666667, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_dec": 1.0, "train_accuracy_recipient_token_dec_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.841897233201581, "step": 284, "train_accuracy_first_token_arguments": 0.8571428571428571, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 0.875, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.841897233201581, "step": 284, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 10, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8888888888888888, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.8448616600790514, "grad_norm": 2.1863171999841655, "learning_rate": 3.0737888173187067e-07, "loss": 0.1506, "step": 285 }, { "epoch": 0.8448616600790514, "step": 285, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 11, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_l": 1.0, "train_accuracy_recipient_token_l_total": 1, "train_accuracy_recipient_token_remove": 1.0, "train_accuracy_recipient_token_remove_total": 1, "train_accuracy_recipient_token_write": 1.0, "train_accuracy_recipient_token_write_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.8448616600790514, "step": 285, "train_accuracy_first_token_arguments": 0.6666666666666666, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_detect": 1.0, "train_accuracy_recipient_token_detect_total": 1, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 19 }, { "epoch": 0.8448616600790514, "step": 285, "train_accuracy_first_token_arguments": 0.76, "train_accuracy_first_token_arguments_total": 25, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_accuracy_recipient_token_write": 1.0, "train_accuracy_recipient_token_write_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.8478260869565217, "grad_norm": 1.3956004649176736, "learning_rate": 2.9590704369780313e-07, "loss": 0.1174, "step": 286 }, { "epoch": 0.8478260869565217, "step": 286, "train_accuracy_first_token_arguments": 0.8235294117647058, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9166666666666666, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_paint": 1.0, "train_accuracy_recipient_token_paint_total": 1, "train_accuracy_recipient_token_rep": 1.0, "train_accuracy_recipient_token_rep_total": 1, "train_accuracy_recipient_token_surface": 1.0, "train_accuracy_recipient_token_surface_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.8478260869565217, "step": 286, "train_accuracy_first_token_arguments": 0.9333333333333333, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.8333333333333334, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.8478260869565217, "step": 286, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 20 }, { "epoch": 0.8507905138339921, "grad_norm": 1.8154026629925726, "learning_rate": 2.8463989305498596e-07, "loss": 0.2343, "step": 287 }, { "epoch": 0.8507905138339921, "step": 287, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.8507905138339921, "step": 287, "train_accuracy_first_token_arguments": 0.8260869565217391, "train_accuracy_first_token_arguments_total": 23, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.8507905138339921, "step": 287, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 10, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 16 }, { "epoch": 0.8537549407114624, "grad_norm": 2.1946988449253766, "learning_rate": 2.7357847614933876e-07, "loss": 0.2904, "step": 288 }, { "epoch": 0.8537549407114624, "step": 288, "train_accuracy_first_token_arguments": 0.8666666666666667, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_f": 1.0, "train_accuracy_recipient_token_f_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_mul": 1.0, "train_accuracy_recipient_token_mul_total": 1, "train_accuracy_recipient_token_weed": 1.0, "train_accuracy_recipient_token_weed_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.8537549407114624, "step": 288, "train_accuracy_first_token_arguments": 0.8846153846153846, "train_accuracy_first_token_arguments_total": 26, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_histor": 1.0, "train_accuracy_recipient_token_histor_total": 1, "train_accuracy_recipient_token_land": 1.0, "train_accuracy_recipient_token_land_total": 3, "train_total_number_recipient_token": 20 }, { "epoch": 0.8537549407114624, "step": 288, "train_accuracy_first_token_arguments": 0.85, "train_accuracy_first_token_arguments_total": 40, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9166666666666666, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_check": 0.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.8567193675889329, "grad_norm": 1.8692598511995904, "learning_rate": 2.6272382022091704e-07, "loss": 0.2141, "step": 289 }, { "epoch": 0.8567193675889329, "step": 289, "train_accuracy_first_token_arguments": 0.9310344827586207, "train_accuracy_first_token_arguments_total": 29, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_repair": 1.0, "train_accuracy_recipient_token_repair_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.8567193675889329, "step": 289, "train_accuracy_first_token_arguments": 0.75, "train_accuracy_first_token_arguments_total": 36, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_design": 1.0, "train_accuracy_recipient_token_design_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_play": 1.0, "train_accuracy_recipient_token_play_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.8567193675889329, "step": 289, "train_accuracy_first_token_arguments": 0.92, "train_accuracy_first_token_arguments_total": 25, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_train": 0.75, "train_accuracy_recipient_token_train_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.8596837944664032, "grad_norm": 1.4764081056636353, "learning_rate": 2.520769333085141e-07, "loss": 0.1611, "step": 290 }, { "epoch": 0.8596837944664032, "step": 290, "train_accuracy_first_token_arguments": 0.9626168224299065, "train_accuracy_first_token_arguments_total": 107, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_check": 0.5, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_mach": 0.6666666666666666, "train_accuracy_recipient_token_mach_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.8596837944664032, "step": 290, "train_accuracy_first_token_arguments": 0.8666666666666667, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.9, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 20 }, { "epoch": 0.8596837944664032, "step": 290, "train_accuracy_first_token_arguments": 0.75, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.8626482213438735, "grad_norm": 1.920082186381157, "learning_rate": 2.4163880415604913e-07, "loss": 0.266, "step": 291 }, { "epoch": 0.8626482213438735, "step": 291, "train_accuracy_first_token_arguments": 0.7333333333333333, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_pig": 1.0, "train_accuracy_recipient_token_pig_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.8626482213438735, "step": 291, "train_accuracy_first_token_arguments": 0.9680851063829787, "train_accuracy_first_token_arguments_total": 94, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_accuracy_recipient_token_track": 1.0, "train_accuracy_recipient_token_track_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.8626482213438735, "step": 291, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 10, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.8656126482213439, "grad_norm": 2.4496922640460923, "learning_rate": 2.3141040212074445e-07, "loss": 0.3954, "step": 292 }, { "epoch": 0.8656126482213439, "step": 292, "train_accuracy_first_token_arguments": 0.8, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_inspect": 1.0, "train_accuracy_recipient_token_inspect_total": 1, "train_accuracy_recipient_token_remove": 1.0, "train_accuracy_recipient_token_remove_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.8656126482213439, "step": 292, "train_accuracy_first_token_arguments": 0.8666666666666667, "train_accuracy_first_token_arguments_total": 30, "train_accuracy_recipient_token": 0.8260869565217391, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_multiply": 0.5, "train_accuracy_recipient_token_multiply_total": 4, "train_accuracy_recipient_token_suggest": 1.0, "train_accuracy_recipient_token_suggest_total": 2, "train_total_number_recipient_token": 23 }, { "epoch": 0.8656126482213439, "step": 292, "train_accuracy_first_token_arguments": 0.90625, "train_accuracy_first_token_arguments_total": 32, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 2, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_join": 0.0, "train_accuracy_recipient_token_join_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.8685770750988142, "grad_norm": 2.4189295709433183, "learning_rate": 2.2139267708310457e-07, "loss": 0.1385, "step": 293 }, { "epoch": 0.8685770750988142, "step": 293, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 10, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_locate": 1.0, "train_accuracy_recipient_token_locate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.8685770750988142, "step": 293, "train_accuracy_first_token_arguments": 0.9629629629629629, "train_accuracy_first_token_arguments_total": 27, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.8685770750988142, "step": 293, "train_accuracy_first_token_arguments": 0.8571428571428571, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_text": 1.0, "train_accuracy_recipient_token_text_total": 6, "train_total_number_recipient_token": 21 }, { "epoch": 0.8715415019762845, "grad_norm": 2.1088624243937426, "learning_rate": 2.1158655935870325e-07, "loss": 0.2126, "step": 294 }, { "epoch": 0.8715415019762845, "step": 294, "train_accuracy_first_token_arguments": 0.8947368421052632, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.8715415019762845, "step": 294, "train_accuracy_first_token_arguments": 0.7230769230769231, "train_accuracy_first_token_arguments_total": 65, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_draw": 1.0, "train_accuracy_recipient_token_draw_total": 5, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 20 }, { "epoch": 0.8715415019762845, "step": 294, "train_accuracy_first_token_arguments": 0.9444444444444444, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 20 }, { "epoch": 0.8745059288537549, "grad_norm": 2.032905445433328, "learning_rate": 2.0199295961178893e-07, "loss": 0.2617, "step": 295 }, { "epoch": 0.8745059288537549, "step": 295, "train_accuracy_first_token_arguments": 0.9166666666666666, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_locate": 1.0, "train_accuracy_recipient_token_locate_total": 1, "train_accuracy_recipient_token_suggest": 1.0, "train_accuracy_recipient_token_suggest_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.8745059288537549, "step": 295, "train_accuracy_first_token_arguments": 0.9210526315789473, "train_accuracy_first_token_arguments_total": 38, "train_accuracy_recipient_token": 0.9333333333333333, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_design": 1.0, "train_accuracy_recipient_token_design_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_extract": 0.5, "train_accuracy_recipient_token_extract_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 15 }, { "epoch": 0.8745059288537549, "step": 295, "train_accuracy_first_token_arguments": 0.7272727272727273, "train_accuracy_first_token_arguments_total": 22, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_att": 1.0, "train_accuracy_recipient_token_att_total": 1, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_explain": 1.0, "train_accuracy_recipient_token_explain_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.8774703557312253, "grad_norm": 2.284386546160239, "learning_rate": 1.9261276877071354e-07, "loss": 0.3743, "step": 296 }, { "epoch": 0.8774703557312253, "step": 296, "train_accuracy_first_token_arguments": 0.8, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_learn": 1.0, "train_accuracy_recipient_token_learn_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.8774703557312253, "step": 296, "train_accuracy_first_token_arguments": 0.9285714285714286, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.8774703557312253, "step": 296, "train_accuracy_first_token_arguments": 0.9069767441860465, "train_accuracy_first_token_arguments_total": 43, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8333333333333334, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 17 }, { "epoch": 0.8804347826086957, "grad_norm": 1.6291217601575605, "learning_rate": 1.8344685794519507e-07, "loss": 0.1641, "step": 297 }, { "epoch": 0.8804347826086957, "step": 297, "train_accuracy_first_token_arguments": 0.9649122807017544, "train_accuracy_first_token_arguments_total": 57, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.8804347826086957, "step": 297, "train_accuracy_first_token_arguments": 0.825, "train_accuracy_first_token_arguments_total": 40, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_accuracy_recipient_token_sort": 0.0, "train_accuracy_recipient_token_sort_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.8804347826086957, "step": 297, "train_accuracy_first_token_arguments": 0.8648648648648649, "train_accuracy_first_token_arguments_total": 37, "train_accuracy_recipient_token": 0.85, "train_accuracy_recipient_token_all": 0.9230769230769231, "train_accuracy_recipient_token_all_total": 13, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_restaurant": 1.0, "train_accuracy_recipient_token_restaurant_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.883399209486166, "grad_norm": 1.7794723846623823, "learning_rate": 1.744960783454186e-07, "loss": 0.2034, "step": 298 }, { "epoch": 0.883399209486166, "step": 298, "train_accuracy_first_token_arguments": 0.8421052631578947, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_accuracy_recipient_token_identify": 0.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.883399209486166, "step": 298, "train_accuracy_first_token_arguments": 0.7872340425531915, "train_accuracy_first_token_arguments_total": 47, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8333333333333334, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_building": 1.0, "train_accuracy_recipient_token_building_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_stat": 1.0, "train_accuracy_recipient_token_stat_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.883399209486166, "step": 298, "train_accuracy_first_token_arguments": 0.8837209302325582, "train_accuracy_first_token_arguments_total": 43, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compile": 1.0, "train_accuracy_recipient_token_compile_total": 1, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 1, "train_accuracy_recipient_token_restaurant": 1.0, "train_accuracy_recipient_token_restaurant_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.8863636363636364, "grad_norm": 1.9655423422483125, "learning_rate": 1.6576126120299046e-07, "loss": 0.3209, "step": 299 }, { "epoch": 0.8863636363636364, "step": 299, "train_accuracy_first_token_arguments": 0.8333333333333334, "train_accuracy_first_token_arguments_total": 30, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_draw": 1.0, "train_accuracy_recipient_token_draw_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 20 }, { "epoch": 0.8863636363636364, "step": 299, "train_accuracy_first_token_arguments": 0.7777777777777778, "train_accuracy_first_token_arguments_total": 9, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_exp": 1.0, "train_accuracy_recipient_token_exp_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_translate": 0.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.8863636363636364, "step": 299, "train_accuracy_first_token_arguments": 0.9473684210526315, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_plant": 1.0, "train_accuracy_recipient_token_plant_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.8893280632411067, "grad_norm": 2.3395853773637407, "learning_rate": 1.5724321769374023e-07, "loss": 0.3217, "step": 300 }, { "epoch": 0.8893280632411067, "step": 300, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 9, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.8893280632411067, "step": 300, "train_accuracy_first_token_arguments": 0.9, "train_accuracy_first_token_arguments_total": 30, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_bird": 1.0, "train_accuracy_recipient_token_bird_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.8893280632411067, "step": 300, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 10, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { "epoch": 0.892292490118577, "grad_norm": 2.265776160882912, "learning_rate": 1.4894273886239208e-07, "loss": 0.3318, "step": 301 }, { "epoch": 0.892292490118577, "step": 301, "train_accuracy_first_token_arguments": 0.9375, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { "epoch": 0.892292490118577, "step": 301, "train_accuracy_first_token_arguments": 0.8888888888888888, "train_accuracy_first_token_arguments_total": 9, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8333333333333334, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 18 }, { "epoch": 0.892292490118577, "step": 301, "train_accuracy_first_token_arguments": 0.9, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 1, "train_accuracy_recipient_token_stat": 1.0, "train_accuracy_recipient_token_stat_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.8952569169960475, "grad_norm": 2.3746950651302927, "learning_rate": 1.4086059554910186e-07, "loss": 0.3669, "step": 302 }, { "epoch": 0.8952569169960475, "step": 302, "train_accuracy_first_token_arguments": 0.6590909090909091, "train_accuracy_first_token_arguments_total": 44, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_object": 1.0, "train_accuracy_recipient_token_object_total": 1, "train_accuracy_recipient_token_people": 1.0, "train_accuracy_recipient_token_people_total": 2, "train_total_number_recipient_token": 20 }, { "epoch": 0.8952569169960475, "step": 302, "train_accuracy_first_token_arguments": 0.9285714285714286, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_verify": 1.0, "train_accuracy_recipient_token_verify_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.8952569169960475, "step": 302, "train_accuracy_first_token_arguments": 0.9090909090909091, "train_accuracy_first_token_arguments_total": 11, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_building": 1.0, "train_accuracy_recipient_token_building_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.8982213438735178, "grad_norm": 2.5356426721293994, "learning_rate": 1.3299753831787193e-07, "loss": 0.2913, "step": 303 }, { "epoch": 0.8982213438735178, "step": 303, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.8982213438735178, "step": 303, "train_accuracy_first_token_arguments": 0.8947368421052632, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 0.875, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.8982213438735178, "step": 303, "train_accuracy_first_token_arguments": 0.9230769230769231, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.9011857707509882, "grad_norm": 2.3231357183648993, "learning_rate": 1.2535429738684822e-07, "loss": 0.2729, "step": 304 }, { "epoch": 0.9011857707509882, "step": 304, "train_accuracy_first_token_arguments": 0.8571428571428571, "train_accuracy_first_token_arguments_total": 21, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.9011857707509882, "step": 304, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 6, "train_accuracy_recipient_token": 0.8, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_recogn": 1.0, "train_accuracy_recipient_token_recogn_total": 1, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 15 }, { "epoch": 0.9011857707509882, "step": 304, "train_accuracy_first_token_arguments": 0.9444444444444444, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 17 }, { "epoch": 0.9041501976284585, "grad_norm": 2.0889543788782716, "learning_rate": 1.1793158256050708e-07, "loss": 0.3167, "step": 305 }, { "epoch": 0.9041501976284585, "step": 305, "train_accuracy_first_token_arguments": 0.9285714285714286, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 20 }, { "epoch": 0.9041501976284585, "step": 305, "train_accuracy_first_token_arguments": 0.8059701492537313, "train_accuracy_first_token_arguments_total": 67, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_event": 1.0, "train_accuracy_recipient_token_event_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 16 }, { "epoch": 0.9041501976284585, "step": 305, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_car": 1.0, "train_accuracy_recipient_token_car_total": 1, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.9071146245059288, "grad_norm": 1.546141851099797, "learning_rate": 1.1073008316373812e-07, "loss": 0.1956, "step": 306 }, { "epoch": 0.9071146245059288, "step": 306, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 30, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_stat": 1.0, "train_accuracy_recipient_token_stat_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.9071146245059288, "step": 306, "train_accuracy_first_token_arguments": 0.9166666666666666, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.9071146245059288, "step": 306, "train_accuracy_first_token_arguments": 0.7857142857142857, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.9523809523809523, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 2, "train_accuracy_recipient_token_search": 0.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 21 }, { "epoch": 0.9100790513833992, "grad_norm": 2.2381733984414756, "learning_rate": 1.0375046797782868e-07, "loss": 0.2507, "step": 307 }, { "epoch": 0.9100790513833992, "step": 307, "train_accuracy_first_token_arguments": 0.8947368421052632, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 2, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.9100790513833992, "step": 307, "train_accuracy_first_token_arguments": 0.8, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_gr": 1.0, "train_accuracy_recipient_token_gr_total": 3, "train_accuracy_recipient_token_identify": 0.5, "train_accuracy_recipient_token_identify_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.9100790513833992, "step": 307, "train_accuracy_first_token_arguments": 0.9565217391304348, "train_accuracy_first_token_arguments_total": 23, "train_accuracy_recipient_token": 0.9523809523809523, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.9090909090909091, "train_accuracy_recipient_token_get_total": 11, "train_total_number_recipient_token": 21 }, { "epoch": 0.9130434782608695, "grad_norm": 1.6564222632639842, "learning_rate": 9.699338517835611e-08, "loss": 0.2674, "step": 308 }, { "epoch": 0.9130434782608695, "step": 308, "train_accuracy_first_token_arguments": 0.9166666666666666, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_verify": 1.0, "train_accuracy_recipient_token_verify_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.9130434782608695, "step": 308, "train_accuracy_first_token_arguments": 0.8571428571428571, "train_accuracy_first_token_arguments_total": 28, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_arr": 0.0, "train_accuracy_recipient_token_arr_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_house": 1.0, "train_accuracy_recipient_token_house_total": 3, "train_accuracy_recipient_token_landscape": 1.0, "train_accuracy_recipient_token_landscape_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.9130434782608695, "step": 308, "train_accuracy_first_token_arguments": 0.8125, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_market": 1.0, "train_accuracy_recipient_token_market_total": 3, "train_accuracy_recipient_token_verify": 1.0, "train_accuracy_recipient_token_verify_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.91600790513834, "grad_norm": 1.503775394527991, "learning_rate": 9.045946227499298e-08, "loss": 0.1276, "step": 309 }, { "epoch": 0.91600790513834, "step": 309, "train_accuracy_first_token_arguments": 0.8947368421052632, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 0.8888888888888888, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 20 }, { "epoch": 0.91600790513834, "step": 309, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 20 }, { "epoch": 0.91600790513834, "step": 309, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 11, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_research": 1.0, "train_accuracy_recipient_token_research_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.9189723320158103, "grad_norm": 2.2000639611250423, "learning_rate": 8.414930605323445e-08, "loss": 0.2596, "step": 310 }, { "epoch": 0.9189723320158103, "step": 310, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 4, "train_accuracy_recipient_token": 0.9285714285714286, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_extract": 0.5, "train_accuracy_recipient_token_extract_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 14 }, { "epoch": 0.9189723320158103, "step": 310, "train_accuracy_first_token_arguments": 0.7272727272727273, "train_accuracy_first_token_arguments_total": 11, "train_accuracy_recipient_token": 0.8125, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_explain": 0.6666666666666666, "train_accuracy_recipient_token_explain_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_suggest": 1.0, "train_accuracy_recipient_token_suggest_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.9189723320158103, "step": 310, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 23, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.9219367588932806, "grad_norm": 2.4294485936947194, "learning_rate": 7.806350251804484e-08, "loss": 0.3957, "step": 311 }, { "epoch": 0.9219367588932806, "step": 311, "train_accuracy_first_token_arguments": 0.7619047619047619, "train_accuracy_first_token_arguments_total": 42, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 16 }, { "epoch": 0.9219367588932806, "step": 311, "train_accuracy_first_token_arguments": 0.8235294117647058, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_property": 1.0, "train_accuracy_recipient_token_property_total": 1, "train_accuracy_recipient_token_schedule": 0.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.9219367588932806, "step": 311, "train_accuracy_first_token_arguments": 0.9444444444444444, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.924901185770751, "grad_norm": 2.1844882128572, "learning_rate": 7.220261683943935e-08, "loss": 0.26, "step": 312 }, { "epoch": 0.924901185770751, "step": 312, "train_accuracy_first_token_arguments": 0.6666666666666666, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_ap": 1.0, "train_accuracy_recipient_token_ap_total": 1, "train_accuracy_recipient_token_ger": 1.0, "train_accuracy_recipient_token_ger_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.924901185770751, "step": 312, "train_accuracy_first_token_arguments": 0.875, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_sand": 0.75, "train_accuracy_recipient_token_sand_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.924901185770751, "step": 312, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8888888888888888, "train_accuracy_recipient_token_get_total": 9, "train_accuracy_recipient_token_identify": 0.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.9278656126482213, "grad_norm": 1.842757768341881, "learning_rate": 6.6567193299997e-08, "loss": 0.1724, "step": 313 }, { "epoch": 0.9278656126482213, "step": 313, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 8, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 16 }, { "epoch": 0.9278656126482213, "step": 313, "train_accuracy_first_token_arguments": 0.5833333333333334, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_arr": 0.0, "train_accuracy_recipient_token_arr_total": 1, "train_accuracy_recipient_token_marine": 1.0, "train_accuracy_recipient_token_marine_total": 1, "train_accuracy_recipient_token_o": 1.0, "train_accuracy_recipient_token_o_total": 1, "train_accuracy_recipient_token_under": 1.0, "train_accuracy_recipient_token_under_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.9278656126482213, "step": 313, "train_accuracy_first_token_arguments": 0.8571428571428571, "train_accuracy_first_token_arguments_total": 35, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 0.5, "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.9308300395256917, "grad_norm": 2.101888459848309, "learning_rate": 6.115775524431711e-08, "loss": 0.3051, "step": 314 }, { "epoch": 0.9308300395256917, "step": 314, "train_accuracy_first_token_arguments": 0.9090909090909091, "train_accuracy_first_token_arguments_total": 22, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_plan": 1.0, "train_accuracy_recipient_token_plan_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.9308300395256917, "step": 314, "train_accuracy_first_token_arguments": 0.7857142857142857, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_stat": 0.5, "train_accuracy_recipient_token_stat_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.9308300395256917, "step": 314, "train_accuracy_first_token_arguments": 0.9411764705882353, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_arch": 1.0, "train_accuracy_recipient_token_arch_total": 2, "train_accuracy_recipient_token_building": 1.0, "train_accuracy_recipient_token_building_total": 3, "train_accuracy_recipient_token_gr": 1.0, "train_accuracy_recipient_token_gr_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.9337944664031621, "grad_norm": 1.8025117128414663, "learning_rate": 5.597480503041486e-08, "loss": 0.3276, "step": 315 }, { "epoch": 0.9337944664031621, "step": 315, "train_accuracy_first_token_arguments": 0.7647058823529411, "train_accuracy_first_token_arguments_total": 102, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_inter": 1.0, "train_accuracy_recipient_token_inter_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.9337944664031621, "step": 315, "train_accuracy_first_token_arguments": 0.8461538461538461, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.9337944664031621, "step": 315, "train_accuracy_first_token_arguments": 0.9473684210526315, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_fabric": 1.0, "train_accuracy_recipient_token_fabric_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_send": 1.0, "train_accuracy_recipient_token_send_total": 1, "train_accuracy_recipient_token_set": 1.0, "train_accuracy_recipient_token_set_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.9367588932806324, "grad_norm": 2.303122934120546, "learning_rate": 5.1018823983070285e-08, "loss": 0.2539, "step": 316 }, { "epoch": 0.9367588932806324, "step": 316, "train_accuracy_first_token_arguments": 0.7, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { "epoch": 0.9367588932806324, "step": 316, "train_accuracy_first_token_arguments": 0.9166666666666666, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_format": 1.0, "train_accuracy_recipient_token_format_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_sum": 1.0, "train_accuracy_recipient_token_sum_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.9367588932806324, "step": 316, "train_accuracy_first_token_arguments": 0.7, "train_accuracy_first_token_arguments_total": 10, "train_accuracy_recipient_token": 0.7777777777777778, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_arr": 0.0, "train_accuracy_recipient_token_arr_total": 1, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.9397233201581028, "grad_norm": 1.7793933316663082, "learning_rate": 4.629027234912986e-08, "loss": 0.1763, "step": 317 }, { "epoch": 0.9397233201581028, "step": 317, "train_accuracy_first_token_arguments": 0.95, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9166666666666666, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_compare": 0.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.9397233201581028, "step": 317, "train_accuracy_first_token_arguments": 0.7428571428571429, "train_accuracy_first_token_arguments_total": 35, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_explain": 1.0, "train_accuracy_recipient_token_explain_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_plane": 1.0, "train_accuracy_recipient_token_plane_total": 3, "train_accuracy_recipient_token_zi": 0.0, "train_accuracy_recipient_token_zi_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.9397233201581028, "step": 317, "train_accuracy_first_token_arguments": 0.8888888888888888, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_arr": 1.0, "train_accuracy_recipient_token_arr_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.9426877470355731, "grad_norm": 1.5774943505040582, "learning_rate": 4.178958925476401e-08, "loss": 0.1834, "step": 318 }, { "epoch": 0.9426877470355731, "eval_accuracy_first_token_arguments": 0.9248395967002749, "eval_accuracy_first_token_arguments_total": 1091, "eval_accuracy_recipient_token": 0.9401330376940134, "eval_accuracy_recipient_token_all": 0.9220272904483431, "eval_accuracy_recipient_token_all_total": 513, "eval_accuracy_recipient_token_calculate": 1.0, "eval_accuracy_recipient_token_calculate_total": 13, "eval_accuracy_recipient_token_find": 1.0, "eval_accuracy_recipient_token_find_total": 57, "eval_accuracy_recipient_token_get": 0.9795918367346939, "eval_accuracy_recipient_token_get_total": 196, "eval_accuracy_recipient_token_identify": 0.8333333333333334, "eval_accuracy_recipient_token_identify_total": 6, "eval_loss": 0.21341286599636078, "eval_perplexity": 1.0333647517086502, "eval_runtime": 112.5857, "eval_samples_per_second": 0.897, "eval_steps_per_second": 0.453, "eval_total_number_recipient_token": 902, "step": 318 }, { "epoch": 0.9426877470355731, "step": 318, "train_accuracy_first_token_arguments": 0.7777777777777778, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_identify": 0.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.9426877470355731, "step": 318, "train_accuracy_first_token_arguments": 0.9375, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { "epoch": 0.9426877470355731, "step": 318, "train_accuracy_first_token_arguments": 0.8461538461538461, "train_accuracy_first_token_arguments_total": 26, "train_accuracy_recipient_token": 0.85, "train_accuracy_recipient_token_all": 0.75, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.9456521739130435, "grad_norm": 2.4406440061460377, "learning_rate": 3.7517192664685844e-08, "loss": 0.3195, "step": 319 }, { "epoch": 0.9456521739130435, "step": 319, "train_accuracy_first_token_arguments": 0.9375, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.9456521739130435, "step": 319, "train_accuracy_first_token_arguments": 0.7142857142857143, "train_accuracy_first_token_arguments_total": 21, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 2, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.9456521739130435, "step": 319, "train_accuracy_first_token_arguments": 0.9375, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_plant": 1.0, "train_accuracy_recipient_token_plant_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.9486166007905138, "grad_norm": 2.041210168988887, "learning_rate": 3.347347934333778e-08, "loss": 0.2436, "step": 320 }, { "epoch": 0.9486166007905138, "step": 320, "train_accuracy_first_token_arguments": 0.9375, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 0.85, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 15, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 20 }, { "epoch": 0.9486166007905138, "step": 320, "train_accuracy_first_token_arguments": 0.5925925925925926, "train_accuracy_first_token_arguments_total": 54, "train_accuracy_recipient_token": 0.8666666666666667, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_predict": 1.0, "train_accuracy_recipient_token_predict_total": 1, "train_accuracy_recipient_token_stamp": 0.5, "train_accuracy_recipient_token_stamp_total": 2, "train_total_number_recipient_token": 15 }, { "epoch": 0.9486166007905138, "step": 320, "train_accuracy_first_token_arguments": 0.9565217391304348, "train_accuracy_first_token_arguments_total": 23, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.9515810276679841, "grad_norm": 2.151431626558894, "learning_rate": 2.9658824818044328e-08, "loss": 0.309, "step": 321 }, { "epoch": 0.9515810276679841, "step": 321, "train_accuracy_first_token_arguments": 0.9811320754716981, "train_accuracy_first_token_arguments_total": 53, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_optimize": 0.0, "train_accuracy_recipient_token_optimize_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.9515810276679841, "step": 321, "train_accuracy_first_token_arguments": 0.90625, "train_accuracy_first_token_arguments_total": 32, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.9515810276679841, "step": 321, "train_accuracy_first_token_arguments": 0.9545454545454546, "train_accuracy_first_token_arguments_total": 22, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 16 }, { "epoch": 0.9545454545454546, "grad_norm": 2.5639718605032216, "learning_rate": 2.607358334413779e-08, "loss": 0.3681, "step": 322 }, { "epoch": 0.9545454545454546, "step": 322, "train_accuracy_first_token_arguments": 0.8148148148148148, "train_accuracy_first_token_arguments_total": 27, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_ne": 1.0, "train_accuracy_recipient_token_ne_total": 3, "train_total_number_recipient_token": 20 }, { "epoch": 0.9545454545454546, "step": 322, "train_accuracy_first_token_arguments": 0.8888888888888888, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 0.9333333333333333, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_convert": 1.0, "train_accuracy_recipient_token_convert_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 15 }, { "epoch": 0.9545454545454546, "step": 322, "train_accuracy_first_token_arguments": 0.9, "train_accuracy_first_token_arguments_total": 30, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_customize": 1.0, "train_accuracy_recipient_token_customize_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.9575098814229249, "grad_norm": 2.8167178374609607, "learning_rate": 2.2718087872060925e-08, "loss": 0.2343, "step": 323 }, { "epoch": 0.9575098814229249, "step": 323, "train_accuracy_first_token_arguments": 0.6744186046511628, "train_accuracy_first_token_arguments_total": 172, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_pool": 1.0, "train_accuracy_recipient_token_pool_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.9575098814229249, "step": 323, "train_accuracy_first_token_arguments": 0.9285714285714286, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_robot": 1.0, "train_accuracy_recipient_token_robot_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.9575098814229249, "step": 323, "train_accuracy_first_token_arguments": 0.9583333333333334, "train_accuracy_first_token_arguments_total": 24, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 0.875, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.9604743083003953, "grad_norm": 1.6492281178442556, "learning_rate": 1.9592650016444503e-08, "loss": 0.1848, "step": 324 }, { "epoch": 0.9604743083003953, "step": 324, "train_accuracy_first_token_arguments": 0.896551724137931, "train_accuracy_first_token_arguments_total": 29, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_learn": 1.0, "train_accuracy_recipient_token_learn_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.9604743083003953, "step": 324, "train_accuracy_first_token_arguments": 0.9166666666666666, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 0.8888888888888888, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 20 }, { "epoch": 0.9604743083003953, "step": 324, "train_accuracy_first_token_arguments": 0.5384615384615384, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 0.9047619047619048, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_count": 0.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_subtract": 1.0, "train_accuracy_recipient_token_subtract_total": 1, "train_accuracy_recipient_token_track": 1.0, "train_accuracy_recipient_token_track_total": 1, "train_total_number_recipient_token": 21 }, { "epoch": 0.9634387351778656, "grad_norm": 1.754485036239026, "learning_rate": 1.6697560027171543e-08, "loss": 0.1782, "step": 325 }, { "epoch": 0.9634387351778656, "step": 325, "train_accuracy_first_token_arguments": 0.9642857142857143, "train_accuracy_first_token_arguments_total": 28, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 0.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_compile": 1.0, "train_accuracy_recipient_token_compile_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.9634387351778656, "step": 325, "train_accuracy_first_token_arguments": 0.7857142857142857, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_extract": 1.0, "train_accuracy_recipient_token_extract_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_recogn": 0.0, "train_accuracy_recipient_token_recogn_total": 1, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.9634387351778656, "step": 325, "train_accuracy_first_token_arguments": 0.5333333333333333, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.9333333333333333, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_verify": 1.0, "train_accuracy_recipient_token_verify_total": 1, "train_total_number_recipient_token": 15 }, { "epoch": 0.9664031620553359, "grad_norm": 2.1745450667504445, "learning_rate": 1.4033086762419989e-08, "loss": 0.3463, "step": 326 }, { "epoch": 0.9664031620553359, "step": 326, "train_accuracy_first_token_arguments": 0.9322033898305084, "train_accuracy_first_token_arguments_total": 59, "train_accuracy_recipient_token": 0.7777777777777778, "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_check": 0.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.9664031620553359, "step": 326, "train_accuracy_first_token_arguments": 0.9375, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compile": 1.0, "train_accuracy_recipient_token_compile_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.9664031620553359, "step": 326, "train_accuracy_first_token_arguments": 0.6875, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_download": 1.0, "train_accuracy_recipient_token_download_total": 1, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_search": 0.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.9693675889328063, "grad_norm": 2.1355214484085425, "learning_rate": 1.1599477663696845e-08, "loss": 0.246, "step": 327 }, { "epoch": 0.9693675889328063, "step": 327, "train_accuracy_first_token_arguments": 0.9230769230769231, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.9693675889328063, "step": 327, "train_accuracy_first_token_arguments": 0.8125, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_deb": 1.0, "train_accuracy_recipient_token_deb_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_plan": 1.0, "train_accuracy_recipient_token_plan_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.9693675889328063, "step": 327, "train_accuracy_first_token_arguments": 0.7727272727272727, "train_accuracy_first_token_arguments_total": 22, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.9723320158102767, "grad_norm": 1.6782990768747863, "learning_rate": 9.396958732856843e-09, "loss": 0.2385, "step": 328 }, { "epoch": 0.9723320158102767, "step": 328, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 29, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 0.75, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.9723320158102767, "step": 328, "train_accuracy_first_token_arguments": 0.85, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.9723320158102767, "step": 328, "train_accuracy_first_token_arguments": 0.631578947368421, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_park": 1.0, "train_accuracy_recipient_token_park_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.9752964426877471, "grad_norm": 2.45983910491564, "learning_rate": 7.425734511117e-09, "loss": 0.3068, "step": 329 }, { "epoch": 0.9752964426877471, "step": 329, "train_accuracy_first_token_arguments": 0.95, "train_accuracy_first_token_arguments_total": 40, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_explain": 1.0, "train_accuracy_recipient_token_explain_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.9752964426877471, "step": 329, "train_accuracy_first_token_arguments": 0.95, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.9752964426877471, "step": 329, "train_accuracy_first_token_arguments": 0.8666666666666667, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.9782608695652174, "grad_norm": 1.9407711881422764, "learning_rate": 5.685988060059045e-09, "loss": 0.2393, "step": 330 }, { "epoch": 0.9782608695652174, "step": 330, "train_accuracy_first_token_arguments": 0.8571428571428571, "train_accuracy_first_token_arguments_total": 28, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8888888888888888, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.9782608695652174, "step": 330, "train_accuracy_first_token_arguments": 0.7142857142857143, "train_accuracy_first_token_arguments_total": 21, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_filter": 1.0, "train_accuracy_recipient_token_filter_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.9782608695652174, "step": 330, "train_accuracy_first_token_arguments": 0.9166666666666666, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_motor": 1.0, "train_accuracy_recipient_token_motor_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.9812252964426877, "grad_norm": 1.9607276731011434, "learning_rate": 4.17788094463023e-09, "loss": 0.1726, "step": 331 }, { "epoch": 0.9812252964426877, "step": 331, "train_accuracy_first_token_arguments": 0.8666666666666667, "train_accuracy_first_token_arguments_total": 15, "train_accuracy_recipient_token": 0.6842105263157895, "train_accuracy_recipient_token_all": 0.6666666666666666, "train_accuracy_recipient_token_all_total": 15, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.9812252964426877, "step": 331, "train_accuracy_first_token_arguments": 0.8571428571428571, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_building": 1.0, "train_accuracy_recipient_token_building_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.9812252964426877, "step": 331, "train_accuracy_first_token_arguments": 0.8, "train_accuracy_first_token_arguments_total": 30, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_schedule": 0.5, "train_accuracy_recipient_token_schedule_total": 2, "train_accuracy_recipient_token_table": 0.5, "train_accuracy_recipient_token_table_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.9841897233201581, "grad_norm": 1.9808670794420047, "learning_rate": 2.9015532181397854e-09, "loss": 0.2188, "step": 332 }, { "epoch": 0.9841897233201581, "step": 332, "train_accuracy_first_token_arguments": 0.8421052631578947, "train_accuracy_first_token_arguments_total": 19, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_generate": 0.5, "train_accuracy_recipient_token_generate_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_sum": 1.0, "train_accuracy_recipient_token_sum_total": 1, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.9841897233201581, "step": 332, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 14, "train_accuracy_recipient_token": 0.7857142857142857, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_evaluate": 0.0, "train_accuracy_recipient_token_evaluate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 14 }, { "epoch": 0.9841897233201581, "step": 332, "train_accuracy_first_token_arguments": 0.96875, "train_accuracy_first_token_arguments_total": 64, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_property": 1.0, "train_accuracy_recipient_token_property_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.9871541501976284, "grad_norm": 2.0556535601550325, "learning_rate": 1.857123409250705e-09, "loss": 0.2487, "step": 333 }, { "epoch": 0.9871541501976284, "step": 333, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_explain": 1.0, "train_accuracy_recipient_token_explain_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 18 }, { "epoch": 0.9871541501976284, "step": 333, "train_accuracy_first_token_arguments": 0.8571428571428571, "train_accuracy_first_token_arguments_total": 21, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 2, "train_accuracy_recipient_token_calculate": 0.5, "train_accuracy_recipient_token_calculate_total": 2, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.9871541501976284, "step": 333, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 17, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_artifact": 1.0, "train_accuracy_recipient_token_artifact_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.9901185770750988, "grad_norm": 2.092295914240381, "learning_rate": 1.0446885109746673e-09, "loss": 0.1504, "step": 334 }, { "epoch": 0.9901185770750988, "step": 334, "train_accuracy_first_token_arguments": 0.8823529411764706, "train_accuracy_first_token_arguments_total": 34, "train_accuracy_recipient_token": 0.7222222222222222, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_schedule": 0.0, "train_accuracy_recipient_token_schedule_total": 1, "train_accuracy_recipient_token_track": 0.0, "train_accuracy_recipient_token_track_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.9901185770750988, "step": 334, "train_accuracy_first_token_arguments": 0.9230769230769231, "train_accuracy_first_token_arguments_total": 13, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.9901185770750988, "step": 334, "train_accuracy_first_token_arguments": 0.9166666666666666, "train_accuracy_first_token_arguments_total": 12, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.9930830039525692, "grad_norm": 2.0272702589635965, "learning_rate": 4.6432397166285e-10, "loss": 0.1775, "step": 335 }, { "epoch": 0.9930830039525692, "step": 335, "train_accuracy_first_token_arguments": 0.4375, "train_accuracy_first_token_arguments_total": 32, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_write": 1.0, "train_accuracy_recipient_token_write_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.9930830039525692, "step": 335, "train_accuracy_first_token_arguments": 0.8769230769230769, "train_accuracy_first_token_arguments_total": 65, "train_accuracy_recipient_token": 0.7222222222222222, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.3333333333333333, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_optimize": 0.0, "train_accuracy_recipient_token_optimize_total": 1, "train_accuracy_recipient_token_plan": 0.0, "train_accuracy_recipient_token_plan_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.9930830039525692, "step": 335, "train_accuracy_first_token_arguments": 0.85, "train_accuracy_first_token_arguments_total": 20, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 15, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.9960474308300395, "grad_norm": 1.9403721941845617, "learning_rate": 1.160836880001459e-10, "loss": 0.3228, "step": 336 }, { "epoch": 0.9960474308300395, "step": 336, "train_accuracy_first_token_arguments": 0.868421052631579, "train_accuracy_first_token_arguments_total": 38, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.8333333333333334, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 2, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.9960474308300395, "step": 336, "train_accuracy_first_token_arguments": 0.9375, "train_accuracy_first_token_arguments_total": 16, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_explain": 1.0, "train_accuracy_recipient_token_explain_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.9960474308300395, "step": 336, "train_accuracy_first_token_arguments": 1.0, "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.8, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 20 }, { "epoch": 0.9990118577075099, "grad_norm": 1.5288849501967092, "learning_rate": 0.0, "loss": 0.1852, "step": 337 }, { "epoch": 0.9990118577075099, "step": 337, "total_flos": 337785957384192.0, "train_loss": 0.2736482885984002, "train_runtime": 6297.3, "train_samples_per_second": 0.321, "train_steps_per_second": 0.054 } ], "logging_steps": 1.0, "max_steps": 337, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100.0, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": false, "should_training_stop": false }, "attributes": {} } }, "total_flos": 337785957384192.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }