diff --git "a/trainer_state.json" "b/trainer_state.json" --- "a/trainer_state.json" +++ "b/trainer_state.json" @@ -1,9 +1,9 @@ { "best_metric": null, "best_model_checkpoint": null, - "epoch": 0.9990118577075099, - "eval_steps": 106, - "global_step": 337, + "epoch": 1.0, + "eval_steps": 168, + "global_step": 506, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, @@ -11,8 +11,6 @@ { "epoch": 0, "step": 0, - "train_accuracy_first_token_arguments": 0.7777777777777778, - "train_accuracy_first_token_arguments_total": 18, "train_accuracy_recipient_token": 0.8, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, @@ -29,73 +27,72 @@ { "epoch": 0, "step": 0, - "train_accuracy_first_token_arguments": 0.6129032258064516, - "train_accuracy_first_token_arguments_total": 31, "train_accuracy_recipient_token": 0.6111111111111112, "train_accuracy_recipient_token_all": 0.6, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, - "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_get": 0.3333333333333333, + "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_identify": 0.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 0.0, + "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0, - "step": 0, - "train_accuracy_first_token_arguments": 0.5555555555555556, - "train_accuracy_first_token_arguments_total": 18, - "train_accuracy_recipient_token": 0.6842105263157895, + "epoch": 0.001976284584980237, + "grad_norm": 2.82898798080985, + "learning_rate": 3.125e-07, + "loss": 0.2927, + "step": 1 + }, + { + "epoch": 0.001976284584980237, + "step": 1, + "train_accuracy_recipient_token": 0.5789473684210527, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_histor": 0.5, + "train_accuracy_recipient_token_histor": 0.0, "train_accuracy_recipient_token_histor_total": 2, - "train_accuracy_recipient_token_med": 1.0, + "train_accuracy_recipient_token_med": 0.5, "train_accuracy_recipient_token_med_total": 2, "train_total_number_recipient_token": 19 }, { - "epoch": 0.0029644268774703555, - "grad_norm": 3.2711954063798103, - "learning_rate": 4.5454545454545457e-07, - "loss": 0.3754, - "step": 1 - }, - { - "epoch": 0.0029644268774703555, + "epoch": 0.001976284584980237, "step": 1, - "train_accuracy_first_token_arguments": 0.5245901639344263, - "train_accuracy_first_token_arguments_total": 61, - "train_accuracy_recipient_token": 0.6, - "train_accuracy_recipient_token_all": 0.6, + "train_accuracy_recipient_token": 0.65, + "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 0.5, "train_accuracy_recipient_token_estimate_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_prior": 0.0, - "train_accuracy_recipient_token_prior_total": 1, + "train_accuracy_recipient_token_p": 0.0, + "train_accuracy_recipient_token_p_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 20 }, { - "epoch": 0.0029644268774703555, - "step": 1, - "train_accuracy_first_token_arguments": 0.8076923076923077, - "train_accuracy_first_token_arguments_total": 26, - "train_accuracy_recipient_token": 0.6666666666666666, + "epoch": 0.003952569169960474, + "grad_norm": 2.796922585421834, + "learning_rate": 6.25e-07, + "loss": 0.3515, + "step": 2 + }, + { + "epoch": 0.003952569169960474, + "step": 2, + "train_accuracy_recipient_token": 0.5555555555555556, "train_accuracy_recipient_token_all": 0.6, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find": 0.3333333333333333, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, @@ -104,16 +101,14 @@ "train_total_number_recipient_token": 18 }, { - "epoch": 0.0029644268774703555, - "step": 1, - "train_accuracy_first_token_arguments": 0.8181818181818182, - "train_accuracy_first_token_arguments_total": 22, - "train_accuracy_recipient_token": 0.7894736842105263, - "train_accuracy_recipient_token_all": 0.8, + "epoch": 0.003952569169960474, + "step": 2, + "train_accuracy_recipient_token": 0.6842105263157895, + "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_find": 0.75, + "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, @@ -123,16 +118,14 @@ }, { "epoch": 0.005928853754940711, - "grad_norm": 2.9342358559874504, - "learning_rate": 9.090909090909091e-07, - "loss": 0.2983, - "step": 2 + "grad_norm": 2.1200152439803803, + "learning_rate": 9.375000000000001e-07, + "loss": 0.1468, + "step": 3 }, { "epoch": 0.005928853754940711, - "step": 2, - "train_accuracy_first_token_arguments": 0.9166666666666666, - "train_accuracy_first_token_arguments_total": 12, + "step": 3, "train_accuracy_recipient_token": 0.7368421052631579, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, @@ -144,9 +137,7 @@ }, { "epoch": 0.005928853754940711, - "step": 2, - "train_accuracy_first_token_arguments": 0.6666666666666666, - "train_accuracy_first_token_arguments_total": 24, + "step": 3, "train_accuracy_recipient_token": 0.8, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, @@ -159,10 +150,15 @@ "train_total_number_recipient_token": 20 }, { - "epoch": 0.005928853754940711, - "step": 2, - "train_accuracy_first_token_arguments": 0.8387096774193549, - "train_accuracy_first_token_arguments_total": 31, + "epoch": 0.007905138339920948, + "grad_norm": 3.190242133778613, + "learning_rate": 1.25e-06, + "loss": 0.2953, + "step": 4 + }, + { + "epoch": 0.007905138339920948, + "step": 4, "train_accuracy_recipient_token": 0.7368421052631579, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, @@ -177,23 +173,14 @@ "train_total_number_recipient_token": 19 }, { - "epoch": 0.008893280632411068, - "grad_norm": 3.7666597300364524, - "learning_rate": 1.3636363636363636e-06, - "loss": 0.3456, - "step": 3 - }, - { - "epoch": 0.008893280632411068, - "step": 3, - "train_accuracy_first_token_arguments": 0.7241379310344828, - "train_accuracy_first_token_arguments_total": 29, - "train_accuracy_recipient_token": 0.6666666666666666, + "epoch": 0.007905138339920948, + "step": 4, + "train_accuracy_recipient_token": 0.6111111111111112, "train_accuracy_recipient_token_all": 0.6, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, - "train_accuracy_recipient_token_find": 0.6666666666666666, + "train_accuracy_recipient_token_find": 0.3333333333333333, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, @@ -202,14 +189,19 @@ "train_total_number_recipient_token": 18 }, { - "epoch": 0.008893280632411068, - "step": 3, - "train_accuracy_first_token_arguments": 0.7333333333333333, - "train_accuracy_first_token_arguments_total": 15, + "epoch": 0.009881422924901186, + "grad_norm": 3.13829102322951, + "learning_rate": 1.5625e-06, + "loss": 0.2338, + "step": 5 + }, + { + "epoch": 0.009881422924901186, + "step": 5, "train_accuracy_recipient_token": 0.631578947368421, "train_accuracy_recipient_token_all": 0.6, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, @@ -220,33 +212,29 @@ "train_total_number_recipient_token": 19 }, { - "epoch": 0.008893280632411068, - "step": 3, - "train_accuracy_first_token_arguments": 0.7058823529411765, - "train_accuracy_first_token_arguments_total": 17, - "train_accuracy_recipient_token": 0.55, - "train_accuracy_recipient_token_all": 0.7, + "epoch": 0.009881422924901186, + "step": 5, + "train_accuracy_recipient_token": 0.7, + "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_count": 0.0, + "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, - "train_accuracy_recipient_token_get": 0.5714285714285714, + "train_accuracy_recipient_token_get": 0.7142857142857143, "train_accuracy_recipient_token_get_total": 7, - "train_accuracy_recipient_token_identify": 0.0, - "train_accuracy_recipient_token_identify_total": 2, + "train_accuracy_recipient_token_ident": 0.0, + "train_accuracy_recipient_token_ident_total": 2, "train_total_number_recipient_token": 20 }, { "epoch": 0.011857707509881422, - "grad_norm": 4.455480512503962, - "learning_rate": 1.8181818181818183e-06, - "loss": 0.3511, - "step": 4 + "grad_norm": 3.8032606174393555, + "learning_rate": 1.8750000000000003e-06, + "loss": 0.2693, + "step": 6 }, { "epoch": 0.011857707509881422, - "step": 4, - "train_accuracy_first_token_arguments": 0.975609756097561, - "train_accuracy_first_token_arguments_total": 41, + "step": 6, "train_accuracy_recipient_token": 0.6, "train_accuracy_recipient_token_all": 0.5, "train_accuracy_recipient_token_all_total": 8, @@ -262,14 +250,12 @@ }, { "epoch": 0.011857707509881422, - "step": 4, - "train_accuracy_first_token_arguments": 0.8125, - "train_accuracy_first_token_arguments_total": 16, - "train_accuracy_recipient_token": 0.6, + "step": 6, + "train_accuracy_recipient_token": 0.6666666666666666, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 0.5, - "train_accuracy_recipient_token_analyze_total": 2, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 2, "train_accuracy_recipient_token_get": 0.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_locate": 0.0, @@ -279,75 +265,74 @@ "train_total_number_recipient_token": 15 }, { - "epoch": 0.011857707509881422, - "step": 4, - "train_accuracy_first_token_arguments": 0.75, - "train_accuracy_first_token_arguments_total": 20, - "train_accuracy_recipient_token": 0.5625, - "train_accuracy_recipient_token_all": 0.6, + "epoch": 0.01383399209486166, + "grad_norm": 4.597413898848223, + "learning_rate": 2.1875000000000002e-06, + "loss": 0.3544, + "step": 7 + }, + { + "epoch": 0.01383399209486166, + "step": 7, + "train_accuracy_recipient_token": 0.5, + "train_accuracy_recipient_token_all": 0.4, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 0.0, - "train_accuracy_recipient_token_analyze_total": 2, + "train_accuracy_recipient_token_analy": 0.5, + "train_accuracy_recipient_token_analy_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_rate": 1.0, "train_accuracy_recipient_token_rate_total": 1, - "train_accuracy_recipient_token_recommend": 0.0, - "train_accuracy_recipient_token_recommend_total": 1, + "train_accuracy_recipient_token_re": 0.0, + "train_accuracy_recipient_token_re_total": 1, "train_total_number_recipient_token": 16 }, { - "epoch": 0.014822134387351778, - "grad_norm": 4.251122308361701, - "learning_rate": 2.2727272727272728e-06, - "loss": 0.5231, - "step": 5 - }, - { - "epoch": 0.014822134387351778, - "step": 5, - "train_accuracy_first_token_arguments": 0.7391304347826086, - "train_accuracy_first_token_arguments_total": 23, - "train_accuracy_recipient_token": 0.6875, + "epoch": 0.01383399209486166, + "step": 7, + "train_accuracy_recipient_token": 0.5625, "train_accuracy_recipient_token_all": 0.6, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_fruit": 1.0, "train_accuracy_recipient_token_fruit_total": 3, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_select": 1.0, + "train_accuracy_recipient_token_ident": 0.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_select": 0.0, "train_accuracy_recipient_token_select_total": 1, "train_total_number_recipient_token": 16 }, { - "epoch": 0.014822134387351778, - "step": 5, - "train_accuracy_first_token_arguments": 0.8888888888888888, - "train_accuracy_first_token_arguments_total": 9, - "train_accuracy_recipient_token": 0.6111111111111112, + "epoch": 0.015810276679841896, + "grad_norm": 2.9210762729436683, + "learning_rate": 2.5e-06, + "loss": 0.346, + "step": 8 + }, + { + "epoch": 0.015810276679841896, + "step": 8, + "train_accuracy_recipient_token": 0.6666666666666666, "train_accuracy_recipient_token_all": 0.6, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 0.7142857142857143, + "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, - "train_accuracy_recipient_token_identify": 0.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 0.0, + "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.014822134387351778, - "step": 5, - "train_accuracy_first_token_arguments": 0.6216216216216216, - "train_accuracy_first_token_arguments_total": 37, - "train_accuracy_recipient_token": 0.7894736842105263, + "epoch": 0.015810276679841896, + "step": 8, + "train_accuracy_recipient_token": 0.7368421052631579, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compose": 0.0, "train_accuracy_recipient_token_compose_total": 1, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 1, - "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_g": 1.0, "train_accuracy_recipient_token_g_total": 4, @@ -355,217 +340,207 @@ }, { "epoch": 0.017786561264822136, - "grad_norm": 2.9047536525561006, - "learning_rate": 2.7272727272727272e-06, - "loss": 0.3063, - "step": 6 + "grad_norm": 2.4385850326725835, + "learning_rate": 2.8125e-06, + "loss": 0.2706, + "step": 9 }, { "epoch": 0.017786561264822136, - "step": 6, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 13, - "train_accuracy_recipient_token": 0.7222222222222222, + "step": 9, + "train_accuracy_recipient_token": 0.6111111111111112, "train_accuracy_recipient_token_all": 0.6363636363636364, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_get": 0.8333333333333334, + "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 6, - "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate": 0.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.017786561264822136, - "step": 6, - "train_accuracy_first_token_arguments": 0.8648648648648649, - "train_accuracy_first_token_arguments_total": 37, - "train_accuracy_recipient_token": 0.5294117647058824, - "train_accuracy_recipient_token_all": 0.6363636363636364, + "step": 9, + "train_accuracy_recipient_token": 0.4117647058823529, + "train_accuracy_recipient_token_all": 0.5454545454545454, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare": 0.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 0.3333333333333333, "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_identify": 0.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 0.0, + "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_validate": 0.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 17 }, { - "epoch": 0.017786561264822136, - "step": 6, - "train_accuracy_first_token_arguments": 0.76, - "train_accuracy_first_token_arguments_total": 25, - "train_accuracy_recipient_token": 0.75, - "train_accuracy_recipient_token_all": 0.8333333333333334, + "epoch": 0.019762845849802372, + "grad_norm": 2.738667329727624, + "learning_rate": 3.125e-06, + "loss": 0.3037, + "step": 10 + }, + { + "epoch": 0.019762845849802372, + "step": 10, + "train_accuracy_recipient_token": 0.625, + "train_accuracy_recipient_token_all": 0.75, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_evaluate": 0.0, "train_accuracy_recipient_token_evaluate_total": 1, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 16 }, { - "epoch": 0.020750988142292492, - "grad_norm": 3.2977375008302783, - "learning_rate": 3.181818181818182e-06, - "loss": 0.4802, - "step": 7 - }, - { - "epoch": 0.020750988142292492, - "step": 7, - "train_accuracy_first_token_arguments": 0.9090909090909091, - "train_accuracy_first_token_arguments_total": 11, - "train_accuracy_recipient_token": 0.7777777777777778, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.019762845849802372, + "step": 10, + "train_accuracy_recipient_token": 0.7222222222222222, + "train_accuracy_recipient_token_all": 0.6, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 0.5714285714285714, + "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { - "epoch": 0.020750988142292492, - "step": 7, - "train_accuracy_first_token_arguments": 0.7692307692307693, - "train_accuracy_first_token_arguments_total": 13, - "train_accuracy_recipient_token": 0.75, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.021739130434782608, + "grad_norm": 3.164747479850259, + "learning_rate": 3.4375e-06, + "loss": 0.4035, + "step": 11 + }, + { + "epoch": 0.021739130434782608, + "step": 11, + "train_accuracy_recipient_token": 0.625, + "train_accuracy_recipient_token_all": 0.5, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 0.5, + "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 16 }, { - "epoch": 0.020750988142292492, - "step": 7, - "train_accuracy_first_token_arguments": 0.8421052631578947, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.7368421052631579, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.021739130434782608, + "step": 11, + "train_accuracy_recipient_token": 0.7894736842105263, + "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 0.5555555555555556, + "train_accuracy_recipient_token_get": 0.7777777777777778, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.023715415019762844, - "grad_norm": 2.6597485903906652, - "learning_rate": 3.6363636363636366e-06, - "loss": 0.4234, - "step": 8 + "grad_norm": 2.361065524720244, + "learning_rate": 3.7500000000000005e-06, + "loss": 0.3349, + "step": 12 }, { "epoch": 0.023715415019762844, - "step": 8, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 16, + "step": 12, "train_accuracy_recipient_token": 0.6666666666666666, - "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all": 0.6, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 0.0, - "train_accuracy_recipient_token_analyze_total": 1, - "train_accuracy_recipient_token_count": 0.5, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, + "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_identify": 0.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 0.0, + "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.023715415019762844, - "step": 8, - "train_accuracy_first_token_arguments": 0.9, - "train_accuracy_first_token_arguments_total": 10, + "step": 12, "train_accuracy_recipient_token": 0.8947368421052632, - "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { - "epoch": 0.023715415019762844, - "step": 8, - "train_accuracy_first_token_arguments": 0.8867924528301887, - "train_accuracy_first_token_arguments_total": 53, + "epoch": 0.025691699604743084, + "grad_norm": 2.0061236591150666, + "learning_rate": 4.0625000000000005e-06, + "loss": 0.2121, + "step": 13 + }, + { + "epoch": 0.025691699604743084, + "step": 13, "train_accuracy_recipient_token": 0.7368421052631579, - "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_describe": 1.0, "train_accuracy_recipient_token_describe_total": 2, - "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_sum": 0.0, + "train_accuracy_recipient_token_sum": 0.5, "train_accuracy_recipient_token_sum_total": 2, "train_total_number_recipient_token": 19 }, { - "epoch": 0.0266798418972332, - "grad_norm": 2.552647042706623, - "learning_rate": 4.0909090909090915e-06, - "loss": 0.3278, - "step": 9 - }, - { - "epoch": 0.0266798418972332, - "step": 9, - "train_accuracy_first_token_arguments": 0.7894736842105263, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.7777777777777778, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.025691699604743084, + "step": 13, + "train_accuracy_recipient_token": 0.6666666666666666, + "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_prior": 1.0, - "train_accuracy_recipient_token_prior_total": 1, + "train_accuracy_recipient_token_p": 1.0, + "train_accuracy_recipient_token_p_total": 1, "train_accuracy_recipient_token_schedule": 0.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.0266798418972332, - "step": 9, - "train_accuracy_first_token_arguments": 0.9333333333333333, - "train_accuracy_first_token_arguments_total": 15, - "train_accuracy_recipient_token": 0.5, - "train_accuracy_recipient_token_all": 0.8, + "epoch": 0.02766798418972332, + "grad_norm": 2.7988165456067735, + "learning_rate": 4.3750000000000005e-06, + "loss": 0.2798, + "step": 14 + }, + { + "epoch": 0.02766798418972332, + "step": 14, + "train_accuracy_recipient_token": 0.625, + "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 0.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 0.0, + "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_load": 0.0, + "train_accuracy_recipient_token_load": 1.0, "train_accuracy_recipient_token_load_total": 1, "train_total_number_recipient_token": 16 }, { - "epoch": 0.0266798418972332, - "step": 9, - "train_accuracy_first_token_arguments": 0.7142857142857143, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 0.95, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.02766798418972332, + "step": 14, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_property": 1.0, "train_accuracy_recipient_token_property_total": 3, @@ -575,97 +550,94 @@ }, { "epoch": 0.029644268774703556, - "grad_norm": 2.3528382096065767, - "learning_rate": 4.5454545454545455e-06, - "loss": 0.2897, - "step": 10 + "grad_norm": 2.4311181509760647, + "learning_rate": 4.6875000000000004e-06, + "loss": 0.2332, + "step": 15 }, { "epoch": 0.029644268774703556, - "step": 10, - "train_accuracy_first_token_arguments": 0.76, - "train_accuracy_first_token_arguments_total": 25, + "step": 15, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_assemble": 1.0, - "train_accuracy_recipient_token_assemble_total": 1, + "train_accuracy_recipient_token_as": 1.0, + "train_accuracy_recipient_token_as_total": 1, "train_accuracy_recipient_token_clean": 1.0, "train_accuracy_recipient_token_clean_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_repair": 0.5, - "train_accuracy_recipient_token_repair_total": 2, + "train_accuracy_recipient_token_rep": 0.5, + "train_accuracy_recipient_token_rep_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.029644268774703556, - "step": 10, - "train_accuracy_first_token_arguments": 0.8260869565217391, - "train_accuracy_first_token_arguments_total": 23, - "train_accuracy_recipient_token": 0.8947368421052632, - "train_accuracy_recipient_token_all": 0.9, + "step": 15, + "train_accuracy_recipient_token": 0.7368421052631579, + "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare": 0.75, "train_accuracy_recipient_token_compare_total": 4, - "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 1, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, "train_accuracy_recipient_token_search": 0.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.029644268774703556, - "step": 10, - "train_accuracy_first_token_arguments": 0.6333333333333333, - "train_accuracy_first_token_arguments_total": 30, - "train_accuracy_recipient_token": 0.6666666666666666, - "train_accuracy_recipient_token_all": 0.8, + "epoch": 0.03162055335968379, + "grad_norm": 2.6718500710390014, + "learning_rate": 5e-06, + "loss": 0.2296, + "step": 16 + }, + { + "epoch": 0.03162055335968379, + "step": 16, + "train_accuracy_recipient_token": 0.7222222222222222, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check": 0.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_suggest": 0.5, - "train_accuracy_recipient_token_suggest_total": 2, + "train_accuracy_recipient_token_s": 0.5, + "train_accuracy_recipient_token_s_total": 2, "train_total_number_recipient_token": 18 }, { - "epoch": 0.03260869565217391, - "grad_norm": 2.351257384059913, - "learning_rate": 5e-06, - "loss": 0.2987, - "step": 11 - }, - { - "epoch": 0.03260869565217391, - "step": 11, - "train_accuracy_first_token_arguments": 0.5106382978723404, - "train_accuracy_first_token_arguments_total": 47, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_add": 1.0, + "epoch": 0.03162055335968379, + "step": 16, + "train_accuracy_recipient_token": 0.7894736842105263, + "train_accuracy_recipient_token_add": 0.8, "train_accuracy_recipient_token_add_total": 5, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check": 0.0, "train_accuracy_recipient_token_check_total": 1, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 19 }, { - "epoch": 0.03260869565217391, - "step": 11, - "train_accuracy_first_token_arguments": 0.9130434782608695, - "train_accuracy_first_token_arguments_total": 23, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.03359683794466403, + "grad_norm": 2.4193634654326193, + "learning_rate": 4.999948617395916e-06, + "loss": 0.2744, + "step": 17 + }, + { + "epoch": 0.03359683794466403, + "step": 17, + "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, @@ -674,18 +646,16 @@ "train_total_number_recipient_token": 20 }, { - "epoch": 0.03260869565217391, - "step": 11, - "train_accuracy_first_token_arguments": 0.7142857142857143, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 0.9444444444444444, + "epoch": 0.03359683794466403, + "step": 17, + "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_fabric": 1.0, - "train_accuracy_recipient_token_fabric_total": 1, + "train_accuracy_recipient_token_f": 1.0, + "train_accuracy_recipient_token_f_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, @@ -693,20 +663,18 @@ }, { "epoch": 0.03557312252964427, - "grad_norm": 2.266360756598947, - "learning_rate": 4.999883916312e-06, - "loss": 0.3119, - "step": 12 + "grad_norm": 1.9915037614380675, + "learning_rate": 4.9997944716957985e-06, + "loss": 0.2374, + "step": 18 }, { "epoch": 0.03557312252964427, - "step": 12, - "train_accuracy_first_token_arguments": 0.696969696969697, - "train_accuracy_first_token_arguments_total": 33, + "step": 18, "train_accuracy_recipient_token": 0.9, - "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 5, - "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find": 0.75, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, @@ -714,84 +682,81 @@ }, { "epoch": 0.03557312252964427, - "step": 12, - "train_accuracy_first_token_arguments": 0.8048780487804879, - "train_accuracy_first_token_arguments_total": 41, - "train_accuracy_recipient_token": 0.7058823529411765, - "train_accuracy_recipient_token_all": 0.7, + "step": 18, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 2, - "train_accuracy_recipient_token_explain": 0.0, - "train_accuracy_recipient_token_explain_total": 1, - "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_ex": 1.0, + "train_accuracy_recipient_token_ex_total": 1, + "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_translate": 0.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 17 }, { - "epoch": 0.03557312252964427, - "step": 12, - "train_accuracy_first_token_arguments": 0.8333333333333334, - "train_accuracy_first_token_arguments_total": 18, - "train_accuracy_recipient_token": 0.9473684210526315, + "epoch": 0.037549407114624504, + "grad_norm": 13.63293630478699, + "learning_rate": 4.999537569235975e-06, + "loss": 0.575, + "step": 19 + }, + { + "epoch": 0.037549407114624504, + "step": 19, + "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate": 0.0, "train_accuracy_recipient_token_calculate_total": 1, - "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { - "epoch": 0.038537549407114624, - "grad_norm": 10.867702658903724, - "learning_rate": 4.999535676028338e-06, - "loss": 0.4711, - "step": 13 - }, - { - "epoch": 0.038537549407114624, - "step": 13, - "train_accuracy_first_token_arguments": 0.9333333333333333, - "train_accuracy_first_token_arguments_total": 15, - "train_accuracy_recipient_token": 0.7058823529411765, - "train_accuracy_recipient_token_all": 0.7, + "epoch": 0.037549407114624504, + "step": 19, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, - "train_accuracy_recipient_token_suggest": 0.0, - "train_accuracy_recipient_token_suggest_total": 1, + "train_accuracy_recipient_token_s": 0.0, + "train_accuracy_recipient_token_s_total": 1, "train_total_number_recipient_token": 17 }, { - "epoch": 0.038537549407114624, - "step": 13, - "train_accuracy_first_token_arguments": 0.875, - "train_accuracy_first_token_arguments_total": 16, + "epoch": 0.039525691699604744, + "grad_norm": 2.7743056709907594, + "learning_rate": 4.9991779205767e-06, + "loss": 0.2632, + "step": 20 + }, + { + "epoch": 0.039525691699604744, + "step": 20, "train_accuracy_recipient_token": 0.8947368421052632, - "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_verify": 1.0, "train_accuracy_recipient_token_verify_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.038537549407114624, - "step": 13, - "train_accuracy_first_token_arguments": 0.9285714285714286, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.8, + "epoch": 0.039525691699604744, + "step": 20, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 1, @@ -799,62 +764,63 @@ "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_track": 1.0, "train_accuracy_recipient_token_track_total": 1, - "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate": 0.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.041501976284584984, - "grad_norm": 2.8220096523893234, - "learning_rate": 4.998955311489025e-06, - "loss": 0.3082, - "step": 14 + "grad_norm": 2.867128204261161, + "learning_rate": 4.99871554050172e-06, + "loss": 0.2, + "step": 21 }, { "epoch": 0.041501976284584984, - "step": 14, - "train_accuracy_first_token_arguments": 0.8636363636363636, - "train_accuracy_first_token_arguments_total": 22, - "train_accuracy_recipient_token": 0.8333333333333334, - "train_accuracy_recipient_token_all": 0.7, + "step": 21, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check": 0.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 18 }, { "epoch": 0.041501976284584984, - "step": 14, - "train_accuracy_first_token_arguments": 0.7777777777777778, - "train_accuracy_first_token_arguments_total": 18, - "train_accuracy_recipient_token": 0.8823529411764706, + "step": 21, + "train_accuracy_recipient_token": 0.7058823529411765, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_buy": 1.0, + "train_accuracy_recipient_token_buy": 0.0, "train_accuracy_recipient_token_buy_total": 1, - "train_accuracy_recipient_token_exp": 1.0, + "train_accuracy_recipient_token_exp": 0.0, "train_accuracy_recipient_token_exp_total": 1, - "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 17 }, { - "epoch": 0.041501976284584984, - "step": 14, - "train_accuracy_first_token_arguments": 0.8285714285714286, - "train_accuracy_first_token_arguments_total": 35, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.043478260869565216, + "grad_norm": 2.8424629908120873, + "learning_rate": 4.99815044801767e-06, + "loss": 0.3206, + "step": 22 + }, + { + "epoch": 0.043478260869565216, + "step": 22, + "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 1, - "train_accuracy_recipient_token_filter": 1.0, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, + "train_accuracy_recipient_token_filter": 0.0, "train_accuracy_recipient_token_filter_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, @@ -863,84 +829,76 @@ "train_total_number_recipient_token": 19 }, { - "epoch": 0.044466403162055336, - "grad_norm": 2.258107289871962, - "learning_rate": 4.998142876590749e-06, - "loss": 0.3097, - "step": 15 - }, - { - "epoch": 0.044466403162055336, - "step": 15, - "train_accuracy_first_token_arguments": 0.75, - "train_accuracy_first_token_arguments_total": 16, - "train_accuracy_recipient_token": 0.9473684210526315, + "epoch": 0.043478260869565216, + "step": 22, + "train_accuracy_recipient_token": 0.7894736842105263, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_clean": 1.0, "train_accuracy_recipient_token_clean_total": 1, - "train_accuracy_recipient_token_generate": 1.0, + "train_accuracy_recipient_token_generate": 0.6, "train_accuracy_recipient_token_generate_total": 5, - "train_accuracy_recipient_token_remove": 1.0, + "train_accuracy_recipient_token_remove": 0.0, "train_accuracy_recipient_token_remove_total": 1, "train_accuracy_recipient_token_replace": 1.0, "train_accuracy_recipient_token_replace_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.044466403162055336, - "step": 15, - "train_accuracy_first_token_arguments": 0.75, - "train_accuracy_first_token_arguments_total": 16, - "train_accuracy_recipient_token": 0.7058823529411765, - "train_accuracy_recipient_token_all": 0.6, + "epoch": 0.045454545454545456, + "grad_norm": 3.269809875569231, + "learning_rate": 4.997482666353287e-06, + "loss": 0.2165, + "step": 23 + }, + { + "epoch": 0.045454545454545456, + "step": 23, + "train_accuracy_recipient_token": 0.6470588235294118, + "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_identify": 0.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_register": 1.0, + "train_accuracy_recipient_token_ident": 0.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_register": 0.0, "train_accuracy_recipient_token_register_total": 1, "train_total_number_recipient_token": 17 }, { - "epoch": 0.044466403162055336, - "step": 15, - "train_accuracy_first_token_arguments": 0.9545454545454546, - "train_accuracy_first_token_arguments_total": 22, - "train_accuracy_recipient_token": 0.8947368421052632, + "epoch": 0.045454545454545456, + "step": 23, + "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_schedule": 0.0, + "train_accuracy_recipient_token_s": 1.0, + "train_accuracy_recipient_token_s_total": 1, + "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, - "train_accuracy_recipient_token_suggest": 1.0, - "train_accuracy_recipient_token_suggest_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.04743083003952569, - "grad_norm": 3.2749980475323635, - "learning_rate": 4.997098446781861e-06, - "loss": 0.2419, - "step": 16 + "grad_norm": 2.399757318906346, + "learning_rate": 4.9967122229584614e-06, + "loss": 0.1488, + "step": 24 }, { "epoch": 0.04743083003952569, - "step": 16, - "train_accuracy_first_token_arguments": 0.85, - "train_accuracy_first_token_arguments_total": 20, + "step": 24, "train_accuracy_recipient_token": 0.9, - "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find": 0.8, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, @@ -950,28 +908,31 @@ }, { "epoch": 0.04743083003952569, - "step": 16, - "train_accuracy_first_token_arguments": 0.7096774193548387, - "train_accuracy_first_token_arguments_total": 62, - "train_accuracy_recipient_token": 0.8947368421052632, - "train_accuracy_recipient_token_all": 0.8, + "step": 24, + "train_accuracy_recipient_token": 0.7894736842105263, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 3, + "train_accuracy_recipient_token_ident": 0.3333333333333333, + "train_accuracy_recipient_token_ident_total": 3, "train_total_number_recipient_token": 19 }, { - "epoch": 0.04743083003952569, - "step": 16, - "train_accuracy_first_token_arguments": 0.9285714285714286, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 0.8823529411764706, + "epoch": 0.04940711462450593, + "grad_norm": 1.7504463644259465, + "learning_rate": 4.995839149503103e-06, + "loss": 0.1377, + "step": 25 + }, + { + "epoch": 0.04940711462450593, + "step": 25, + "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_c": 1.0, @@ -985,22 +946,13 @@ "train_total_number_recipient_token": 17 }, { - "epoch": 0.05039525691699605, - "grad_norm": 2.3223843516425937, - "learning_rate": 4.9958221190553705e-06, - "loss": 0.2414, - "step": 17 - }, - { - "epoch": 0.05039525691699605, - "step": 17, - "train_accuracy_first_token_arguments": 0.9583333333333334, - "train_accuracy_first_token_arguments_total": 24, + "epoch": 0.04940711462450593, + "step": 25, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, @@ -1008,10 +960,15 @@ "train_total_number_recipient_token": 18 }, { - "epoch": 0.05039525691699605, - "step": 17, - "train_accuracy_first_token_arguments": 0.92, - "train_accuracy_first_token_arguments_total": 25, + "epoch": 0.05138339920948617, + "grad_norm": 2.7693470356484107, + "learning_rate": 4.994863481875842e-06, + "loss": 0.2731, + "step": 26 + }, + { + "epoch": 0.05138339920948617, + "step": 26, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, @@ -1026,10 +983,8 @@ "train_total_number_recipient_token": 20 }, { - "epoch": 0.05039525691699605, - "step": 17, - "train_accuracy_first_token_arguments": 0.8571428571428571, - "train_accuracy_first_token_arguments_total": 21, + "epoch": 0.05138339920948617, + "step": 26, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, @@ -1039,28 +994,26 @@ "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_nutrition": 1.0, - "train_accuracy_recipient_token_nutrition_total": 1, + "train_accuracy_recipient_token_nut": 1.0, + "train_accuracy_recipient_token_nut_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.0533596837944664, - "grad_norm": 2.7061439489759453, - "learning_rate": 4.994314011939941e-06, - "loss": 0.2925, - "step": 18 + "grad_norm": 2.9947438603207726, + "learning_rate": 4.993785260182552e-06, + "loss": 0.2053, + "step": 27 }, { "epoch": 0.0533596837944664, - "step": 18, - "train_accuracy_first_token_arguments": 0.9666666666666667, - "train_accuracy_first_token_arguments_total": 30, - "train_accuracy_recipient_token": 0.8823529411764706, - "train_accuracy_recipient_token_all": 0.9090909090909091, + "step": 27, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_building": 1.0, "train_accuracy_recipient_token_building_total": 1, - "train_accuracy_recipient_token_find": 0.75, + "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_property": 1.0, "train_accuracy_recipient_token_property_total": 1, @@ -1068,9 +1021,7 @@ }, { "epoch": 0.0533596837944664, - "step": 18, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 13, + "step": 27, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, @@ -1079,49 +1030,50 @@ "train_total_number_recipient_token": 19 }, { - "epoch": 0.0533596837944664, - "step": 18, - "train_accuracy_first_token_arguments": 0.6428571428571429, - "train_accuracy_first_token_arguments_total": 14, + "epoch": 0.05533596837944664, + "grad_norm": 2.8975618558727008, + "learning_rate": 4.992604528744705e-06, + "loss": 0.2712, + "step": 28 + }, + { + "epoch": 0.05533596837944664, + "step": 28, "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_decode": 0.0, + "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { - "epoch": 0.05632411067193676, - "grad_norm": 2.8315352410694907, - "learning_rate": 4.992574265488883e-06, - "loss": 0.3402, - "step": 19 - }, - { - "epoch": 0.05632411067193676, - "step": 19, - "train_accuracy_first_token_arguments": 0.7916666666666666, - "train_accuracy_first_token_arguments_total": 24, - "train_accuracy_recipient_token": 0.7894736842105263, + "epoch": 0.05533596837944664, + "step": 28, + "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_plan": 0.0, "train_accuracy_recipient_token_plan_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.05632411067193676, - "step": 19, - "train_accuracy_first_token_arguments": 0.8, - "train_accuracy_first_token_arguments_total": 15, + "epoch": 0.05731225296442688, + "grad_norm": 2.5657098582780398, + "learning_rate": 4.991321336097546e-06, + "loss": 0.2796, + "step": 29 + }, + { + "epoch": 0.05731225296442688, + "step": 29, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, @@ -1130,12 +1082,10 @@ "train_total_number_recipient_token": 19 }, { - "epoch": 0.05632411067193676, - "step": 19, - "train_accuracy_first_token_arguments": 0.8571428571428571, - "train_accuracy_first_token_arguments_total": 14, + "epoch": 0.05731225296442688, + "step": 29, "train_accuracy_recipient_token": 0.85, - "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, @@ -1149,103 +1099,98 @@ }, { "epoch": 0.05928853754940711, - "grad_norm": 3.725703665798811, - "learning_rate": 4.990603041267144e-06, - "loss": 0.2884, - "step": 20 + "grad_norm": 2.7207547834543275, + "learning_rate": 4.989935734988098e-06, + "loss": 0.2106, + "step": 30 }, { "epoch": 0.05928853754940711, - "step": 20, - "train_accuracy_first_token_arguments": 0.7241379310344828, - "train_accuracy_first_token_arguments_total": 29, + "step": 30, "train_accuracy_recipient_token": 0.95, - "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 0.8, + "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, - "train_accuracy_recipient_token_suggest": 1.0, - "train_accuracy_recipient_token_suggest_total": 1, + "train_accuracy_recipient_token_s": 1.0, + "train_accuracy_recipient_token_s_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.05928853754940711, - "step": 20, - "train_accuracy_first_token_arguments": 0.8, - "train_accuracy_first_token_arguments_total": 20, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 1.0, + "step": 30, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 0.8888888888888888, + "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { - "epoch": 0.05928853754940711, - "step": 20, - "train_accuracy_first_token_arguments": 0.8, - "train_accuracy_first_token_arguments_total": 15, - "train_accuracy_recipient_token": 0.5625, - "train_accuracy_recipient_token_all": 0.6363636363636364, + "epoch": 0.06126482213438735, + "grad_norm": 2.0609870060002202, + "learning_rate": 4.988447782372996e-06, + "loss": 0.2457, + "step": 31 + }, + { + "epoch": 0.06126482213438735, + "step": 31, + "train_accuracy_recipient_token": 0.6875, + "train_accuracy_recipient_token_all": 0.5454545454545454, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_deploy": 1.0, "train_accuracy_recipient_token_deploy_total": 1, - "train_accuracy_recipient_token_identify": 0.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_install": 1.0, "train_accuracy_recipient_token_install_total": 1, - "train_accuracy_recipient_token_schedule": 0.0, + "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 16 }, { - "epoch": 0.06225296442687747, - "grad_norm": 2.4295313275927413, - "learning_rate": 4.988400522336304e-06, - "loss": 0.2966, - "step": 21 - }, - { - "epoch": 0.06225296442687747, - "step": 21, - "train_accuracy_first_token_arguments": 0.9473684210526315, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.7647058823529411, - "train_accuracy_recipient_token_all": 0.8, + "epoch": 0.06126482213438735, + "step": 31, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_order": 0.0, + "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, - "train_accuracy_recipient_token_prior": 1.0, - "train_accuracy_recipient_token_prior_total": 1, + "train_accuracy_recipient_token_p": 1.0, + "train_accuracy_recipient_token_p_total": 1, "train_total_number_recipient_token": 17 }, { - "epoch": 0.06225296442687747, - "step": 21, - "train_accuracy_first_token_arguments": 0.8461538461538461, - "train_accuracy_first_token_arguments_total": 13, - "train_accuracy_recipient_token": 0.7647058823529411, + "epoch": 0.06324110671936758, + "grad_norm": 2.6519364596617034, + "learning_rate": 4.986857539416144e-06, + "loss": 0.2742, + "step": 32 + }, + { + "epoch": 0.06324110671936758, + "step": 32, + "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 0.7142857142857143, + "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { - "epoch": 0.06225296442687747, - "step": 21, - "train_accuracy_first_token_arguments": 0.9, - "train_accuracy_first_token_arguments_total": 20, + "epoch": 0.06324110671936758, + "step": 32, "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, @@ -1253,56 +1198,57 @@ "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_translate": 0.0, + "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.06521739130434782, - "grad_norm": 3.0399009498719938, - "learning_rate": 4.985966913237581e-06, - "loss": 0.3759, - "step": 22 + "grad_norm": 3.3761072923479434, + "learning_rate": 4.985165071486201e-06, + "loss": 0.3322, + "step": 33 }, { "epoch": 0.06521739130434782, - "step": 22, - "train_accuracy_first_token_arguments": 0.9230769230769231, - "train_accuracy_first_token_arguments_total": 13, - "train_accuracy_recipient_token": 0.7777777777777778, + "step": 33, + "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_estimate": 0.0, + "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_style": 1.0, "train_accuracy_recipient_token_style_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.06521739130434782, - "step": 22, - "train_accuracy_first_token_arguments": 0.5, - "train_accuracy_first_token_arguments_total": 6, + "step": 33, "train_accuracy_recipient_token": 0.7333333333333333, - "train_accuracy_recipient_token_all": 0.7, + "train_accuracy_recipient_token_all": 0.6, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_c": 0.0, + "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 15 }, { - "epoch": 0.06521739130434782, - "step": 22, - "train_accuracy_first_token_arguments": 0.7804878048780488, - "train_accuracy_first_token_arguments_total": 41, + "epoch": 0.06719367588932806, + "grad_norm": 3.2773278792649094, + "learning_rate": 4.983370448153896e-06, + "loss": 0.4958, + "step": 34 + }, + { + "epoch": 0.06719367588932806, + "step": 34, "train_accuracy_recipient_token": 0.8947368421052632, - "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, @@ -1315,17 +1261,8 @@ "train_total_number_recipient_token": 19 }, { - "epoch": 0.06818181818181818, - "grad_norm": 2.939285198394914, - "learning_rate": 4.9833024399728295e-06, - "loss": 0.4389, - "step": 23 - }, - { - "epoch": 0.06818181818181818, - "step": 23, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 17, + "epoch": 0.06719367588932806, + "step": 34, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, @@ -1336,24 +1273,27 @@ "train_total_number_recipient_token": 17 }, { - "epoch": 0.06818181818181818, - "step": 23, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 0.5, + "epoch": 0.0691699604743083, + "grad_norm": 2.1733891801745626, + "learning_rate": 4.981473743189163e-06, + "loss": 0.1996, + "step": 35 + }, + { + "epoch": 0.0691699604743083, + "step": 35, + "train_accuracy_recipient_token": 0.75, "train_accuracy_recipient_token_all": 0.6363636363636364, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_get": 0.25, + "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_track": 0.0, + "train_accuracy_recipient_token_track": 1.0, "train_accuracy_recipient_token_track_total": 1, "train_total_number_recipient_token": 16 }, { - "epoch": 0.06818181818181818, - "step": 23, - "train_accuracy_first_token_arguments": 0.9375, - "train_accuracy_first_token_arguments_total": 16, + "epoch": 0.0691699604743083, + "step": 35, "train_accuracy_recipient_token": 0.8125, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, @@ -1367,18 +1307,16 @@ }, { "epoch": 0.07114624505928854, - "grad_norm": 3.2811864983396806, - "learning_rate": 4.980407349983556e-06, - "loss": 0.2645, - "step": 24 + "grad_norm": 2.8245080994396585, + "learning_rate": 4.979475034558115e-06, + "loss": 0.2105, + "step": 36 }, { "epoch": 0.07114624505928854, - "step": 24, - "train_accuracy_first_token_arguments": 0.9166666666666666, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, + "step": 36, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, @@ -1388,11 +1326,9 @@ }, { "epoch": 0.07114624505928854, - "step": 24, - "train_accuracy_first_token_arguments": 0.9016393442622951, - "train_accuracy_first_token_arguments_total": 61, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, + "step": 36, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, @@ -1400,17 +1336,22 @@ "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 20 }, { - "epoch": 0.07114624505928854, - "step": 24, - "train_accuracy_first_token_arguments": 0.76, - "train_accuracy_first_token_arguments_total": 25, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.9090909090909091, + "epoch": 0.07312252964426877, + "grad_norm": 1.6572734803173075, + "learning_rate": 4.977374404419838e-06, + "loss": 0.2089, + "step": 37 + }, + { + "epoch": 0.07312252964426877, + "step": 37, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, @@ -1423,49 +1364,43 @@ "train_total_number_recipient_token": 18 }, { - "epoch": 0.0741106719367589, - "grad_norm": 1.8226124894214228, - "learning_rate": 4.9772819121279395e-06, - "loss": 0.2212, - "step": 25 - }, - { - "epoch": 0.0741106719367589, - "step": 25, - "train_accuracy_first_token_arguments": 0.9193548387096774, - "train_accuracy_first_token_arguments_total": 62, - "train_accuracy_recipient_token": 0.9444444444444444, + "epoch": 0.07312252964426877, + "step": 37, + "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.7142857142857143, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { - "epoch": 0.0741106719367589, - "step": 25, - "train_accuracy_first_token_arguments": 0.76, - "train_accuracy_first_token_arguments_total": 25, + "epoch": 0.07509881422924901, + "grad_norm": 2.053920013378982, + "learning_rate": 4.9751719391230055e-06, + "loss": 0.1193, + "step": 38 + }, + { + "epoch": 0.07509881422924901, + "step": 38, "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.875, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { - "epoch": 0.0741106719367589, - "step": 25, - "train_accuracy_first_token_arguments": 0.6666666666666666, - "train_accuracy_first_token_arguments_total": 24, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.07509881422924901, + "step": 38, + "train_accuracy_recipient_token": 0.7777777777777778, + "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_plan": 0.0, "train_accuracy_recipient_token_plan_total": 1, @@ -1473,36 +1408,32 @@ }, { "epoch": 0.07707509881422925, - "grad_norm": 1.524883365948892, - "learning_rate": 4.973926416655863e-06, - "loss": 0.1674, - "step": 26 + "grad_norm": 1.2542268957003366, + "learning_rate": 4.9728677292023405e-06, + "loss": 0.1507, + "step": 39 }, { "epoch": 0.07707509881422925, - "step": 26, - "train_accuracy_first_token_arguments": 0.8333333333333334, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.7777777777777778, - "train_accuracy_recipient_token_all": 0.8, + "step": 39, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_animal": 0.6666666666666666, + "train_accuracy_recipient_token_animal": 0.0, "train_accuracy_recipient_token_animal_total": 3, - "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_im": 1.0, "train_accuracy_recipient_token_im_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.07707509881422925, - "step": 26, - "train_accuracy_first_token_arguments": 0.8260869565217391, - "train_accuracy_first_token_arguments_total": 23, - "train_accuracy_recipient_token": 0.8235294117647058, - "train_accuracy_recipient_token_all": 0.8, + "step": 39, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, @@ -1515,31 +1446,27 @@ "train_total_number_recipient_token": 17 }, { - "epoch": 0.07707509881422925, - "step": 26, - "train_accuracy_first_token_arguments": 0.75, - "train_accuracy_first_token_arguments_total": 20, - "train_accuracy_recipient_token": 0.9047619047619048, + "epoch": 0.07905138339920949, + "grad_norm": 1.7847434934616466, + "learning_rate": 4.97046186937489e-06, + "loss": 0.1873, + "step": 40 + }, + { + "epoch": 0.07905138339920949, + "step": 40, + "train_accuracy_recipient_token": 0.8571428571428571, "train_accuracy_recipient_token_all": 0.9166666666666666, "train_accuracy_recipient_token_all_total": 12, - "train_accuracy_recipient_token_get": 0.875, + "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 8, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 21 }, { - "epoch": 0.0800395256916996, - "grad_norm": 2.0177323408527643, - "learning_rate": 4.970341175181957e-06, - "loss": 0.2729, - "step": 27 - }, - { - "epoch": 0.0800395256916996, - "step": 27, - "train_accuracy_first_token_arguments": 0.8888888888888888, - "train_accuracy_first_token_arguments_total": 9, + "epoch": 0.07905138339920949, + "step": 40, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, @@ -1552,14 +1479,19 @@ "train_total_number_recipient_token": 17 }, { - "epoch": 0.0800395256916996, - "step": 27, - "train_accuracy_first_token_arguments": 0.8787878787878788, - "train_accuracy_first_token_arguments_total": 33, - "train_accuracy_recipient_token": 0.8888888888888888, + "epoch": 0.08102766798418973, + "grad_norm": 2.2812627178319715, + "learning_rate": 4.967954458536126e-06, + "loss": 0.3651, + "step": 41 + }, + { + "epoch": 0.08102766798418973, + "step": 41, + "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_draw": 1.0, + "train_accuracy_recipient_token_draw": 0.8, "train_accuracy_recipient_token_draw_total": 5, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, @@ -1568,31 +1500,27 @@ "train_total_number_recipient_token": 18 }, { - "epoch": 0.0800395256916996, - "step": 27, - "train_accuracy_first_token_arguments": 0.7692307692307693, - "train_accuracy_first_token_arguments_total": 13, - "train_accuracy_recipient_token": 0.85, + "epoch": 0.08102766798418973, + "step": 41, + "train_accuracy_recipient_token": 0.8, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, - "train_accuracy_recipient_token_get": 0.875, + "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 20 }, { "epoch": 0.08300395256916997, - "grad_norm": 2.0988150229289273, - "learning_rate": 4.966526520656663e-06, - "loss": 0.3187, - "step": 28 + "grad_norm": 2.184012199227446, + "learning_rate": 4.965345599755888e-06, + "loss": 0.2094, + "step": 42 }, { "epoch": 0.08300395256916997, - "step": 28, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 15, + "step": 42, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.7692307692307693, "train_accuracy_recipient_token_all_total": 13, @@ -1604,9 +1532,7 @@ }, { "epoch": 0.08300395256916997, - "step": 28, - "train_accuracy_first_token_arguments": 0.7857142857142857, - "train_accuracy_first_token_arguments_total": 14, + "step": 42, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, @@ -1614,19 +1540,24 @@ "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.08300395256916997, - "step": 28, - "train_accuracy_first_token_arguments": 0.7241379310344828, - "train_accuracy_first_token_arguments_total": 29, - "train_accuracy_recipient_token": 0.8333333333333334, - "train_accuracy_recipient_token_all": 0.7272727272727273, + "epoch": 0.08498023715415019, + "grad_norm": 2.2059440276239743, + "learning_rate": 4.9626354002741424e-06, + "loss": 0.2288, + "step": 43 + }, + { + "epoch": 0.08498023715415019, + "step": 43, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, @@ -1639,31 +1570,27 @@ "train_total_number_recipient_token": 18 }, { - "epoch": 0.08596837944664032, - "grad_norm": 2.2452954917456447, - "learning_rate": 4.9624828073353144e-06, - "loss": 0.2444, - "step": 29 - }, - { - "epoch": 0.08596837944664032, - "step": 29, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 24, - "train_accuracy_recipient_token": 0.8888888888888888, + "epoch": 0.08498023715415019, + "step": 43, + "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 18 }, { - "epoch": 0.08596837944664032, - "step": 29, - "train_accuracy_first_token_arguments": 0.9, - "train_accuracy_first_token_arguments_total": 10, + "epoch": 0.08695652173913043, + "grad_norm": 2.04652434683755, + "learning_rate": 4.959823971496575e-06, + "loss": 0.2356, + "step": 44 + }, + { + "epoch": 0.08695652173913043, + "step": 44, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, @@ -1674,12 +1601,10 @@ "train_total_number_recipient_token": 19 }, { - "epoch": 0.08596837944664032, - "step": 29, - "train_accuracy_first_token_arguments": 0.75, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.8, - "train_accuracy_recipient_token_all": 0.8, + "epoch": 0.08695652173913043, + "step": 44, + "train_accuracy_recipient_token": 0.8666666666666667, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, @@ -1691,16 +1616,14 @@ }, { "epoch": 0.08893280632411067, - "grad_norm": 2.1813680120346604, - "learning_rate": 4.958210410745237e-06, - "loss": 0.2787, - "step": 30 + "grad_norm": 2.367283225731573, + "learning_rate": 4.95691142899001e-06, + "loss": 0.2289, + "step": 45 }, { "epoch": 0.08893280632411067, - "step": 30, - "train_accuracy_first_token_arguments": 0.9259259259259259, - "train_accuracy_first_token_arguments_total": 54, + "step": 45, "train_accuracy_recipient_token": 0.8125, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, @@ -1712,11 +1635,9 @@ }, { "epoch": 0.08893280632411067, - "step": 30, - "train_accuracy_first_token_arguments": 0.8235294117647058, - "train_accuracy_first_token_arguments_total": 17, - "train_accuracy_recipient_token": 0.7777777777777778, - "train_accuracy_recipient_token_all": 0.6666666666666666, + "step": 45, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.75, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, @@ -1727,10 +1648,15 @@ "train_total_number_recipient_token": 18 }, { - "epoch": 0.08893280632411067, - "step": 30, - "train_accuracy_first_token_arguments": 0.88, - "train_accuracy_first_token_arguments_total": 25, + "epoch": 0.09090909090909091, + "grad_norm": 2.580407364292025, + "learning_rate": 4.953897892477664e-06, + "loss": 0.4122, + "step": 46 + }, + { + "epoch": 0.09090909090909091, + "step": 46, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, @@ -1745,33 +1671,29 @@ "train_total_number_recipient_token": 19 }, { - "epoch": 0.09189723320158102, - "grad_norm": 1.9535807669288685, - "learning_rate": 4.95370972765087e-06, - "loss": 0.384, - "step": 31 - }, - { - "epoch": 0.09189723320158102, - "step": 31, - "train_accuracy_first_token_arguments": 0.6666666666666666, - "train_accuracy_first_token_arguments_total": 15, + "epoch": 0.09090909090909091, + "step": 46, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_convert": 1.0, + "train_accuracy_recipient_token_convert": 0.8, "train_accuracy_recipient_token_convert_total": 5, - "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 19 }, { - "epoch": 0.09189723320158102, - "step": 31, - "train_accuracy_first_token_arguments": 0.75, - "train_accuracy_first_token_arguments_total": 16, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9090909090909091, + "epoch": 0.09288537549407115, + "grad_norm": 2.516544751703755, + "learning_rate": 4.950783485834218e-06, + "loss": 0.2026, + "step": 47 + }, + { + "epoch": 0.09288537549407115, + "step": 47, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_evaluate": 1.0, "train_accuracy_recipient_token_evaluate_total": 1, @@ -1782,10 +1704,8 @@ "train_total_number_recipient_token": 19 }, { - "epoch": 0.09189723320158102, - "step": 31, - "train_accuracy_first_token_arguments": 0.7368421052631579, - "train_accuracy_first_token_arguments_total": 19, + "epoch": 0.09288537549407115, + "step": 47, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, @@ -1793,58 +1713,59 @@ "train_accuracy_recipient_token_filter_total": 5, "train_accuracy_recipient_token_food": 1.0, "train_accuracy_recipient_token_food_total": 4, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.09486166007905138, - "grad_norm": 2.5184069445357213, - "learning_rate": 4.94898117601693e-06, - "loss": 0.255, - "step": 32 + "grad_norm": 2.1047249182046577, + "learning_rate": 4.947568337080733e-06, + "loss": 0.2317, + "step": 48 }, { "epoch": 0.09486166007905138, - "step": 32, - "train_accuracy_first_token_arguments": 0.7272727272727273, - "train_accuracy_first_token_arguments_total": 22, - "train_accuracy_recipient_token": 0.8888888888888888, + "step": 48, + "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate": 0.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.09486166007905138, - "step": 32, - "train_accuracy_first_token_arguments": 0.8461538461538461, - "train_accuracy_first_token_arguments_total": 13, - "train_accuracy_recipient_token": 0.8333333333333334, + "step": 48, + "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_schedule": 0.0, + "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 2, "train_total_number_recipient_token": 18 }, { - "epoch": 0.09486166007905138, - "step": 32, - "train_accuracy_first_token_arguments": 0.9444444444444444, - "train_accuracy_first_token_arguments_total": 18, + "epoch": 0.09683794466403162, + "grad_norm": 2.1380480543246896, + "learning_rate": 4.944252578379379e-06, + "loss": 0.2049, + "step": 49 + }, + { + "epoch": 0.09683794466403162, + "step": 49, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, @@ -1852,40 +1773,36 @@ "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, - "train_accuracy_recipient_token_suggest": 1.0, - "train_accuracy_recipient_token_suggest_total": 1, + "train_accuracy_recipient_token_s": 1.0, + "train_accuracy_recipient_token_s_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.09782608695652174, - "grad_norm": 1.7872388531938566, - "learning_rate": 4.944025194969586e-06, - "loss": 0.241, - "step": 33 - }, - { - "epoch": 0.09782608695652174, - "step": 33, - "train_accuracy_first_token_arguments": 0.6428571428571429, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compare": 1.0, + "epoch": 0.09683794466403162, + "step": 49, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { - "epoch": 0.09782608695652174, - "step": 33, - "train_accuracy_first_token_arguments": 0.9528301886792453, - "train_accuracy_first_token_arguments_total": 106, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.09881422924901186, + "grad_norm": 1.887568271122836, + "learning_rate": 4.940836346028011e-06, + "loss": 0.3046, + "step": 50 + }, + { + "epoch": 0.09881422924901186, + "step": 50, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_assert": 1.0, + "train_accuracy_recipient_token_assert": 0.8, "train_accuracy_recipient_token_assert_total": 5, "train_accuracy_recipient_token_describe": 1.0, "train_accuracy_recipient_token_describe_total": 1, @@ -1894,131 +1811,122 @@ "train_total_number_recipient_token": 18 }, { - "epoch": 0.09782608695652174, - "step": 33, - "train_accuracy_first_token_arguments": 0.9811320754716981, - "train_accuracy_first_token_arguments_total": 53, + "epoch": 0.09881422924901186, + "step": 50, "train_accuracy_recipient_token": 0.8947368421052632, - "train_accuracy_recipient_token_all": 0.9166666666666666, + "train_accuracy_recipient_token_all": 0.8333333333333334, "train_accuracy_recipient_token_all_total": 12, - "train_accuracy_recipient_token_get": 0.8571428571428571, + "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { "epoch": 0.1007905138339921, - "grad_norm": 1.826928371292299, - "learning_rate": 4.938842244755683e-06, - "loss": 0.3343, - "step": 34 + "grad_norm": 1.5031262286748293, + "learning_rate": 4.937319780454559e-06, + "loss": 0.2287, + "step": 51 }, { "epoch": 0.1007905138339921, - "step": 34, - "train_accuracy_first_token_arguments": 0.75, - "train_accuracy_first_token_arguments_total": 8, - "train_accuracy_recipient_token": 0.8235294117647058, + "step": 51, + "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 0.8571428571428571, + "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.1007905138339921, - "step": 34, - "train_accuracy_first_token_arguments": 0.8888888888888888, - "train_accuracy_first_token_arguments_total": 18, + "step": 51, "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 0.8888888888888888, + "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { - "epoch": 0.1007905138339921, - "step": 34, - "train_accuracy_first_token_arguments": 0.9333333333333333, - "train_accuracy_first_token_arguments_total": 30, - "train_accuracy_recipient_token": 0.7894736842105263, + "epoch": 0.10276679841897234, + "grad_norm": 2.1419992157156957, + "learning_rate": 4.933703026211262e-06, + "loss": 0.2536, + "step": 52 + }, + { + "epoch": 0.10276679841897234, + "step": 52, + "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_compare": 0.0, + "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 1, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.10375494071146245, - "grad_norm": 2.586283846482814, - "learning_rate": 4.933432806700004e-06, - "loss": 0.2806, - "step": 35 - }, - { - "epoch": 0.10375494071146245, - "step": 35, - "train_accuracy_first_token_arguments": 0.6956521739130435, - "train_accuracy_first_token_arguments_total": 23, - "train_accuracy_recipient_token": 0.8947368421052632, + "epoch": 0.10276679841897234, + "step": 52, + "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_validate": 0.0, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.10375494071146245, - "step": 35, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 22, - "train_accuracy_recipient_token": 0.9, + "epoch": 0.10474308300395258, + "grad_norm": 2.61662813745291, + "learning_rate": 4.92998623196872e-06, + "loss": 0.2495, + "step": 53 + }, + { + "epoch": 0.10474308300395258, + "step": 53, + "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get": 0.9, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 20 }, { - "epoch": 0.10375494071146245, - "step": 35, - "train_accuracy_first_token_arguments": 0.6666666666666666, - "train_accuracy_first_token_arguments_total": 33, - "train_accuracy_recipient_token": 0.9473684210526315, + "epoch": 0.10474308300395258, + "step": 53, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_art": 0.6666666666666666, + "train_accuracy_recipient_token_all_total": 9, + "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 3, - "train_accuracy_recipient_token_inspect": 1.0, - "train_accuracy_recipient_token_inspect_total": 1, - "train_accuracy_recipient_token_repair": 1.0, - "train_accuracy_recipient_token_repair_total": 1, + "train_accuracy_recipient_token_ins": 1.0, + "train_accuracy_recipient_token_ins_total": 1, + "train_accuracy_recipient_token_rep": 1.0, + "train_accuracy_recipient_token_rep_total": 1, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, - "train_total_number_recipient_token": 19 + "train_total_number_recipient_token": 18 }, { "epoch": 0.1067193675889328, - "grad_norm": 2.28772413603275, - "learning_rate": 4.927797383160561e-06, - "loss": 0.3219, - "step": 36 + "grad_norm": 2.206222398671073, + "learning_rate": 4.926169550509787e-06, + "loss": 0.2654, + "step": 54 }, { "epoch": 0.1067193675889328, - "step": 36, - "train_accuracy_first_token_arguments": 0.9090909090909091, - "train_accuracy_first_token_arguments_total": 11, + "step": 54, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, @@ -2028,9 +1936,7 @@ }, { "epoch": 0.1067193675889328, - "step": 36, - "train_accuracy_first_token_arguments": 0.8, - "train_accuracy_first_token_arguments_total": 25, + "step": 54, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, @@ -2039,53 +1945,54 @@ "train_total_number_recipient_token": 17 }, { - "epoch": 0.1067193675889328, - "step": 36, - "train_accuracy_first_token_arguments": 0.6333333333333333, - "train_accuracy_first_token_arguments_total": 30, + "epoch": 0.10869565217391304, + "grad_norm": 1.7197195903001685, + "learning_rate": 4.9222531387232885e-06, + "loss": 0.2115, + "step": 55 + }, + { + "epoch": 0.10869565217391304, + "step": 55, "train_accuracy_recipient_token": 0.8823529411764706, - "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 0.0, "train_accuracy_recipient_token_calculate_total": 1, - "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 17 }, { - "epoch": 0.10968379446640317, - "grad_norm": 1.7442627302155183, - "learning_rate": 4.921936497481956e-06, - "loss": 0.3006, - "step": 37 - }, - { - "epoch": 0.10968379446640317, - "step": 37, - "train_accuracy_first_token_arguments": 0.8, - "train_accuracy_first_token_arguments_total": 15, - "train_accuracy_recipient_token": 0.9473684210526315, + "epoch": 0.10869565217391304, + "step": 55, + "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.10968379446640317, - "step": 37, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 13, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.11067193675889328, + "grad_norm": 2.1768195760337314, + "learning_rate": 4.918237157597574e-06, + "loss": 0.2403, + "step": 56 + }, + { + "epoch": 0.11067193675889328, + "step": 56, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, @@ -2096,12 +2003,10 @@ "train_total_number_recipient_token": 18 }, { - "epoch": 0.10968379446640317, - "step": 37, - "train_accuracy_first_token_arguments": 0.9166666666666666, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.11067193675889328, + "step": 56, + "train_accuracy_recipient_token": 0.9523809523809523, + "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, @@ -2111,23 +2016,21 @@ }, { "epoch": 0.11264822134387352, - "grad_norm": 2.0180662009823, - "learning_rate": 4.915850693946766e-06, - "loss": 0.266, - "step": 38 + "grad_norm": 1.969156469443854, + "learning_rate": 4.914121772213898e-06, + "loss": 0.28, + "step": 57 }, { "epoch": 0.11264822134387352, - "step": 38, - "train_accuracy_first_token_arguments": 0.8387096774193549, - "train_accuracy_first_token_arguments_total": 31, - "train_accuracy_recipient_token": 0.8947368421052632, + "step": 57, + "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_log": 1.0, "train_accuracy_recipient_token_log_total": 1, "train_accuracy_recipient_token_plant": 1.0, @@ -2136,11 +2039,9 @@ }, { "epoch": 0.11264822134387352, - "step": 38, - "train_accuracy_first_token_arguments": 0.6274509803921569, - "train_accuracy_first_token_arguments_total": 51, + "step": 57, "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, @@ -2153,37 +2054,33 @@ "train_total_number_recipient_token": 18 }, { - "epoch": 0.11264822134387352, - "step": 38, - "train_accuracy_first_token_arguments": 0.6862745098039216, - "train_accuracy_first_token_arguments_total": 51, + "epoch": 0.11462450592885376, + "grad_norm": 1.8351868382101044, + "learning_rate": 4.909907151739634e-06, + "loss": 0.2386, + "step": 58 + }, + { + "epoch": 0.11462450592885376, + "step": 58, "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_upgrade": 1.0, "train_accuracy_recipient_token_upgrade_total": 2, "train_total_number_recipient_token": 19 }, { - "epoch": 0.11561264822134387, - "grad_norm": 2.0013783198828907, - "learning_rate": 4.909540537725007e-06, - "loss": 0.3129, - "step": 39 - }, - { - "epoch": 0.11561264822134387, - "step": 39, - "train_accuracy_first_token_arguments": 0.8888888888888888, - "train_accuracy_first_token_arguments_total": 9, - "train_accuracy_recipient_token": 0.8823529411764706, - "train_accuracy_recipient_token_all": 0.8, + "epoch": 0.11462450592885376, + "step": 58, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, @@ -2192,10 +2089,15 @@ "train_total_number_recipient_token": 17 }, { - "epoch": 0.11561264822134387, - "step": 39, - "train_accuracy_first_token_arguments": 0.9375, - "train_accuracy_first_token_arguments_total": 16, + "epoch": 0.116600790513834, + "grad_norm": 3.0727021394959544, + "learning_rate": 4.905593469421323e-06, + "loss": 0.3147, + "step": 59 + }, + { + "epoch": 0.116600790513834, + "step": 59, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, @@ -2210,10 +2112,8 @@ "train_total_number_recipient_token": 19 }, { - "epoch": 0.11561264822134387, - "step": 39, - "train_accuracy_first_token_arguments": 0.7560975609756098, - "train_accuracy_first_token_arguments_total": 41, + "epoch": 0.116600790513834, + "step": 59, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, @@ -2229,46 +2129,47 @@ }, { "epoch": 0.11857707509881422, - "grad_norm": 2.383877337768723, - "learning_rate": 4.903006614821645e-06, - "loss": 0.2901, - "step": 40 + "grad_norm": 2.181993064766873, + "learning_rate": 4.901180902577549e-06, + "loss": 0.2224, + "step": 60 }, { "epoch": 0.11857707509881422, - "step": 40, - "train_accuracy_first_token_arguments": 0.7666666666666667, - "train_accuracy_first_token_arguments_total": 30, - "train_accuracy_recipient_token": 0.9, + "step": 60, + "train_accuracy_recipient_token": 0.85, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 2, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 2, "train_accuracy_recipient_token_translate": 0.5, "train_accuracy_recipient_token_translate_total": 2, "train_total_number_recipient_token": 20 }, { "epoch": 0.11857707509881422, - "step": 40, - "train_accuracy_first_token_arguments": 0.8571428571428571, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 0.95, + "step": 60, + "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.9, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 20 }, { - "epoch": 0.11857707509881422, - "step": 40, - "train_accuracy_first_token_arguments": 0.9183673469387755, - "train_accuracy_first_token_arguments_total": 49, - "train_accuracy_recipient_token": 0.7, - "train_accuracy_recipient_token_all": 0.6666666666666666, + "epoch": 0.12055335968379446, + "grad_norm": 2.1374748114773303, + "learning_rate": 4.896669632591652e-06, + "loss": 0.1559, + "step": 61 + }, + { + "epoch": 0.12055335968379446, + "step": 61, + "train_accuracy_recipient_token": 0.8, + "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 15, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, @@ -2277,52 +2178,46 @@ "train_total_number_recipient_token": 20 }, { - "epoch": 0.12154150197628459, - "grad_norm": 2.3845477671070725, - "learning_rate": 4.8962495320221714e-06, - "loss": 0.2275, - "step": 41 - }, - { - "epoch": 0.12154150197628459, - "step": 41, - "train_accuracy_first_token_arguments": 0.65, - "train_accuracy_first_token_arguments_total": 20, - "train_accuracy_recipient_token": 0.8888888888888888, + "epoch": 0.12055335968379446, + "step": 61, + "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare": 0.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.12154150197628459, - "step": 41, - "train_accuracy_first_token_arguments": 0.8888888888888888, - "train_accuracy_first_token_arguments_total": 18, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.1225296442687747, + "grad_norm": 2.467435650737327, + "learning_rate": 4.892059844904273e-06, + "loss": 0.2946, + "step": 62 + }, + { + "epoch": 0.1225296442687747, + "step": 62, + "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 1, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.875, "train_accuracy_recipient_token_get_total": 8, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 20 }, { - "epoch": 0.12154150197628459, - "step": 41, - "train_accuracy_first_token_arguments": 0.5769230769230769, - "train_accuracy_first_token_arguments_total": 52, - "train_accuracy_recipient_token": 0.8947368421052632, + "epoch": 0.1225296442687747, + "step": 62, + "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, @@ -2331,22 +2226,20 @@ "train_accuracy_recipient_token_list_total": 1, "train_accuracy_recipient_token_merge": 1.0, "train_accuracy_recipient_token_merge_total": 1, - "train_accuracy_recipient_token_recommend": 0.75, - "train_accuracy_recipient_token_recommend_total": 4, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.12450592885375494, - "grad_norm": 2.2451847636665194, - "learning_rate": 4.8892699168362626e-06, - "loss": 0.2985, - "step": 42 + "grad_norm": 3.655476447614754, + "learning_rate": 4.8873517290057265e-06, + "loss": 0.2478, + "step": 63 }, { "epoch": 0.12450592885375494, - "step": 42, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 20, + "step": 63, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, @@ -2358,11 +2251,9 @@ }, { "epoch": 0.12450592885375494, - "step": 42, - "train_accuracy_first_token_arguments": 0.8, - "train_accuracy_first_token_arguments_total": 15, - "train_accuracy_recipient_token": 0.9375, - "train_accuracy_recipient_token_all": 0.9, + "step": 63, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, @@ -2373,10 +2264,15 @@ "train_total_number_recipient_token": 16 }, { - "epoch": 0.12450592885375494, - "step": 42, - "train_accuracy_first_token_arguments": 0.8823529411764706, - "train_accuracy_first_token_arguments_total": 17, + "epoch": 0.12648221343873517, + "grad_norm": 2.202751393561073, + "learning_rate": 4.882545478428219e-06, + "loss": 0.2624, + "step": 64 + }, + { + "epoch": 0.12648221343873517, + "step": 64, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, @@ -2385,22 +2281,13 @@ "train_total_number_recipient_token": 19 }, { - "epoch": 0.1274703557312253, - "grad_norm": 2.1597197573051443, - "learning_rate": 4.8820684174394935e-06, - "loss": 0.3257, - "step": 43 - }, - { - "epoch": 0.1274703557312253, - "step": 43, - "train_accuracy_first_token_arguments": 0.6428571428571429, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 0.9375, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.12648221343873517, + "step": 64, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 1.0, @@ -2408,24 +2295,27 @@ "train_total_number_recipient_token": 16 }, { - "epoch": 0.1274703557312253, - "step": 43, - "train_accuracy_first_token_arguments": 0.75, - "train_accuracy_first_token_arguments_total": 8, - "train_accuracy_recipient_token": 0.8235294117647058, - "train_accuracy_recipient_token_all": 0.8461538461538461, + "epoch": 0.12845849802371542, + "grad_norm": 2.26174226561165, + "learning_rate": 4.8776412907378845e-06, + "loss": 0.3378, + "step": 65 + }, + { + "epoch": 0.12845849802371542, + "step": 65, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.9230769230769231, "train_accuracy_recipient_token_all_total": 13, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 17 }, { - "epoch": 0.1274703557312253, - "step": 43, - "train_accuracy_first_token_arguments": 0.8333333333333334, - "train_accuracy_first_token_arguments_total": 6, + "epoch": 0.12845849802371542, + "step": 65, "train_accuracy_recipient_token": 0.8666666666666667, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, @@ -2435,16 +2325,14 @@ }, { "epoch": 0.13043478260869565, - "grad_norm": 2.4009249785875033, - "learning_rate": 4.874645702613152e-06, - "loss": 0.4131, - "step": 44 + "grad_norm": 2.6607823334587644, + "learning_rate": 4.872639367526672e-06, + "loss": 0.3342, + "step": 66 }, { "epoch": 0.13043478260869565, - "step": 44, - "train_accuracy_first_token_arguments": 0.7727272727272727, - "train_accuracy_first_token_arguments_total": 22, + "step": 66, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, @@ -2452,7 +2340,7 @@ "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, - "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, @@ -2460,88 +2348,85 @@ }, { "epoch": 0.13043478260869565, - "step": 44, - "train_accuracy_first_token_arguments": 0.8095238095238095, - "train_accuracy_first_token_arguments_total": 21, + "step": 66, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_artifact": 1.0, - "train_accuracy_recipient_token_artifact_total": 3, + "train_accuracy_recipient_token_art": 1.0, + "train_accuracy_recipient_token_art_total": 3, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 20 }, { - "epoch": 0.13043478260869565, - "step": 44, - "train_accuracy_first_token_arguments": 0.6, - "train_accuracy_first_token_arguments_total": 35, - "train_accuracy_recipient_token": 0.8947368421052632, - "train_accuracy_recipient_token_all": 0.9090909090909091, + "epoch": 0.1324110671936759, + "grad_norm": 2.3676089848491957, + "learning_rate": 4.8675399144040535e-06, + "loss": 0.4039, + "step": 67 + }, + { + "epoch": 0.1324110671936759, + "step": 67, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_build": 1.0, "train_accuracy_recipient_token_build_total": 1, - "train_accuracy_recipient_token_calculate": 0.75, + "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 4, "train_accuracy_recipient_token_fire": 1.0, "train_accuracy_recipient_token_fire_total": 3, "train_total_number_recipient_token": 19 }, { - "epoch": 0.13339920948616601, - "grad_norm": 2.2178791639658186, - "learning_rate": 4.867002461682129e-06, - "loss": 0.401, - "step": 45 - }, - { - "epoch": 0.13339920948616601, - "step": 45, - "train_accuracy_first_token_arguments": 0.8235294117647058, - "train_accuracy_first_token_arguments_total": 34, - "train_accuracy_recipient_token": 0.7368421052631579, + "epoch": 0.1324110671936759, + "step": 67, + "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 0.5, "train_accuracy_recipient_token_compare_total": 2, - "train_accuracy_recipient_token_find": 0.4, + "train_accuracy_recipient_token_find": 0.8, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 19 }, { - "epoch": 0.13339920948616601, - "step": 45, - "train_accuracy_first_token_arguments": 0.9333333333333333, - "train_accuracy_first_token_arguments_total": 15, - "train_accuracy_recipient_token": 0.8947368421052632, + "epoch": 0.13438735177865613, + "grad_norm": 2.1673908938305124, + "learning_rate": 4.862343140988573e-06, + "loss": 0.1757, + "step": 68 + }, + { + "epoch": 0.13438735177865613, + "step": 68, + "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_ex": 0.3333333333333333, + "train_accuracy_recipient_token_ex": 0.6666666666666666, "train_accuracy_recipient_token_ex_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.13339920948616601, - "step": 45, - "train_accuracy_first_token_arguments": 0.8709677419354839, - "train_accuracy_first_token_arguments_total": 31, - "train_accuracy_recipient_token": 0.85, + "epoch": 0.13438735177865613, + "step": 68, + "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_log": 1.0, "train_accuracy_recipient_token_log_total": 1, - "train_accuracy_recipient_token_review": 0.5, + "train_accuracy_recipient_token_review": 1.0, "train_accuracy_recipient_token_review_total": 2, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, @@ -2549,20 +2434,18 @@ }, { "epoch": 0.13636363636363635, - "grad_norm": 1.7428907811945011, - "learning_rate": 4.8591394044508985e-06, - "loss": 0.2065, - "step": 46 + "grad_norm": 1.8900015178295946, + "learning_rate": 4.857049260899233e-06, + "loss": 0.2066, + "step": 69 }, { "epoch": 0.13636363636363635, - "step": 46, - "train_accuracy_first_token_arguments": 0.9, - "train_accuracy_first_token_arguments_total": 200, - "train_accuracy_recipient_token": 0.8823529411764706, + "step": 69, + "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_filter": 0.75, + "train_accuracy_recipient_token_filter": 1.0, "train_accuracy_recipient_token_filter_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, @@ -2570,10 +2453,8 @@ }, { "epoch": 0.13636363636363635, - "step": 46, - "train_accuracy_first_token_arguments": 0.8235294117647058, - "train_accuracy_first_token_arguments_total": 17, - "train_accuracy_recipient_token": 0.85, + "step": 69, + "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, @@ -2587,16 +2468,21 @@ "train_total_number_recipient_token": 20 }, { - "epoch": 0.13636363636363635, - "step": 46, - "train_accuracy_first_token_arguments": 0.9259259259259259, - "train_accuracy_first_token_arguments_total": 27, + "epoch": 0.1383399209486166, + "grad_norm": 2.536038040555671, + "learning_rate": 4.851658491746707e-06, + "loss": 0.2677, + "step": 70 + }, + { + "epoch": 0.1383399209486166, + "step": 70, "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, - "train_accuracy_recipient_token_city": 0.75, + "train_accuracy_recipient_token_city": 1.0, "train_accuracy_recipient_token_city_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, @@ -2605,17 +2491,8 @@ "train_total_number_recipient_token": 18 }, { - "epoch": 0.13932806324110672, - "grad_norm": 2.3699963492078684, - "learning_rate": 4.851057261137608e-06, - "loss": 0.2718, - "step": 47 - }, - { - "epoch": 0.13932806324110672, - "step": 47, - "train_accuracy_first_token_arguments": 0.8, - "train_accuracy_first_token_arguments_total": 15, + "epoch": 0.1383399209486166, + "step": 70, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, @@ -2626,28 +2503,31 @@ "train_total_number_recipient_token": 18 }, { - "epoch": 0.13932806324110672, - "step": 47, - "train_accuracy_first_token_arguments": 0.59375, - "train_accuracy_first_token_arguments_total": 32, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.14031620553359683, + "grad_norm": 1.7887214738919732, + "learning_rate": 4.846171055124401e-06, + "loss": 0.1541, + "step": 71 + }, + { + "epoch": 0.14031620553359683, + "step": 71, + "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 2, + "train_accuracy_recipient_token_analy": 0.5, + "train_accuracy_recipient_token_analy_total": 2, "train_accuracy_recipient_token_detect": 1.0, "train_accuracy_recipient_token_detect_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 2, + "train_accuracy_recipient_token_ident": 0.5, + "train_accuracy_recipient_token_ident_total": 2, "train_total_number_recipient_token": 19 }, { - "epoch": 0.13932806324110672, - "step": 47, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 22, + "epoch": 0.14031620553359683, + "step": 71, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.8333333333333334, "train_accuracy_recipient_token_all_total": 12, @@ -2659,34 +2539,30 @@ }, { "epoch": 0.1422924901185771, - "grad_norm": 2.553448245564224, - "learning_rate": 4.842756782306261e-06, - "loss": 0.3375, - "step": 48 + "grad_norm": 3.415062910528756, + "learning_rate": 4.8405871765993435e-06, + "loss": 0.3752, + "step": 72 }, { "epoch": 0.1422924901185771, - "step": 48, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 0.6470588235294118, + "step": 72, + "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_organ": 0.0, + "train_accuracy_recipient_token_organ": 1.0, "train_accuracy_recipient_token_organ_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.1422924901185771, - "step": 48, - "train_accuracy_first_token_arguments": 0.5384615384615384, - "train_accuracy_first_token_arguments_total": 13, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9, + "step": 72, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 2, @@ -2699,51 +2575,52 @@ "train_total_number_recipient_token": 19 }, { - "epoch": 0.1422924901185771, - "step": 48, - "train_accuracy_first_token_arguments": 0.7894736842105263, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.95, + "epoch": 0.1442687747035573, + "grad_norm": 2.100833710936171, + "learning_rate": 4.834907085702909e-06, + "loss": 0.2565, + "step": 73 + }, + { + "epoch": 0.1442687747035573, + "step": 73, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_c": 0.0, + "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 2, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 2, "train_total_number_recipient_token": 20 }, { - "epoch": 0.14525691699604742, - "grad_norm": 2.047932501451832, - "learning_rate": 4.8342387387970105e-06, - "loss": 0.2978, - "step": 49 - }, - { - "epoch": 0.14525691699604742, - "step": 49, - "train_accuracy_first_token_arguments": 0.9090909090909091, - "train_accuracy_first_token_arguments_total": 33, - "train_accuracy_recipient_token": 0.95, + "epoch": 0.1442687747035573, + "step": 73, + "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.875, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 20 }, { - "epoch": 0.14525691699604742, - "step": 49, - "train_accuracy_first_token_arguments": 0.85, - "train_accuracy_first_token_arguments_total": 40, + "epoch": 0.14624505928853754, + "grad_norm": 1.8239921154406062, + "learning_rate": 4.829131015921386e-06, + "loss": 0.1924, + "step": 74 + }, + { + "epoch": 0.14624505928853754, + "step": 74, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 8, @@ -2751,25 +2628,23 @@ "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_setup": 1.0, + "train_accuracy_recipient_token_setup": 0.0, "train_accuracy_recipient_token_setup_total": 1, "train_total_number_recipient_token": 16 }, { - "epoch": 0.14525691699604742, - "step": 49, - "train_accuracy_first_token_arguments": 0.8181818181818182, - "train_accuracy_first_token_arguments_total": 55, - "train_accuracy_recipient_token": 0.8235294117647058, - "train_accuracy_recipient_token_all": 0.8, + "epoch": 0.14624505928853754, + "step": 74, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, - "train_accuracy_recipient_token_g": 0.6666666666666666, + "train_accuracy_recipient_token_g": 1.0, "train_accuracy_recipient_token_g_total": 3, "train_accuracy_recipient_token_track": 1.0, "train_accuracy_recipient_token_track_total": 1, @@ -2777,18 +2652,16 @@ }, { "epoch": 0.1482213438735178, - "grad_norm": 2.251584762112145, - "learning_rate": 4.825503921654582e-06, - "loss": 0.1679, - "step": 50 + "grad_norm": 2.1914980081973794, + "learning_rate": 4.82325920468638e-06, + "loss": 0.1502, + "step": 75 }, { "epoch": 0.1482213438735178, - "step": 50, - "train_accuracy_first_token_arguments": 0.9090909090909091, - "train_accuracy_first_token_arguments_total": 11, - "train_accuracy_recipient_token": 0.8125, - "train_accuracy_recipient_token_all": 0.7, + "step": 75, + "train_accuracy_recipient_token": 0.9375, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, @@ -2800,11 +2673,9 @@ }, { "epoch": 0.1482213438735178, - "step": 50, - "train_accuracy_first_token_arguments": 0.6923076923076923, - "train_accuracy_first_token_arguments_total": 13, - "train_accuracy_recipient_token": 0.875, - "train_accuracy_recipient_token_all": 0.8, + "step": 75, + "train_accuracy_recipient_token": 0.9375, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, @@ -2815,15 +2686,20 @@ "train_total_number_recipient_token": 16 }, { - "epoch": 0.1482213438735178, - "step": 50, - "train_accuracy_first_token_arguments": 0.8, - "train_accuracy_first_token_arguments_total": 10, - "train_accuracy_recipient_token": 0.8823529411764706, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.15019762845849802, + "grad_norm": 2.315784596207221, + "learning_rate": 4.817291893365055e-06, + "loss": 0.1811, + "step": 76 + }, + { + "epoch": 0.15019762845849802, + "step": 76, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 0.5, - "train_accuracy_recipient_token_analyze_total": 2, + "train_accuracy_recipient_token_analy": 0.5, + "train_accuracy_recipient_token_analy_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, @@ -2833,40 +2709,36 @@ "train_total_number_recipient_token": 17 }, { - "epoch": 0.15118577075098813, - "grad_norm": 2.326983001481526, - "learning_rate": 4.816553142054806e-06, - "loss": 0.2279, - "step": 51 - }, - { - "epoch": 0.15118577075098813, - "step": 51, - "train_accuracy_first_token_arguments": 0.7692307692307693, - "train_accuracy_first_token_arguments_total": 13, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.15019762845849802, + "step": 76, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, - "train_accuracy_recipient_token_plan": 1.0, + "train_accuracy_recipient_token_plan": 0.0, "train_accuracy_recipient_token_plan_total": 1, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 3, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 3, "train_total_number_recipient_token": 19 }, { - "epoch": 0.15118577075098813, - "step": 51, - "train_accuracy_first_token_arguments": 0.875, - "train_accuracy_first_token_arguments_total": 16, - "train_accuracy_recipient_token": 0.875, - "train_accuracy_recipient_token_all": 0.8181818181818182, + "epoch": 0.15217391304347827, + "grad_norm": 3.9917762289000542, + "learning_rate": 4.811229327250204e-06, + "loss": 0.2349, + "step": 77 + }, + { + "epoch": 0.15217391304347827, + "step": 77, + "train_accuracy_recipient_token": 0.9375, + "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, @@ -2876,10 +2748,8 @@ "train_total_number_recipient_token": 16 }, { - "epoch": 0.15118577075098813, - "step": 51, - "train_accuracy_first_token_arguments": 0.8421052631578947, - "train_accuracy_first_token_arguments_total": 19, + "epoch": 0.15217391304347827, + "step": 77, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, @@ -2887,41 +2757,37 @@ "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.1541501976284585, - "grad_norm": 2.104029483073964, - "learning_rate": 4.807387231229287e-06, - "loss": 0.3356, - "step": 52 + "grad_norm": 2.2903178058367315, + "learning_rate": 4.805071755550177e-06, + "loss": 0.3338, + "step": 78 }, { "epoch": 0.1541501976284585, - "step": 52, - "train_accuracy_first_token_arguments": 0.7, - "train_accuracy_first_token_arguments_total": 10, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9, + "step": 78, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.1541501976284585, - "step": 52, - "train_accuracy_first_token_arguments": 0.8620689655172413, - "train_accuracy_first_token_arguments_total": 29, - "train_accuracy_recipient_token": 0.9473684210526315, + "step": 78, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_evaluate": 1.0, @@ -2935,12 +2801,17 @@ "train_total_number_recipient_token": 19 }, { - "epoch": 0.1541501976284585, - "step": 52, - "train_accuracy_first_token_arguments": 0.782608695652174, - "train_accuracy_first_token_arguments_total": 23, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.15612648221343872, + "grad_norm": 2.526628452772315, + "learning_rate": 4.7988194313786275e-06, + "loss": 0.2483, + "step": 79 + }, + { + "epoch": 0.15612648221343872, + "step": 79, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_display": 1.0, "train_accuracy_recipient_token_display_total": 1, @@ -2948,28 +2819,19 @@ "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_h": 1.0, "train_accuracy_recipient_token_h_total": 3, - "train_accuracy_recipient_token_suggest": 1.0, - "train_accuracy_recipient_token_suggest_total": 2, + "train_accuracy_recipient_token_s": 1.0, + "train_accuracy_recipient_token_s_total": 2, "train_total_number_recipient_token": 18 }, { - "epoch": 0.15711462450592886, - "grad_norm": 2.1920664784328907, - "learning_rate": 4.798007040388212e-06, - "loss": 0.2247, - "step": 53 - }, - { - "epoch": 0.15711462450592886, - "step": 53, - "train_accuracy_first_token_arguments": 0.9428571428571428, - "train_accuracy_first_token_arguments_total": 35, + "epoch": 0.15612648221343872, + "step": 79, "train_accuracy_recipient_token": 0.8947368421052632, - "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_calculate": 0.0, "train_accuracy_recipient_token_calculate_total": 1, - "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check": 0.5, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, @@ -2978,22 +2840,25 @@ "train_total_number_recipient_token": 19 }, { - "epoch": 0.15711462450592886, - "step": 53, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 11, - "train_accuracy_recipient_token": 0.9411764705882353, + "epoch": 0.15810276679841898, + "grad_norm": 1.3952343413076715, + "learning_rate": 4.7924726117441135e-06, + "loss": 0.1396, + "step": 80 + }, + { + "epoch": 0.15810276679841898, + "step": 80, + "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { - "epoch": 0.15711462450592886, - "step": 53, - "train_accuracy_first_token_arguments": 0.6857142857142857, - "train_accuracy_first_token_arguments_total": 35, + "epoch": 0.15810276679841898, + "step": 80, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, @@ -3007,30 +2872,26 @@ }, { "epoch": 0.1600790513833992, - "grad_norm": 1.6828766971614149, - "learning_rate": 4.788413440641297e-06, - "loss": 0.2468, - "step": 54 + "grad_norm": 1.8632305360852095, + "learning_rate": 4.786031557539532e-06, + "loss": 0.2135, + "step": 81 }, { "epoch": 0.1600790513833992, - "step": 54, - "train_accuracy_first_token_arguments": 0.9130434782608695, - "train_accuracy_first_token_arguments_total": 23, - "train_accuracy_recipient_token": 0.9444444444444444, + "step": 81, + "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_flag": 0.6666666666666666, "train_accuracy_recipient_token_flag_total": 3, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 18 }, { "epoch": 0.1600790513833992, - "step": 54, - "train_accuracy_first_token_arguments": 0.8461538461538461, - "train_accuracy_first_token_arguments_total": 26, + "step": 81, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, @@ -3041,33 +2902,29 @@ "train_total_number_recipient_token": 19 }, { - "epoch": 0.1600790513833992, - "step": 54, - "train_accuracy_first_token_arguments": 0.5476190476190477, - "train_accuracy_first_token_arguments_total": 42, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.16205533596837945, + "grad_norm": 2.2968333669223, + "learning_rate": 4.779496533531393e-06, + "loss": 0.3168, + "step": 82 + }, + { + "epoch": 0.16205533596837945, + "step": 82, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, - "train_accuracy_recipient_token_log": 1.0, + "train_accuracy_recipient_token_log": 0.0, "train_accuracy_recipient_token_log_total": 1, "train_total_number_recipient_token": 17 }, { - "epoch": 0.16304347826086957, - "grad_norm": 2.3713370104618146, - "learning_rate": 4.778607322916896e-06, - "loss": 0.4323, - "step": 55 - }, - { - "epoch": 0.16304347826086957, - "step": 55, - "train_accuracy_first_token_arguments": 0.8627450980392157, - "train_accuracy_first_token_arguments_total": 51, - "train_accuracy_recipient_token": 0.8823529411764706, - "train_accuracy_recipient_token_all": 0.8, + "epoch": 0.16205533596837945, + "step": 82, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 1, @@ -3080,12 +2937,17 @@ "train_total_number_recipient_token": 17 }, { - "epoch": 0.16304347826086957, - "step": 55, - "train_accuracy_first_token_arguments": 0.9473684210526315, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.16403162055335968, + "grad_norm": 2.249381257252995, + "learning_rate": 4.7728678083489375e-06, + "loss": 0.3738, + "step": 83 + }, + { + "epoch": 0.16403162055335968, + "step": 83, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_build": 0.5, "train_accuracy_recipient_token_build_total": 2, @@ -3096,16 +2958,14 @@ "train_total_number_recipient_token": 18 }, { - "epoch": 0.16304347826086957, - "step": 55, - "train_accuracy_first_token_arguments": 0.7368421052631579, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.9473684210526315, + "epoch": 0.16403162055335968, + "step": 83, + "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, - "train_accuracy_recipient_token_fetch": 1.0, + "train_accuracy_recipient_token_fetch": 0.0, "train_accuracy_recipient_token_fetch_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, @@ -3113,18 +2973,16 @@ }, { "epoch": 0.16600790513833993, - "grad_norm": 1.892415917915294, - "learning_rate": 4.7685895978792564e-06, - "loss": 0.2455, - "step": 56 + "grad_norm": 1.8402947400679557, + "learning_rate": 4.766145654473096e-06, + "loss": 0.1812, + "step": 84 }, { "epoch": 0.16600790513833993, - "step": 56, - "train_accuracy_first_token_arguments": 0.68, - "train_accuracy_first_token_arguments_total": 25, - "train_accuracy_recipient_token": 0.8823529411764706, - "train_accuracy_recipient_token_all": 0.9090909090909091, + "step": 84, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, @@ -3138,16 +2996,14 @@ }, { "epoch": 0.16600790513833993, - "step": 56, - "train_accuracy_first_token_arguments": 0.9761904761904762, - "train_accuracy_first_token_arguments_total": 42, + "step": 84, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 1, - "train_accuracy_recipient_token_explain": 1.0, - "train_accuracy_recipient_token_explain_total": 1, + "train_accuracy_recipient_token_ex": 1.0, + "train_accuracy_recipient_token_ex_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_number": 1.0, @@ -3155,10 +3011,15 @@ "train_total_number_recipient_token": 20 }, { - "epoch": 0.16600790513833993, - "step": 56, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 14, + "epoch": 0.16798418972332016, + "grad_norm": 1.9475841569891188, + "learning_rate": 4.7593303482252835e-06, + "loss": 0.2456, + "step": 85 + }, + { + "epoch": 0.16798418972332016, + "step": 85, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, @@ -3171,17 +3032,8 @@ "train_total_number_recipient_token": 20 }, { - "epoch": 0.16897233201581027, - "grad_norm": 1.7624413054210974, - "learning_rate": 4.7583611958439514e-06, - "loss": 0.2711, - "step": 57 - }, - { - "epoch": 0.16897233201581027, - "step": 57, - "train_accuracy_first_token_arguments": 0.8205128205128205, - "train_accuracy_first_token_arguments_total": 39, + "epoch": 0.16798418972332016, + "step": 85, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, @@ -3191,29 +3043,32 @@ "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_hire": 1.0, "train_accuracy_recipient_token_hire_total": 1, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.16897233201581027, - "step": 57, - "train_accuracy_first_token_arguments": 0.8717948717948718, - "train_accuracy_first_token_arguments_total": 39, - "train_accuracy_recipient_token": 0.9411764705882353, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.16996047430830039, + "grad_norm": 2.2098134551670943, + "learning_rate": 4.752422169756048e-06, + "loss": 0.2644, + "step": 86 + }, + { + "epoch": 0.16996047430830039, + "step": 86, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { - "epoch": 0.16897233201581027, - "step": 57, - "train_accuracy_first_token_arguments": 0.7735849056603774, - "train_accuracy_first_token_arguments_total": 106, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.16996047430830039, + "step": 86, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_ex": 1.0, "train_accuracy_recipient_token_ex_total": 3, @@ -3225,30 +3080,26 @@ }, { "epoch": 0.17193675889328064, - "grad_norm": 2.096786172099922, - "learning_rate": 4.747923066691487e-06, - "loss": 0.2883, - "step": 58 + "grad_norm": 1.9178477232600728, + "learning_rate": 4.745421403033548e-06, + "loss": 0.2145, + "step": 87 }, { "epoch": 0.17193675889328064, - "step": 58, - "train_accuracy_first_token_arguments": 0.9090909090909091, - "train_accuracy_first_token_arguments_total": 11, - "train_accuracy_recipient_token": 1.0, + "step": 87, + "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.9, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 20 }, { "epoch": 0.17193675889328064, - "step": 58, - "train_accuracy_first_token_arguments": 0.58, - "train_accuracy_first_token_arguments_total": 50, - "train_accuracy_recipient_token": 0.95, - "train_accuracy_recipient_token_all": 0.9090909090909091, + "step": 87, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_f": 1.0, "train_accuracy_recipient_token_f_total": 1, @@ -3256,15 +3107,20 @@ "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 1, - "train_accuracy_recipient_token_wall": 1.0, - "train_accuracy_recipient_token_wall_total": 1, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_window": 1.0, + "train_accuracy_recipient_token_window_total": 1, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.17193675889328064, - "step": 58, - "train_accuracy_first_token_arguments": 0.5, - "train_accuracy_first_token_arguments_total": 12, + "epoch": 0.17391304347826086, + "grad_norm": 1.4953922375484354, + "learning_rate": 4.738328335831883e-06, + "loss": 0.1738, + "step": 88 + }, + { + "epoch": 0.17391304347826086, + "step": 88, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, @@ -3272,25 +3128,16 @@ "train_accuracy_recipient_token_arch_total": 1, "train_accuracy_recipient_token_building": 1.0, "train_accuracy_recipient_token_building_total": 1, - "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 18 }, { - "epoch": 0.17490118577075098, - "grad_norm": 1.6122336767210732, - "learning_rate": 4.7372761797790836e-06, - "loss": 0.2393, - "step": 59 - }, - { - "epoch": 0.17490118577075098, - "step": 59, - "train_accuracy_first_token_arguments": 0.6153846153846154, - "train_accuracy_first_token_arguments_total": 26, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.17391304347826086, + "step": 88, + "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, @@ -3299,31 +3146,34 @@ "train_accuracy_recipient_token_detect_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, - "train_accuracy_recipient_token_truck": 1.0, - "train_accuracy_recipient_token_truck_total": 1, + "train_accuracy_recipient_token_tr": 1.0, + "train_accuracy_recipient_token_tr_total": 1, "train_total_number_recipient_token": 17 }, { - "epoch": 0.17490118577075098, - "step": 59, - "train_accuracy_first_token_arguments": 0.6428571428571429, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 0.9411764705882353, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.17588932806324112, + "grad_norm": 1.9501250918192976, + "learning_rate": 4.7311432597192655e-06, + "loss": 0.224, + "step": 89 + }, + { + "epoch": 0.17588932806324112, + "step": 89, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 17 }, { - "epoch": 0.17490118577075098, - "step": 59, - "train_accuracy_first_token_arguments": 0.75, - "train_accuracy_first_token_arguments_total": 12, + "epoch": 0.17588932806324112, + "step": 89, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, @@ -3335,16 +3185,14 @@ }, { "epoch": 0.17786561264822134, - "grad_norm": 1.7027092503573524, - "learning_rate": 4.726421523850662e-06, - "loss": 0.1559, - "step": 60 + "grad_norm": 1.7132307123231727, + "learning_rate": 4.72386647004603e-06, + "loss": 0.1349, + "step": 90 }, { "epoch": 0.17786561264822134, - "step": 60, - "train_accuracy_first_token_arguments": 0.7647058823529411, - "train_accuracy_first_token_arguments_total": 17, + "step": 90, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, @@ -3360,27 +3208,30 @@ }, { "epoch": 0.17786561264822134, - "step": 60, - "train_accuracy_first_token_arguments": 0.8947368421052632, - "train_accuracy_first_token_arguments_total": 19, + "step": 90, "train_accuracy_recipient_token": 0.8125, - "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_generate": 0.0, + "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 16 }, { - "epoch": 0.17786561264822134, - "step": 60, - "train_accuracy_first_token_arguments": 0.7692307692307693, - "train_accuracy_first_token_arguments_total": 26, + "epoch": 0.17984189723320157, + "grad_norm": 1.862599179005162, + "learning_rate": 4.716498265932501e-06, + "loss": 0.2707, + "step": 91 + }, + { + "epoch": 0.17984189723320157, + "step": 91, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, @@ -3390,22 +3241,13 @@ "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 2, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 2, "train_total_number_recipient_token": 20 }, { - "epoch": 0.1808300395256917, - "grad_norm": 1.640939742247205, - "learning_rate": 4.715360106945015e-06, - "loss": 0.2414, - "step": 61 - }, - { - "epoch": 0.1808300395256917, - "step": 61, - "train_accuracy_first_token_arguments": 0.875, - "train_accuracy_first_token_arguments_total": 24, + "epoch": 0.17984189723320157, + "step": 91, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, @@ -3418,30 +3260,33 @@ "train_total_number_recipient_token": 19 }, { - "epoch": 0.1808300395256917, - "step": 61, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.9, + "epoch": 0.18181818181818182, + "grad_norm": 1.5411465996887044, + "learning_rate": 4.7090389502566884e-06, + "loss": 0.1157, + "step": 92 + }, + { + "epoch": 0.18181818181818182, + "step": 92, + "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_plant": 0.75, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_plant": 1.0, "train_accuracy_recipient_token_plant_total": 4, "train_total_number_recipient_token": 20 }, { - "epoch": 0.1808300395256917, - "step": 61, - "train_accuracy_first_token_arguments": 0.8947368421052632, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.8823529411764706, + "epoch": 0.18181818181818182, + "step": 92, + "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, @@ -3449,28 +3294,24 @@ }, { "epoch": 0.18379446640316205, - "grad_norm": 2.3050792009276817, - "learning_rate": 4.7040929563021975e-06, - "loss": 0.2552, - "step": 62 + "grad_norm": 2.100818863534559, + "learning_rate": 4.701488829641845e-06, + "loss": 0.2582, + "step": 93 }, { "epoch": 0.18379446640316205, - "step": 62, - "train_accuracy_first_token_arguments": 0.9411764705882353, - "train_accuracy_first_token_arguments_total": 17, - "train_accuracy_recipient_token": 0.9411764705882353, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, + "step": 93, + "train_accuracy_recipient_token": 0.875, + "train_accuracy_recipient_token_all": 0.7777777777777778, + "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, - "train_total_number_recipient_token": 17 + "train_total_number_recipient_token": 16 }, { "epoch": 0.18379446640316205, - "step": 62, - "train_accuracy_first_token_arguments": 0.8, - "train_accuracy_first_token_arguments_total": 20, + "step": 93, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, @@ -3483,78 +3324,77 @@ "train_total_number_recipient_token": 18 }, { - "epoch": 0.18379446640316205, - "step": 62, - "train_accuracy_first_token_arguments": 0.8695652173913043, - "train_accuracy_first_token_arguments_total": 23, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.1857707509881423, + "grad_norm": 2.5798405707090852, + "learning_rate": 4.693848214443858e-06, + "loss": 0.4023, + "step": 94 + }, + { + "epoch": 0.1857707509881423, + "step": 94, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, - "train_accuracy_recipient_token_flower": 1.0, + "train_accuracy_recipient_token_flower": 0.6666666666666666, "train_accuracy_recipient_token_flower_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.18675889328063242, - "grad_norm": 2.559937679792714, - "learning_rate": 4.6926211182681295e-06, - "loss": 0.4525, - "step": 63 - }, - { - "epoch": 0.18675889328063242, - "step": 63, - "train_accuracy_first_token_arguments": 0.9565217391304348, - "train_accuracy_first_token_arguments_total": 46, - "train_accuracy_recipient_token": 0.7058823529411765, + "epoch": 0.1857707509881423, + "step": 94, + "train_accuracy_recipient_token": 0.7647058823529411, "train_accuracy_recipient_token_all": 0.6666666666666666, "train_accuracy_recipient_token_all_total": 12, - "train_accuracy_recipient_token_compare": 0.0, + "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 2, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 2, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 17 }, { - "epoch": 0.18675889328063242, - "step": 63, - "train_accuracy_first_token_arguments": 0.6190476190476191, - "train_accuracy_first_token_arguments_total": 21, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.18774703557312253, + "grad_norm": 2.1919145048059283, + "learning_rate": 4.686117418738489e-06, + "loss": 0.2804, + "step": 95 + }, + { + "epoch": 0.18774703557312253, + "step": 95, + "train_accuracy_recipient_token": 0.7894736842105263, + "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_emergency": 1.0, - "train_accuracy_recipient_token_emergency_total": 1, - "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_em": 1.0, + "train_accuracy_recipient_token_em_total": 1, + "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_first": 0.0, + "train_accuracy_recipient_token_first": 1.0, "train_accuracy_recipient_token_first_total": 1, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 19 }, { - "epoch": 0.18675889328063242, - "step": 63, - "train_accuracy_first_token_arguments": 0.6129032258064516, - "train_accuracy_first_token_arguments_total": 31, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.18774703557312253, + "step": 95, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_explain": 1.0, - "train_accuracy_recipient_token_explain_total": 1, + "train_accuracy_recipient_token_ex": 1.0, + "train_accuracy_recipient_token_ex_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, @@ -3563,16 +3403,14 @@ }, { "epoch": 0.18972332015810275, - "grad_norm": 2.2063521775652393, - "learning_rate": 4.680945658197425e-06, - "loss": 0.2713, - "step": 64 + "grad_norm": 2.276698142449688, + "learning_rate": 4.678296760308474e-06, + "loss": 0.2569, + "step": 96 }, { "epoch": 0.18972332015810275, - "step": 64, - "train_accuracy_first_token_arguments": 0.9230769230769231, - "train_accuracy_first_token_arguments_total": 13, + "step": 96, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, @@ -3584,21 +3422,24 @@ }, { "epoch": 0.18972332015810275, - "step": 64, - "train_accuracy_first_token_arguments": 0.75, - "train_accuracy_first_token_arguments_total": 8, - "train_accuracy_recipient_token": 0.9411764705882353, + "step": 96, + "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { - "epoch": 0.18972332015810275, - "step": 64, - "train_accuracy_first_token_arguments": 0.95, - "train_accuracy_first_token_arguments_total": 20, + "epoch": 0.191699604743083, + "grad_norm": 2.2258003018865185, + "learning_rate": 4.670386560630446e-06, + "loss": 0.2672, + "step": 97 + }, + { + "epoch": 0.191699604743083, + "step": 97, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, @@ -3609,37 +3450,33 @@ "train_total_number_recipient_token": 19 }, { - "epoch": 0.19268774703557312, - "grad_norm": 2.1111555785819514, - "learning_rate": 4.669067660354456e-06, - "loss": 0.2806, - "step": 65 - }, - { - "epoch": 0.19268774703557312, - "step": 65, - "train_accuracy_first_token_arguments": 0.7272727272727273, - "train_accuracy_first_token_arguments_total": 11, - "train_accuracy_recipient_token": 0.8888888888888888, + "epoch": 0.191699604743083, + "step": 97, + "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, - "train_accuracy_recipient_token_gr": 0.8, + "train_accuracy_recipient_token_gr": 1.0, "train_accuracy_recipient_token_gr_total": 5, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.19268774703557312, - "step": 65, - "train_accuracy_first_token_arguments": 0.7692307692307693, - "train_accuracy_first_token_arguments_total": 13, - "train_accuracy_recipient_token": 0.8666666666666667, - "train_accuracy_recipient_token_all": 0.8, + "epoch": 0.19367588932806323, + "grad_norm": 2.1162036419771804, + "learning_rate": 4.6623871448617345e-06, + "loss": 0.2356, + "step": 98 + }, + { + "epoch": 0.19367588932806323, + "step": 98, + "train_accuracy_recipient_token": 0.9333333333333333, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, @@ -3650,10 +3487,8 @@ "train_total_number_recipient_token": 15 }, { - "epoch": 0.19268774703557312, - "step": 65, - "train_accuracy_first_token_arguments": 0.9230769230769231, - "train_accuracy_first_token_arguments_total": 13, + "epoch": 0.19367588932806323, + "step": 98, "train_accuracy_recipient_token": 0.7777777777777778, "train_accuracy_recipient_token_all": 0.6666666666666666, "train_accuracy_recipient_token_all_total": 12, @@ -3667,18 +3502,16 @@ }, { "epoch": 0.1956521739130435, - "grad_norm": 2.438873695871041, - "learning_rate": 4.656988227812658e-06, - "loss": 0.3234, - "step": 66 + "grad_norm": 2.752408100243012, + "learning_rate": 4.654298841826988e-06, + "loss": 0.276, + "step": 99 }, { "epoch": 0.1956521739130435, - "step": 66, - "train_accuracy_first_token_arguments": 0.9534883720930233, - "train_accuracy_first_token_arguments_total": 43, - "train_accuracy_recipient_token": 0.7058823529411765, - "train_accuracy_recipient_token_all": 0.6923076923076923, + "step": 99, + "train_accuracy_recipient_token": 0.7647058823529411, + "train_accuracy_recipient_token_all": 0.7692307692307693, "train_accuracy_recipient_token_all_total": 13, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, @@ -3690,11 +3523,9 @@ }, { "epoch": 0.1956521739130435, - "step": 66, - "train_accuracy_first_token_arguments": 0.55, - "train_accuracy_first_token_arguments_total": 20, - "train_accuracy_recipient_token": 0.7, - "train_accuracy_recipient_token_add": 0.0, + "step": 99, + "train_accuracy_recipient_token": 0.8, + "train_accuracy_recipient_token_add": 0.5, "train_accuracy_recipient_token_add_total": 2, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, @@ -3702,15 +3533,20 @@ "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_multiply": 0.0, - "train_accuracy_recipient_token_multiply_total": 2, + "train_accuracy_recipient_token_m": 0.5, + "train_accuracy_recipient_token_m_total": 2, "train_total_number_recipient_token": 20 }, { - "epoch": 0.1956521739130435, - "step": 66, - "train_accuracy_first_token_arguments": 0.9090909090909091, - "train_accuracy_first_token_arguments_total": 11, + "epoch": 0.1976284584980237, + "grad_norm": 2.6892659000959456, + "learning_rate": 4.646121984004666e-06, + "loss": 0.2694, + "step": 100 + }, + { + "epoch": 0.1976284584980237, + "step": 100, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, @@ -3723,17 +3559,8 @@ "train_total_number_recipient_token": 19 }, { - "epoch": 0.19861660079051383, - "grad_norm": 2.7417774764849683, - "learning_rate": 4.644708482352093e-06, - "loss": 0.2789, - "step": 67 - }, - { - "epoch": 0.19861660079051383, - "step": 67, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 12, + "epoch": 0.1976284584980237, + "step": 100, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, @@ -3746,12 +3573,17 @@ "train_total_number_recipient_token": 20 }, { - "epoch": 0.19861660079051383, - "step": 67, - "train_accuracy_first_token_arguments": 0.9047619047619048, - "train_accuracy_first_token_arguments_total": 21, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.19960474308300397, + "grad_norm": 1.8294943574863658, + "learning_rate": 4.637856907513366e-06, + "loss": 0.1149, + "step": 101 + }, + { + "epoch": 0.19960474308300397, + "step": 101, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, @@ -3764,17 +3596,15 @@ "train_total_number_recipient_token": 18 }, { - "epoch": 0.19861660079051383, - "step": 67, - "train_accuracy_first_token_arguments": 0.8846153846153846, - "train_accuracy_first_token_arguments_total": 26, - "train_accuracy_recipient_token": 0.8947368421052632, + "epoch": 0.19960474308300397, + "step": 101, + "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_build": 1.0, "train_accuracy_recipient_token_build_total": 1, - "train_accuracy_recipient_token_conduct": 0.0, - "train_accuracy_recipient_token_conduct_total": 1, + "train_accuracy_recipient_token_con": 0.0, + "train_accuracy_recipient_token_con_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_max": 1.0, @@ -3783,18 +3613,16 @@ }, { "epoch": 0.2015810276679842, - "grad_norm": 2.3471565473735976, - "learning_rate": 4.632229564355275e-06, - "loss": 0.1919, - "step": 68 + "grad_norm": 2.408233398782672, + "learning_rate": 4.629503952098011e-06, + "loss": 0.1773, + "step": 102 }, { "epoch": 0.2015810276679842, - "step": 68, - "train_accuracy_first_token_arguments": 0.8888888888888888, - "train_accuracy_first_token_arguments_total": 18, - "train_accuracy_recipient_token": 0.75, - "train_accuracy_recipient_token_all": 0.8, + "step": 102, + "train_accuracy_recipient_token": 0.6875, + "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_be": 0.0, "train_accuracy_recipient_token_be_total": 1, @@ -3802,16 +3630,14 @@ "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.2015810276679842, - "step": 68, - "train_accuracy_first_token_arguments": 0.6551724137931034, - "train_accuracy_first_token_arguments_total": 29, - "train_accuracy_recipient_token": 0.8947368421052632, + "step": 102, + "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, @@ -3820,109 +3646,104 @@ "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_identify": 0.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.2015810276679842, - "step": 68, - "train_accuracy_first_token_arguments": 0.6, - "train_accuracy_first_token_arguments_total": 25, - "train_accuracy_recipient_token": 0.8333333333333334, + "epoch": 0.20355731225296442, + "grad_norm": 3.07819121819936, + "learning_rate": 4.621063461115882e-06, + "loss": 0.2781, + "step": 103 + }, + { + "epoch": 0.20355731225296442, + "step": 103, + "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_choose": 0.0, + "train_accuracy_recipient_token_choose": 1.0, "train_accuracy_recipient_token_choose_total": 1, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_order": 0.0, "train_accuracy_recipient_token_order_total": 1, - "train_accuracy_recipient_token_plan": 0.0, + "train_accuracy_recipient_token_plan": 1.0, "train_accuracy_recipient_token_plan_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.20454545454545456, - "grad_norm": 2.845262992508057, - "learning_rate": 4.619552632701263e-06, - "loss": 0.2993, - "step": 69 - }, - { - "epoch": 0.20454545454545456, - "step": 69, - "train_accuracy_first_token_arguments": 0.9166666666666666, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.9473684210526315, + "epoch": 0.20355731225296442, + "step": 103, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_armor": 1.0, "train_accuracy_recipient_token_armor_total": 4, - "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { - "epoch": 0.20454545454545456, - "step": 69, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 36, - "train_accuracy_recipient_token": 0.8421052631578947, - "train_accuracy_recipient_token_all": 0.8461538461538461, + "epoch": 0.20553359683794467, + "grad_norm": 2.3567297901560926, + "learning_rate": 4.612535781522504e-06, + "loss": 0.2303, + "step": 104 + }, + { + "epoch": 0.20553359683794467, + "step": 104, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9230769230769231, "train_accuracy_recipient_token_all_total": 13, - "train_accuracy_recipient_token_customize": 1.0, - "train_accuracy_recipient_token_customize_total": 2, + "train_accuracy_recipient_token_custom": 1.0, + "train_accuracy_recipient_token_custom_total": 2, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 19 }, { - "epoch": 0.20454545454545456, - "step": 69, - "train_accuracy_first_token_arguments": 0.9166666666666666, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.8888888888888888, + "epoch": 0.20553359683794467, + "step": 104, + "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_get": 0.7142857142857143, + "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { "epoch": 0.2075098814229249, - "grad_norm": 2.2300979588264487, - "learning_rate": 4.606678864658039e-06, - "loss": 0.2752, - "step": 70 + "grad_norm": 2.007174401193821, + "learning_rate": 4.6039212638573835e-06, + "loss": 0.2304, + "step": 105 }, { "epoch": 0.2075098814229249, - "step": 70, - "train_accuracy_first_token_arguments": 0.6666666666666666, - "train_accuracy_first_token_arguments_total": 18, - "train_accuracy_recipient_token": 0.8947368421052632, - "train_accuracy_recipient_token_all": 0.9, + "step": 105, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 2, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 2, "train_accuracy_recipient_token_convert": 1.0, "train_accuracy_recipient_token_convert_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_render": 0.0, + "train_accuracy_recipient_token_render": 1.0, "train_accuracy_recipient_token_render_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.2075098814229249, - "step": 70, - "train_accuracy_first_token_arguments": 0.9827586206896551, - "train_accuracy_first_token_arguments_total": 58, - "train_accuracy_recipient_token": 0.8947368421052632, + "step": 105, + "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compare": 0.0, + "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, @@ -3933,51 +3754,52 @@ "train_total_number_recipient_token": 19 }, { - "epoch": 0.2075098814229249, - "step": 70, - "train_accuracy_first_token_arguments": 0.9047619047619048, - "train_accuracy_first_token_arguments_total": 21, - "train_accuracy_recipient_token": 0.8823529411764706, + "epoch": 0.20948616600790515, + "grad_norm": 1.8768688322152753, + "learning_rate": 4.5952202622296015e-06, + "loss": 0.1971, + "step": 106 + }, + { + "epoch": 0.20948616600790515, + "step": 106, + "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 17 }, { - "epoch": 0.21047430830039526, - "grad_norm": 2.1107565687083176, - "learning_rate": 4.5936094557731815e-06, - "loss": 0.228, - "step": 71 - }, - { - "epoch": 0.21047430830039526, - "step": 71, - "train_accuracy_first_token_arguments": 0.7307692307692307, - "train_accuracy_first_token_arguments_total": 26, - "train_accuracy_recipient_token": 0.7894736842105263, + "epoch": 0.20948616600790515, + "step": 106, + "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 0.42857142857142855, + "train_accuracy_recipient_token_get": 0.5714285714285714, "train_accuracy_recipient_token_get_total": 7, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 2, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 2, "train_total_number_recipient_token": 19 }, { - "epoch": 0.21047430830039526, - "step": 71, - "train_accuracy_first_token_arguments": 0.7777777777777778, - "train_accuracy_first_token_arguments_total": 18, - "train_accuracy_recipient_token": 0.8823529411764706, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.21146245059288538, + "grad_norm": 2.341750694388968, + "learning_rate": 4.586433134303257e-06, + "loss": 0.2459, + "step": 107 + }, + { + "epoch": 0.21146245059288538, + "step": 107, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, @@ -3988,113 +3810,106 @@ "train_total_number_recipient_token": 17 }, { - "epoch": 0.21047430830039526, - "step": 71, - "train_accuracy_first_token_arguments": 0.7333333333333333, - "train_accuracy_first_token_arguments_total": 15, - "train_accuracy_recipient_token": 0.9411764705882353, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.21146245059288538, + "step": 107, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 0.8571428571428571, + "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.2134387351778656, - "grad_norm": 2.1086876066685165, - "learning_rate": 4.5803456197628374e-06, - "loss": 0.2305, - "step": 72 + "grad_norm": 2.1944375595196965, + "learning_rate": 4.57756024128276e-06, + "loss": 0.199, + "step": 108 }, { "epoch": 0.2134387351778656, - "step": 72, - "train_accuracy_first_token_arguments": 0.8823529411764706, - "train_accuracy_first_token_arguments_total": 17, + "step": 108, "train_accuracy_recipient_token": 0.95, - "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_get": 0.8888888888888888, + "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 20 }, { "epoch": 0.2134387351778656, - "step": 72, - "train_accuracy_first_token_arguments": 0.5333333333333333, - "train_accuracy_first_token_arguments_total": 30, - "train_accuracy_recipient_token": 0.8421052631578947, + "step": 108, + "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_draw": 1.0, "train_accuracy_recipient_token_draw_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, - "train_accuracy_recipient_token_get": 0.6, + "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { - "epoch": 0.2134387351778656, - "step": 72, - "train_accuracy_first_token_arguments": 0.9, - "train_accuracy_first_token_arguments_total": 20, + "epoch": 0.21541501976284586, + "grad_norm": 2.5062076998625566, + "learning_rate": 4.5686019478979915e-06, + "loss": 0.3176, + "step": 109 + }, + { + "epoch": 0.21541501976284586, + "step": 109, "train_accuracy_recipient_token": 0.8, - "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_prior": 0.0, - "train_accuracy_recipient_token_prior_total": 1, + "train_accuracy_recipient_token_p": 1.0, + "train_accuracy_recipient_token_p_total": 1, "train_accuracy_recipient_token_report": 0.0, "train_accuracy_recipient_token_report_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, - "train_accuracy_recipient_token_store": 0.5, + "train_accuracy_recipient_token_store": 1.0, "train_accuracy_recipient_token_store_total": 2, "train_total_number_recipient_token": 15 }, { - "epoch": 0.21640316205533597, - "grad_norm": 2.406206005805186, - "learning_rate": 4.566888588399007e-06, - "loss": 0.3295, - "step": 73 - }, - { - "epoch": 0.21640316205533597, - "step": 73, - "train_accuracy_first_token_arguments": 0.8846153846153846, - "train_accuracy_first_token_arguments_total": 26, - "train_accuracy_recipient_token": 0.8947368421052632, + "epoch": 0.21541501976284586, + "step": 109, + "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 0.8888888888888888, + "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { - "epoch": 0.21640316205533597, - "step": 73, - "train_accuracy_first_token_arguments": 0.875, - "train_accuracy_first_token_arguments_total": 24, + "epoch": 0.21739130434782608, + "grad_norm": 1.746999082447299, + "learning_rate": 4.559558622389304e-06, + "loss": 0.1385, + "step": 110 + }, + { + "epoch": 0.21739130434782608, + "step": 110, "train_accuracy_recipient_token": 0.875, - "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_log": 0.5, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_log": 1.0, "train_accuracy_recipient_token_log_total": 2, "train_total_number_recipient_token": 16 }, { - "epoch": 0.21640316205533597, - "step": 73, - "train_accuracy_first_token_arguments": 0.9285714285714286, - "train_accuracy_first_token_arguments_total": 28, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, + "epoch": 0.21739130434782608, + "step": 110, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.7777777777777778, + "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, @@ -4103,20 +3918,18 @@ "train_accuracy_recipient_token_hire_total": 1, "train_accuracy_recipient_token_track": 1.0, "train_accuracy_recipient_token_track_total": 1, - "train_total_number_recipient_token": 18 + "train_total_number_recipient_token": 17 }, { "epoch": 0.21936758893280633, - "grad_norm": 2.1304599088968823, - "learning_rate": 4.553239611395156e-06, - "loss": 0.2566, - "step": 74 + "grad_norm": 2.7961010968564795, + "learning_rate": 4.55043063649239e-06, + "loss": 0.3181, + "step": 111 }, { "epoch": 0.21936758893280633, - "step": 74, - "train_accuracy_first_token_arguments": 0.8947368421052632, - "train_accuracy_first_token_arguments_total": 19, + "step": 111, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, @@ -4132,9 +3945,7 @@ }, { "epoch": 0.21936758893280633, - "step": 74, - "train_accuracy_first_token_arguments": 0.5517241379310345, - "train_accuracy_first_token_arguments_total": 29, + "step": 111, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, @@ -4142,22 +3953,27 @@ "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 2, - "train_accuracy_recipient_token_suggest": 1.0, - "train_accuracy_recipient_token_suggest_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 2, + "train_accuracy_recipient_token_s": 1.0, + "train_accuracy_recipient_token_s_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.21936758893280633, - "step": 74, - "train_accuracy_first_token_arguments": 0.75, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.7058823529411765, - "train_accuracy_recipient_token_all": 0.6363636363636364, + "epoch": 0.22134387351778656, + "grad_norm": 1.9989358452463166, + "learning_rate": 4.541218365422997e-06, + "loss": 0.2531, + "step": 112 + }, + { + "epoch": 0.22134387351778656, + "step": 112, + "train_accuracy_recipient_token": 0.7647058823529411, + "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_church": 1.0, - "train_accuracy_recipient_token_church_total": 1, + "train_accuracy_recipient_token_ch": 1.0, + "train_accuracy_recipient_token_ch_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.75, @@ -4165,75 +3981,67 @@ "train_total_number_recipient_token": 17 }, { - "epoch": 0.22233201581027667, - "grad_norm": 2.03869187550964, - "learning_rate": 4.539399956290152e-06, - "loss": 0.3017, - "step": 75 - }, - { - "epoch": 0.22233201581027667, - "step": 75, - "train_accuracy_first_token_arguments": 0.7391304347826086, - "train_accuracy_first_token_arguments_total": 46, - "train_accuracy_recipient_token": 0.8235294117647058, + "epoch": 0.22134387351778656, + "step": 112, + "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_f": 0.0, "train_accuracy_recipient_token_f_total": 1, "train_accuracy_recipient_token_plant": 1.0, "train_accuracy_recipient_token_plant_total": 1, - "train_accuracy_recipient_token_se": 0.0, + "train_accuracy_recipient_token_se": 1.0, "train_accuracy_recipient_token_se_total": 1, "train_accuracy_recipient_token_trim": 1.0, "train_accuracy_recipient_token_trim_total": 1, "train_total_number_recipient_token": 17 }, { - "epoch": 0.22233201581027667, - "step": 75, - "train_accuracy_first_token_arguments": 0.9285714285714286, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 0.7857142857142857, - "train_accuracy_recipient_token_all": 0.8, + "epoch": 0.22332015810276679, + "grad_norm": 2.571319112719595, + "learning_rate": 4.531922187861507e-06, + "loss": 0.2898, + "step": 113 + }, + { + "epoch": 0.22332015810276679, + "step": 113, + "train_accuracy_recipient_token": 0.9285714285714286, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_f": 1.0, "train_accuracy_recipient_token_f_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_restore": 0.0, + "train_accuracy_recipient_token_restore": 1.0, "train_accuracy_recipient_token_restore_total": 1, "train_total_number_recipient_token": 14 }, { - "epoch": 0.22233201581027667, - "step": 75, - "train_accuracy_first_token_arguments": 0.9411764705882353, - "train_accuracy_first_token_arguments_total": 17, + "epoch": 0.22332015810276679, + "step": 113, "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_gr": 0.6666666666666666, + "train_accuracy_recipient_token_gr": 1.0, "train_accuracy_recipient_token_gr_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.22529644268774704, - "grad_norm": 2.1715977940434494, - "learning_rate": 4.525370908330564e-06, - "loss": 0.2781, - "step": 76 + "grad_norm": 1.8363624831376564, + "learning_rate": 4.522542485937369e-06, + "loss": 0.2103, + "step": 114 }, { "epoch": 0.22529644268774704, - "step": 76, - "train_accuracy_first_token_arguments": 0.8421052631578947, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.7647058823529411, + "step": 114, + "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 0.0, @@ -4242,31 +4050,34 @@ "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 0.5, + "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.22529644268774704, - "step": 76, - "train_accuracy_first_token_arguments": 0.71875, - "train_accuracy_first_token_arguments_total": 64, - "train_accuracy_recipient_token": 0.8421052631578947, - "train_accuracy_recipient_token_all": 0.7272727272727273, - "train_accuracy_recipient_token_all_total": 11, + "step": 114, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, - "train_total_number_recipient_token": 19 + "train_total_number_recipient_token": 18 }, { - "epoch": 0.22529644268774704, - "step": 76, - "train_accuracy_first_token_arguments": 0.9523809523809523, - "train_accuracy_first_token_arguments_total": 21, + "epoch": 0.22727272727272727, + "grad_norm": 2.6890155172544046, + "learning_rate": 4.513079645213391e-06, + "loss": 0.3333, + "step": 115 + }, + { + "epoch": 0.22727272727272727, + "step": 115, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, @@ -4277,33 +4088,29 @@ "train_total_number_recipient_token": 19 }, { - "epoch": 0.22826086956521738, - "grad_norm": 2.6029973414283223, - "learning_rate": 4.511153770351288e-06, - "loss": 0.3502, - "step": 77 - }, - { - "epoch": 0.22826086956521738, - "step": 77, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 15, - "train_accuracy_recipient_token": 0.9473684210526315, + "epoch": 0.22727272727272727, + "step": 115, + "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { - "epoch": 0.22826086956521738, - "step": 77, - "train_accuracy_first_token_arguments": 0.6153846153846154, - "train_accuracy_first_token_arguments_total": 13, + "epoch": 0.22924901185770752, + "grad_norm": 3.4159406545291486, + "learning_rate": 4.5035340546698915e-06, + "loss": 0.2674, + "step": 116 + }, + { + "epoch": 0.22924901185770752, + "step": 116, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, @@ -4311,36 +4118,32 @@ "train_accuracy_recipient_token_flower_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 20 }, { - "epoch": 0.22826086956521738, - "step": 77, - "train_accuracy_first_token_arguments": 0.7049180327868853, - "train_accuracy_first_token_arguments_total": 61, + "epoch": 0.22924901185770752, + "step": 116, "train_accuracy_recipient_token": 0.9333333333333333, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 15 }, { "epoch": 0.23122529644268774, - "grad_norm": 2.67404592050148, - "learning_rate": 4.496749862654574e-06, - "loss": 0.3309, - "step": 78 + "grad_norm": 2.099346426532093, + "learning_rate": 4.493906106688712e-06, + "loss": 0.3201, + "step": 117 }, { "epoch": 0.23122529644268774, - "step": 78, - "train_accuracy_first_token_arguments": 0.9, - "train_accuracy_first_token_arguments_total": 20, + "step": 117, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_add": 1.0, "train_accuracy_recipient_token_add_total": 1, @@ -4354,22 +4157,25 @@ }, { "epoch": 0.23122529644268774, - "step": 78, - "train_accuracy_first_token_arguments": 0.95, - "train_accuracy_first_token_arguments_total": 20, - "train_accuracy_recipient_token": 0.95, + "step": 117, + "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.8888888888888888, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 20 }, { - "epoch": 0.23122529644268774, - "step": 78, - "train_accuracy_first_token_arguments": 0.84, - "train_accuracy_first_token_arguments_total": 25, - "train_accuracy_recipient_token": 0.9523809523809523, + "epoch": 0.233201581027668, + "grad_norm": 1.9975724683282075, + "learning_rate": 4.484196197037082e-06, + "loss": 0.2154, + "step": 118 + }, + { + "epoch": 0.233201581027668, + "step": 118, + "train_accuracy_recipient_token": 0.9047619047619048, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_extract": 1.0, @@ -4383,31 +4189,27 @@ "train_total_number_recipient_token": 21 }, { - "epoch": 0.2341897233201581, - "grad_norm": 1.8475653464555015, - "learning_rate": 4.482160522887404e-06, - "loss": 0.228, - "step": 79 - }, - { - "epoch": 0.2341897233201581, - "step": 79, - "train_accuracy_first_token_arguments": 0.6153846153846154, - "train_accuracy_first_token_arguments_total": 13, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.233201581027668, + "step": 118, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { - "epoch": 0.2341897233201581, - "step": 79, - "train_accuracy_first_token_arguments": 0.75, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.8947368421052632, - "train_accuracy_recipient_token_all": 0.8, + "epoch": 0.23517786561264822, + "grad_norm": 1.8241360509276188, + "learning_rate": 4.474404724851356e-06, + "loss": 0.2666, + "step": 119 + }, + { + "epoch": 0.23517786561264822, + "step": 119, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_arc": 1.0, "train_accuracy_recipient_token_arc_total": 4, @@ -4418,10 +4220,8 @@ "train_total_number_recipient_token": 19 }, { - "epoch": 0.2341897233201581, - "step": 79, - "train_accuracy_first_token_arguments": 0.4444444444444444, - "train_accuracy_first_token_arguments_total": 45, + "epoch": 0.23517786561264822, + "step": 119, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, @@ -4435,16 +4235,14 @@ }, { "epoch": 0.23715415019762845, - "grad_norm": 2.189256435693527, - "learning_rate": 4.467387105917269e-06, - "loss": 0.3682, - "step": 80 + "grad_norm": 2.374987760349219, + "learning_rate": 4.464532092620607e-06, + "loss": 0.3071, + "step": 120 }, { "epoch": 0.23715415019762845, - "step": 80, - "train_accuracy_first_token_arguments": 0.8275862068965517, - "train_accuracy_first_token_arguments_total": 29, + "step": 120, "train_accuracy_recipient_token": 0.8125, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, @@ -4460,11 +4258,9 @@ }, { "epoch": 0.23715415019762845, - "step": 80, - "train_accuracy_first_token_arguments": 0.7333333333333333, - "train_accuracy_first_token_arguments_total": 15, - "train_accuracy_recipient_token": 0.8235294117647058, - "train_accuracy_recipient_token_all": 0.8, + "step": 120, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, @@ -4472,17 +4268,22 @@ "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 1, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, "train_total_number_recipient_token": 17 }, { - "epoch": 0.23715415019762845, - "step": 80, - "train_accuracy_first_token_arguments": 0.8571428571428571, - "train_accuracy_first_token_arguments_total": 21, - "train_accuracy_recipient_token": 0.8947368421052632, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.2391304347826087, + "grad_norm": 1.5783968329951958, + "learning_rate": 4.454578706170075e-06, + "loss": 0.1226, + "step": 121 + }, + { + "epoch": 0.2391304347826087, + "step": 121, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, @@ -4490,56 +4291,50 @@ "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.24011857707509882, - "grad_norm": 1.6646080394168112, - "learning_rate": 4.452430983706351e-06, - "loss": 0.1579, - "step": 81 - }, - { - "epoch": 0.24011857707509882, - "step": 81, - "train_accuracy_first_token_arguments": 0.875, - "train_accuracy_first_token_arguments_total": 8, + "epoch": 0.2391304347826087, + "step": 121, "train_accuracy_recipient_token": 0.8823529411764706, - "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 2, - "train_accuracy_recipient_token_explain": 1.0, - "train_accuracy_recipient_token_explain_total": 1, - "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_ex": 1.0, + "train_accuracy_recipient_token_ex_total": 1, + "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_sc": 1.0, "train_accuracy_recipient_token_sc_total": 2, "train_total_number_recipient_token": 17 }, { - "epoch": 0.24011857707509882, - "step": 81, - "train_accuracy_first_token_arguments": 0.8, - "train_accuracy_first_token_arguments_total": 15, - "train_accuracy_recipient_token": 0.8125, - "train_accuracy_recipient_token_all": 0.7, + "epoch": 0.24110671936758893, + "grad_norm": 1.848245708230007, + "learning_rate": 4.444544974644493e-06, + "loss": 0.2494, + "step": 122 + }, + { + "epoch": 0.24110671936758893, + "step": 122, + "train_accuracy_recipient_token": 0.9375, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 1, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, "train_total_number_recipient_token": 16 }, { - "epoch": 0.24011857707509882, - "step": 81, - "train_accuracy_first_token_arguments": 0.75, - "train_accuracy_first_token_arguments_total": 16, + "epoch": 0.24110671936758893, + "step": 122, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, @@ -4549,22 +4344,20 @@ "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.24308300395256918, - "grad_norm": 1.9011227117633136, - "learning_rate": 4.437293545184111e-06, - "loss": 0.2544, - "step": 82 + "grad_norm": 1.747036665792108, + "learning_rate": 4.434431310491267e-06, + "loss": 0.1616, + "step": 123 }, { "epoch": 0.24308300395256918, - "step": 82, - "train_accuracy_first_token_arguments": 0.9090909090909091, - "train_accuracy_first_token_arguments_total": 11, + "step": 123, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, @@ -4580,9 +4373,7 @@ }, { "epoch": 0.24308300395256918, - "step": 82, - "train_accuracy_first_token_arguments": 0.631578947368421, - "train_accuracy_first_token_arguments_total": 38, + "step": 123, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, @@ -4590,17 +4381,22 @@ "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, - "train_accuracy_recipient_token_inspect": 1.0, - "train_accuracy_recipient_token_inspect_total": 1, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 1, + "train_accuracy_recipient_token_ins": 1.0, + "train_accuracy_recipient_token_ins_total": 1, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.24308300395256918, - "step": 82, - "train_accuracy_first_token_arguments": 0.8846153846153846, - "train_accuracy_first_token_arguments_total": 26, + "epoch": 0.2450592885375494, + "grad_norm": 2.235367946517307, + "learning_rate": 4.424238129443515e-06, + "loss": 0.2711, + "step": 124 + }, + { + "epoch": 0.2450592885375494, + "step": 124, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 9, @@ -4611,17 +4407,8 @@ "train_total_number_recipient_token": 19 }, { - "epoch": 0.24604743083003952, - "grad_norm": 1.9670272993707267, - "learning_rate": 4.421976196118297e-06, - "loss": 0.2366, - "step": 83 - }, - { - "epoch": 0.24604743083003952, - "step": 83, - "train_accuracy_first_token_arguments": 0.9090909090909091, - "train_accuracy_first_token_arguments_total": 22, + "epoch": 0.2450592885375494, + "step": 124, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, @@ -4630,32 +4417,35 @@ "train_total_number_recipient_token": 18 }, { - "epoch": 0.24604743083003952, - "step": 83, - "train_accuracy_first_token_arguments": 0.7368421052631579, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.7777777777777778, - "train_accuracy_recipient_token_all": 0.8, + "epoch": 0.24703557312252963, + "grad_norm": 1.8124306151814813, + "learning_rate": 4.413965850502987e-06, + "loss": 0.1279, + "step": 125 + }, + { + "epoch": 0.24703557312252963, + "step": 125, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compare": 0.5, + "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_get": 0.5, + "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 2, "train_total_number_recipient_token": 18 }, { - "epoch": 0.24604743083003952, - "step": 83, - "train_accuracy_first_token_arguments": 0.7, - "train_accuracy_first_token_arguments_total": 10, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.24703557312252963, + "step": 125, + "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compile": 1.0, + "train_accuracy_recipient_token_compile": 0.5, "train_accuracy_recipient_token_compile_total": 2, "train_accuracy_recipient_token_compress": 1.0, "train_accuracy_recipient_token_compress_total": 1, @@ -4667,16 +4457,14 @@ }, { "epoch": 0.2490118577075099, - "grad_norm": 2.214361868557108, - "learning_rate": 4.40648035898441e-06, - "loss": 0.3157, - "step": 84 + "grad_norm": 2.3030811699911142, + "learning_rate": 4.4036148959228365e-06, + "loss": 0.3385, + "step": 126 }, { "epoch": 0.2490118577075099, - "step": 84, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 15, + "step": 126, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, @@ -4686,84 +4474,81 @@ }, { "epoch": 0.2490118577075099, - "step": 84, - "train_accuracy_first_token_arguments": 0.9285714285714286, - "train_accuracy_first_token_arguments_total": 28, - "train_accuracy_recipient_token": 0.9473684210526315, + "step": 126, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, - "train_accuracy_recipient_token_get": 0.6666666666666666, - "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_research": 1.0, "train_accuracy_recipient_token_research_total": 2, - "train_total_number_recipient_token": 19 + "train_total_number_recipient_token": 18 }, { - "epoch": 0.2490118577075099, - "step": 84, - "train_accuracy_first_token_arguments": 0.8888888888888888, - "train_accuracy_first_token_arguments_total": 9, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.2509881422924901, + "grad_norm": 1.7077330162259072, + "learning_rate": 4.3931856911902635e-06, + "loss": 0.1785, + "step": 127 + }, + { + "epoch": 0.2509881422924901, + "step": 127, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compose": 1.0, "train_accuracy_recipient_token_compose_total": 4, - "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.2519762845849802, - "grad_norm": 1.79425216853278, - "learning_rate": 4.390807472833585e-06, - "loss": 0.2629, - "step": 85 - }, - { - "epoch": 0.2519762845849802, - "step": 85, - "train_accuracy_first_token_arguments": 0.9375, - "train_accuracy_first_token_arguments_total": 16, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9090909090909091, + "epoch": 0.2509881422924901, + "step": 127, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_write": 1.0, "train_accuracy_recipient_token_write_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.2519762845849802, - "step": 85, - "train_accuracy_first_token_arguments": 0.8461538461538461, - "train_accuracy_first_token_arguments_total": 13, - "train_accuracy_recipient_token": 0.8421052631578947, + "epoch": 0.25296442687747034, + "grad_norm": 2.2097826093919095, + "learning_rate": 4.382678665009028e-06, + "loss": 0.2991, + "step": 128 + }, + { + "epoch": 0.25296442687747034, + "step": 128, + "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 0.7777777777777778, + "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { - "epoch": 0.2519762845849802, - "step": 85, - "train_accuracy_first_token_arguments": 0.7272727272727273, - "train_accuracy_first_token_arguments_total": 11, - "train_accuracy_recipient_token": 0.8888888888888888, + "epoch": 0.25296442687747034, + "step": 128, + "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_explain": 1.0, - "train_accuracy_recipient_token_explain_total": 1, + "train_accuracy_recipient_token_ex": 1.0, + "train_accuracy_recipient_token_ex_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 1, @@ -4771,48 +4556,49 @@ }, { "epoch": 0.2549407114624506, - "grad_norm": 2.256598889106997, - "learning_rate": 4.374958993158965e-06, - "loss": 0.2383, - "step": 86 + "grad_norm": 2.2527852681173846, + "learning_rate": 4.37209424928182e-06, + "loss": 0.1948, + "step": 129 }, { "epoch": 0.2549407114624506, - "step": 86, - "train_accuracy_first_token_arguments": 0.7368421052631579, - "train_accuracy_first_token_arguments_total": 38, - "train_accuracy_recipient_token": 1.0, + "step": 129, + "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_flag": 1.0, + "train_accuracy_recipient_token_flag": 0.75, "train_accuracy_recipient_token_flag_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.2549407114624506, - "step": 86, - "train_accuracy_first_token_arguments": 0.8695652173913043, - "train_accuracy_first_token_arguments_total": 23, - "train_accuracy_recipient_token": 0.7894736842105263, - "train_accuracy_recipient_token_all": 0.8461538461538461, + "step": 129, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9230769230769231, "train_accuracy_recipient_token_all_total": 13, "train_accuracy_recipient_token_compare": 0.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_describe": 1.0, "train_accuracy_recipient_token_describe_total": 4, - "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.2549407114624506, - "step": 86, - "train_accuracy_first_token_arguments": 0.8378378378378378, - "train_accuracy_first_token_arguments_total": 37, + "epoch": 0.25691699604743085, + "grad_norm": 2.248286132307088, + "learning_rate": 4.361432879092518e-06, + "loss": 0.345, + "step": 130 + }, + { + "epoch": 0.25691699604743085, + "step": 130, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, @@ -4820,40 +4606,36 @@ "train_accuracy_recipient_token_ass_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_plan": 1.0, "train_accuracy_recipient_token_plan_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.25790513833992096, - "grad_norm": 2.49197628899743, - "learning_rate": 4.358936391760524e-06, - "loss": 0.4685, - "step": 87 - }, - { - "epoch": 0.25790513833992096, - "step": 87, - "train_accuracy_first_token_arguments": 0.9285714285714286, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.25691699604743085, + "step": 130, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.875, "train_accuracy_recipient_token_get_total": 8, - "train_accuracy_recipient_token_identify": 0.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.25790513833992096, - "step": 87, - "train_accuracy_first_token_arguments": 0.9583333333333334, - "train_accuracy_first_token_arguments_total": 24, - "train_accuracy_recipient_token": 0.9411764705882353, - "train_accuracy_recipient_token_all": 0.9166666666666666, + "epoch": 0.25889328063241107, + "grad_norm": 2.7851285907266368, + "learning_rate": 4.350694992688289e-06, + "loss": 0.4034, + "step": 131 + }, + { + "epoch": 0.25889328063241107, + "step": 131, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.75, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, @@ -4864,12 +4646,10 @@ "train_total_number_recipient_token": 17 }, { - "epoch": 0.25790513833992096, - "step": 87, - "train_accuracy_first_token_arguments": 0.9047619047619048, - "train_accuracy_first_token_arguments_total": 21, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.25889328063241107, + "step": 131, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, @@ -4879,18 +4659,16 @@ }, { "epoch": 0.2608695652173913, - "grad_norm": 2.280644257269206, - "learning_rate": 4.342741156608392e-06, - "loss": 0.2713, - "step": 88 + "grad_norm": 2.0403309814666697, + "learning_rate": 4.339881031461588e-06, + "loss": 0.2414, + "step": 132 }, { "epoch": 0.2608695652173913, - "step": 88, - "train_accuracy_first_token_arguments": 0.9444444444444444, - "train_accuracy_first_token_arguments_total": 18, - "train_accuracy_recipient_token": 0.875, - "train_accuracy_recipient_token_all": 0.9, + "step": 132, + "train_accuracy_recipient_token": 0.8125, + "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, @@ -4898,10 +4676,8 @@ }, { "epoch": 0.2608695652173913, - "step": 88, - "train_accuracy_first_token_arguments": 0.7407407407407407, - "train_accuracy_first_token_arguments_total": 27, - "train_accuracy_recipient_token": 0.9444444444444444, + "step": 132, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, @@ -4910,15 +4686,20 @@ "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_track": 1.0, "train_accuracy_recipient_token_track_total": 1, - "train_accuracy_recipient_token_write": 0.0, + "train_accuracy_recipient_token_write": 1.0, "train_accuracy_recipient_token_write_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.2608695652173913, - "step": 88, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 14, + "epoch": 0.2628458498023715, + "grad_norm": 2.697975383814613, + "learning_rate": 4.328991439932003e-06, + "loss": 0.313, + "step": 133 + }, + { + "epoch": 0.2628458498023715, + "step": 133, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, @@ -4933,35 +4714,31 @@ "train_total_number_recipient_token": 19 }, { - "epoch": 0.26383399209486164, - "grad_norm": 2.337688603854322, - "learning_rate": 4.32637479170467e-06, - "loss": 0.3134, - "step": 89 - }, - { - "epoch": 0.26383399209486164, - "step": 89, - "train_accuracy_first_token_arguments": 0.6666666666666666, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.9444444444444444, + "epoch": 0.2628458498023715, + "step": 133, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_explain": 0.0, - "train_accuracy_recipient_token_explain_total": 1, + "train_accuracy_recipient_token_ex": 1.0, + "train_accuracy_recipient_token_ex_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_track": 1.0, "train_accuracy_recipient_token_track_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.26383399209486164, - "step": 89, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 20, + "epoch": 0.2648221343873518, + "grad_norm": 2.5542065987019664, + "learning_rate": 4.318026665727993e-06, + "loss": 0.2338, + "step": 134 + }, + { + "epoch": 0.2648221343873518, + "step": 134, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, @@ -4976,14 +4753,12 @@ "train_total_number_recipient_token": 20 }, { - "epoch": 0.26383399209486164, - "step": 89, - "train_accuracy_first_token_arguments": 0.9047619047619048, - "train_accuracy_first_token_arguments_total": 21, - "train_accuracy_recipient_token": 0.8125, + "epoch": 0.2648221343873518, + "step": 134, + "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_add": 1.0, "train_accuracy_recipient_token_add_total": 1, - "train_accuracy_recipient_token_all": 0.7692307692307693, + "train_accuracy_recipient_token_all": 0.8461538461538461, "train_accuracy_recipient_token_all_total": 13, "train_accuracy_recipient_token_apply": 1.0, "train_accuracy_recipient_token_apply_total": 1, @@ -4993,16 +4768,14 @@ }, { "epoch": 0.26679841897233203, - "grad_norm": 2.596535079636875, - "learning_rate": 4.309838816943755e-06, - "loss": 0.2432, - "step": 90 + "grad_norm": 2.4386171945089004, + "learning_rate": 4.3069871595684795e-06, + "loss": 0.1782, + "step": 135 }, { "epoch": 0.26679841897233203, - "step": 90, - "train_accuracy_first_token_arguments": 0.76, - "train_accuracy_first_token_arguments_total": 25, + "step": 135, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, @@ -5014,9 +4787,7 @@ }, { "epoch": 0.26679841897233203, - "step": 90, - "train_accuracy_first_token_arguments": 0.9333333333333333, - "train_accuracy_first_token_arguments_total": 15, + "step": 135, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, @@ -5029,10 +4800,15 @@ "train_total_number_recipient_token": 18 }, { - "epoch": 0.26679841897233203, - "step": 90, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 17, + "epoch": 0.26877470355731226, + "grad_norm": 1.9475949445800667, + "learning_rate": 4.295873375244319e-06, + "loss": 0.2018, + "step": 136 + }, + { + "epoch": 0.26877470355731226, + "step": 136, "train_accuracy_recipient_token": 0.75, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, @@ -5043,19 +4819,10 @@ "train_total_number_recipient_token": 16 }, { - "epoch": 0.26976284584980237, - "grad_norm": 2.2037749517640015, - "learning_rate": 4.293134767971193e-06, - "loss": 0.2113, - "step": 91 - }, - { - "epoch": 0.26976284584980237, - "step": 91, - "train_accuracy_first_token_arguments": 0.7, - "train_accuracy_first_token_arguments_total": 10, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.26877470355731226, + "step": 136, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_ele": 1.0, "train_accuracy_recipient_token_ele_total": 2, @@ -5064,55 +4831,56 @@ "train_total_number_recipient_token": 17 }, { - "epoch": 0.26976284584980237, - "step": 91, - "train_accuracy_first_token_arguments": 0.6363636363636364, - "train_accuracy_first_token_arguments_total": 11, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, + "epoch": 0.2707509881422925, + "grad_norm": 1.932745049719722, + "learning_rate": 4.284685769599658e-06, + "loss": 0.3062, + "step": 137 + }, + { + "epoch": 0.2707509881422925, + "step": 137, + "train_accuracy_recipient_token": 0.9375, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 1, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 16 }, { - "epoch": 0.26976284584980237, - "step": 91, - "train_accuracy_first_token_arguments": 0.8103448275862069, - "train_accuracy_first_token_arguments_total": 58, + "epoch": 0.2707509881422925, + "step": 137, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 2, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 2, "train_accuracy_recipient_token_object": 1.0, "train_accuracy_recipient_token_object_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.2727272727272727, - "grad_norm": 2.1267599638494628, - "learning_rate": 4.276264196041074e-06, - "loss": 0.3654, - "step": 92 + "grad_norm": 2.5725457529065006, + "learning_rate": 4.273424802513145e-06, + "loss": 0.2588, + "step": 138 }, { "epoch": 0.2727272727272727, - "step": 92, - "train_accuracy_first_token_arguments": 0.8292682926829268, - "train_accuracy_first_token_arguments_total": 41, - "train_accuracy_recipient_token": 0.8823529411764706, + "step": 138, + "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_deploy": 1.0, "train_accuracy_recipient_token_deploy_total": 1, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_plan": 1.0, "train_accuracy_recipient_token_plan_total": 1, @@ -5120,9 +4888,7 @@ }, { "epoch": 0.2727272727272727, - "step": 92, - "train_accuracy_first_token_arguments": 0.7105263157894737, - "train_accuracy_first_token_arguments_total": 38, + "step": 138, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, @@ -5137,12 +4903,17 @@ "train_total_number_recipient_token": 19 }, { - "epoch": 0.2727272727272727, - "step": 92, - "train_accuracy_first_token_arguments": 0.8421052631578947, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.8235294117647058, - "train_accuracy_recipient_token_all": 0.8, + "epoch": 0.274703557312253, + "grad_norm": 1.86126996030237, + "learning_rate": 4.26209093687903e-06, + "loss": 0.205, + "step": 139 + }, + { + "epoch": 0.274703557312253, + "step": 139, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_detect": 0.0, "train_accuracy_recipient_token_detect_total": 1, @@ -5153,37 +4924,33 @@ "train_total_number_recipient_token": 17 }, { - "epoch": 0.2756916996047431, - "grad_norm": 2.2199589796264414, - "learning_rate": 4.259228667871963e-06, - "loss": 0.3508, - "step": 93 - }, - { - "epoch": 0.2756916996047431, - "step": 93, - "train_accuracy_first_token_arguments": 0.9444444444444444, - "train_accuracy_first_token_arguments_total": 18, + "epoch": 0.274703557312253, + "step": 139, "train_accuracy_recipient_token": 0.7894736842105263, "train_accuracy_recipient_token_add": 1.0, "train_accuracy_recipient_token_add_total": 1, - "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_calculate": 0.5, + "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 2, - "train_accuracy_recipient_token_decode": 1.0, + "train_accuracy_recipient_token_decode": 0.75, "train_accuracy_recipient_token_decode_total": 4, - "train_accuracy_recipient_token_subtract": 0.0, + "train_accuracy_recipient_token_subtract": 1.0, "train_accuracy_recipient_token_subtract_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.2756916996047431, - "step": 93, - "train_accuracy_first_token_arguments": 0.7407407407407407, - "train_accuracy_first_token_arguments_total": 27, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.2766798418972332, + "grad_norm": 3.716815765106261, + "learning_rate": 4.2506846385881375e-06, + "loss": 0.4463, + "step": 140 + }, + { + "epoch": 0.2766798418972332, + "step": 140, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, @@ -5192,10 +4959,8 @@ "train_total_number_recipient_token": 19 }, { - "epoch": 0.2756916996047431, - "step": 93, - "train_accuracy_first_token_arguments": 0.875, - "train_accuracy_first_token_arguments_total": 24, + "epoch": 0.2766798418972332, + "step": 140, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, @@ -5209,18 +4974,16 @@ }, { "epoch": 0.27865612648221344, - "grad_norm": 2.6263899594688755, - "learning_rate": 4.242029765501411e-06, - "loss": 0.2873, - "step": 94 + "grad_norm": 1.8259425268376976, + "learning_rate": 4.239206376508716e-06, + "loss": 0.2031, + "step": 141 }, { "epoch": 0.27865612648221344, - "step": 94, - "train_accuracy_first_token_arguments": 0.9473684210526315, - "train_accuracy_first_token_arguments_total": 76, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.9, + "step": 141, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_d": 0.5, "train_accuracy_recipient_token_d_total": 2, @@ -5232,9 +4995,7 @@ }, { "epoch": 0.27865612648221344, - "step": 94, - "train_accuracy_first_token_arguments": 0.64, - "train_accuracy_first_token_arguments_total": 25, + "step": 141, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, @@ -5249,12 +5010,17 @@ "train_total_number_recipient_token": 18 }, { - "epoch": 0.27865612648221344, - "step": 94, - "train_accuracy_first_token_arguments": 0.8125, - "train_accuracy_first_token_arguments_total": 16, - "train_accuracy_recipient_token": 0.8823529411764706, - "train_accuracy_recipient_token_all": 0.8, + "epoch": 0.28063241106719367, + "grad_norm": 1.9633184170686586, + "learning_rate": 4.227656622467162e-06, + "loss": 0.2176, + "step": 142 + }, + { + "epoch": 0.28063241106719367, + "step": 142, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, @@ -5267,19 +5033,10 @@ "train_total_number_recipient_token": 17 }, { - "epoch": 0.2816205533596838, - "grad_norm": 1.8705277085481686, - "learning_rate": 4.22466908613903e-06, - "loss": 0.26, - "step": 95 - }, - { - "epoch": 0.2816205533596838, - "step": 95, - "train_accuracy_first_token_arguments": 0.8076923076923077, - "train_accuracy_first_token_arguments_total": 26, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9090909090909091, + "epoch": 0.28063241106719367, + "step": 142, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_arr": 1.0, "train_accuracy_recipient_token_arr_total": 1, @@ -5287,15 +5044,20 @@ "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, - "train_accuracy_recipient_token_suggest": 1.0, - "train_accuracy_recipient_token_suggest_total": 1, + "train_accuracy_recipient_token_s": 1.0, + "train_accuracy_recipient_token_s_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.2816205533596838, - "step": 95, - "train_accuracy_first_token_arguments": 0.5510204081632653, - "train_accuracy_first_token_arguments_total": 49, + "epoch": 0.2826086956521739, + "grad_norm": 2.1587360641406623, + "learning_rate": 4.2160358512286266e-06, + "loss": 0.3984, + "step": 143 + }, + { + "epoch": 0.2826086956521739, + "step": 143, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, @@ -5310,15 +5072,13 @@ "train_total_number_recipient_token": 18 }, { - "epoch": 0.2816205533596838, - "step": 95, - "train_accuracy_first_token_arguments": 0.8947368421052632, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.8235294117647058, - "train_accuracy_recipient_token_all": 0.7272727272727273, + "epoch": 0.2826086956521739, + "step": 143, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_sh": 1.0, @@ -5327,30 +5087,26 @@ }, { "epoch": 0.2845849802371542, - "grad_norm": 2.0123868841853065, - "learning_rate": 4.20714824201817e-06, - "loss": 0.3846, - "step": 96 + "grad_norm": 2.0510066673643084, + "learning_rate": 4.204344540477499e-06, + "loss": 0.2632, + "step": 144 }, { "epoch": 0.2845849802371542, - "step": 96, - "train_accuracy_first_token_arguments": 0.8928571428571429, - "train_accuracy_first_token_arguments_total": 28, - "train_accuracy_recipient_token": 0.8333333333333334, - "train_accuracy_recipient_token_all": 0.7272727272727273, + "step": 144, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 18 }, { "epoch": 0.2845849802371542, - "step": 96, - "train_accuracy_first_token_arguments": 0.88, - "train_accuracy_first_token_arguments_total": 25, + "step": 144, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, @@ -5365,49 +5121,50 @@ "train_total_number_recipient_token": 20 }, { - "epoch": 0.2845849802371542, - "step": 96, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 26, - "train_accuracy_recipient_token": 0.8947368421052632, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.2865612648221344, + "grad_norm": 2.0576342231497513, + "learning_rate": 4.192583170797775e-06, + "loss": 0.163, + "step": 145 + }, + { + "epoch": 0.2865612648221344, + "step": 145, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 0.8888888888888888, + "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { - "epoch": 0.2875494071146245, - "grad_norm": 1.6695567253827632, - "learning_rate": 4.189468860246192e-06, - "loss": 0.1603, - "step": 97 - }, - { - "epoch": 0.2875494071146245, - "step": 97, - "train_accuracy_first_token_arguments": 0.75, - "train_accuracy_first_token_arguments_total": 36, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.2865612648221344, + "step": 145, + "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_enh": 1.0, "train_accuracy_recipient_token_enh_total": 1, - "train_accuracy_recipient_token_explain": 1.0, - "train_accuracy_recipient_token_explain_total": 1, - "train_accuracy_recipient_token_public": 1.0, + "train_accuracy_recipient_token_ex": 1.0, + "train_accuracy_recipient_token_ex_total": 1, + "train_accuracy_recipient_token_public": 0.0, "train_accuracy_recipient_token_public_total": 1, "train_accuracy_recipient_token_street": 1.0, "train_accuracy_recipient_token_street_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.2875494071146245, - "step": 97, - "train_accuracy_first_token_arguments": 0.9090909090909091, - "train_accuracy_first_token_arguments_total": 44, - "train_accuracy_recipient_token": 0.8235294117647058, - "train_accuracy_recipient_token_all": 0.75, + "epoch": 0.2885375494071146, + "grad_norm": 1.7728828833133476, + "learning_rate": 4.1807522256532925e-06, + "loss": 0.2257, + "step": 146 + }, + { + "epoch": 0.2885375494071146, + "step": 146, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8333333333333334, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, @@ -5416,38 +5173,34 @@ "train_total_number_recipient_token": 17 }, { - "epoch": 0.2875494071146245, - "step": 97, - "train_accuracy_first_token_arguments": 0.8947368421052632, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.9411764705882353, + "epoch": 0.2885375494071146, + "step": 146, + "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 17 }, { "epoch": 0.29051383399209485, - "grad_norm": 2.3366857423927834, - "learning_rate": 4.171632582653368e-06, - "loss": 0.3712, - "step": 98 + "grad_norm": 2.6572528287169903, + "learning_rate": 4.1688521913678706e-06, + "loss": 0.3191, + "step": 147 }, { "epoch": 0.29051383399209485, - "step": 98, - "train_accuracy_first_token_arguments": 0.8636363636363636, - "train_accuracy_first_token_arguments_total": 22, + "step": 147, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 0.0, - "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_analy": 0.0, + "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_c": 0.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_compare": 1.0, @@ -5458,14 +5211,12 @@ }, { "epoch": 0.29051383399209485, - "step": 98, - "train_accuracy_first_token_arguments": 0.8260869565217391, - "train_accuracy_first_token_arguments_total": 23, - "train_accuracy_recipient_token": 0.8125, - "train_accuracy_recipient_token_all": 0.8, + "step": 147, + "train_accuracy_recipient_token": 0.875, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 0.0, - "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_analy": 0.0, + "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_assign": 1.0, "train_accuracy_recipient_token_assign_total": 1, "train_accuracy_recipient_token_configure": 1.0, @@ -5475,46 +5226,47 @@ "train_total_number_recipient_token": 16 }, { - "epoch": 0.29051383399209485, - "step": 98, - "train_accuracy_first_token_arguments": 0.6153846153846154, - "train_accuracy_first_token_arguments_total": 52, + "epoch": 0.2924901185770751, + "grad_norm": 2.487728724207803, + "learning_rate": 4.156883557105308e-06, + "loss": 0.2116, + "step": 148 + }, + { + "epoch": 0.2924901185770751, + "step": 148, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 20 }, { - "epoch": 0.29347826086956524, - "grad_norm": 2.2368586683604175, - "learning_rate": 4.153641065640402e-06, - "loss": 0.238, - "step": 99 - }, - { - "epoch": 0.29347826086956524, - "step": 99, - "train_accuracy_first_token_arguments": 0.9375, - "train_accuracy_first_token_arguments_total": 16, - "train_accuracy_recipient_token": 0.95, + "epoch": 0.2924901185770751, + "step": 148, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 0.8, + "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 20 }, { - "epoch": 0.29347826086956524, - "step": 99, - "train_accuracy_first_token_arguments": 0.8, - "train_accuracy_first_token_arguments_total": 20, - "train_accuracy_recipient_token": 0.875, + "epoch": 0.29446640316205536, + "grad_norm": 1.8557301698416115, + "learning_rate": 4.144846814849282e-06, + "loss": 0.1994, + "step": 149 + }, + { + "epoch": 0.29446640316205536, + "step": 149, + "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_clean": 1.0, @@ -5523,15 +5275,13 @@ "train_accuracy_recipient_token_install_total": 1, "train_accuracy_recipient_token_remove": 0.0, "train_accuracy_recipient_token_remove_total": 1, - "train_accuracy_recipient_token_replace": 0.0, + "train_accuracy_recipient_token_replace": 1.0, "train_accuracy_recipient_token_replace_total": 1, "train_total_number_recipient_token": 16 }, { - "epoch": 0.29347826086956524, - "step": 99, - "train_accuracy_first_token_arguments": 0.9411764705882353, - "train_accuracy_first_token_arguments_total": 17, + "epoch": 0.29446640316205536, + "step": 149, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, @@ -5545,58 +5295,59 @@ }, { "epoch": 0.2964426877470356, - "grad_norm": 2.6698723558857944, - "learning_rate": 4.1354959800246155e-06, - "loss": 0.2791, - "step": 100 + "grad_norm": 3.3999140933900542, + "learning_rate": 4.132742459383122e-06, + "loss": 0.2816, + "step": 150 }, { "epoch": 0.2964426877470356, - "step": 100, - "train_accuracy_first_token_arguments": 0.6666666666666666, - "train_accuracy_first_token_arguments_total": 21, - "train_accuracy_recipient_token": 0.8888888888888888, + "step": 150, + "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, - "train_accuracy_recipient_token_facility": 1.0, - "train_accuracy_recipient_token_facility_total": 3, + "train_accuracy_recipient_token_fac": 0.6666666666666666, + "train_accuracy_recipient_token_fac_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.2964426877470356, - "step": 100, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 15, - "train_accuracy_recipient_token": 0.9473684210526315, + "step": 150, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_check": 0.0, + "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_mount": 1.0, "train_accuracy_recipient_token_mount_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.2964426877470356, - "step": 100, - "train_accuracy_first_token_arguments": 0.71875, - "train_accuracy_first_token_arguments_total": 32, + "epoch": 0.2984189723320158, + "grad_norm": 1.6480321230907027, + "learning_rate": 4.120570988269472e-06, + "loss": 0.1801, + "step": 151 + }, + { + "epoch": 0.2984189723320158, + "step": 151, "train_accuracy_recipient_token": 0.8421052631578947, - "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 0.0, "train_accuracy_recipient_token_calculate_total": 1, - "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, @@ -5605,35 +5356,31 @@ "train_total_number_recipient_token": 19 }, { - "epoch": 0.2994071146245059, - "grad_norm": 1.5724579692892722, - "learning_rate": 4.1171990108847705e-06, - "loss": 0.1851, - "step": 101 - }, - { - "epoch": 0.2994071146245059, - "step": 101, - "train_accuracy_first_token_arguments": 0.9069767441860465, - "train_accuracy_first_token_arguments_total": 43, - "train_accuracy_recipient_token": 0.9, + "epoch": 0.2984189723320158, + "step": 151, + "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_accuracy_recipient_token_schedule": 0.0, "train_accuracy_recipient_token_schedule_total": 1, - "train_total_number_recipient_token": 20 + "train_total_number_recipient_token": 19 }, { - "epoch": 0.2994071146245059, - "step": 101, - "train_accuracy_first_token_arguments": 0.5909090909090909, - "train_accuracy_first_token_arguments_total": 22, + "epoch": 0.30039525691699603, + "grad_norm": 1.2847579476886248, + "learning_rate": 4.1083329018298356e-06, + "loss": 0.1073, + "step": 152 + }, + { + "epoch": 0.30039525691699603, + "step": 152, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, @@ -5648,14 +5395,12 @@ "train_total_number_recipient_token": 19 }, { - "epoch": 0.2994071146245059, - "step": 101, - "train_accuracy_first_token_arguments": 0.7692307692307693, - "train_accuracy_first_token_arguments_total": 13, - "train_accuracy_recipient_token": 0.8333333333333334, + "epoch": 0.30039525691699603, + "step": 152, + "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, @@ -5663,24 +5408,22 @@ }, { "epoch": 0.30237154150197626, - "grad_norm": 1.3808933034142918, - "learning_rate": 4.098751857404595e-06, - "loss": 0.1429, - "step": 102 + "grad_norm": 1.6318246749351943, + "learning_rate": 4.096028703124014e-06, + "loss": 0.1468, + "step": 153 }, { "epoch": 0.30237154150197626, - "step": 102, - "train_accuracy_first_token_arguments": 0.9166666666666666, - "train_accuracy_first_token_arguments_total": 24, - "train_accuracy_recipient_token": 0.8235294117647058, - "train_accuracy_recipient_token_all": 0.7272727272727273, + "step": 153, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_detect": 1.0, "train_accuracy_recipient_token_detect_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_sum": 1.0, + "train_accuracy_recipient_token_sum": 0.0, "train_accuracy_recipient_token_sum_total": 1, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 1, @@ -5688,9 +5431,7 @@ }, { "epoch": 0.30237154150197626, - "step": 102, - "train_accuracy_first_token_arguments": 0.5833333333333334, - "train_accuracy_first_token_arguments_total": 12, + "step": 153, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, @@ -5700,44 +5441,40 @@ "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 17 }, { - "epoch": 0.30237154150197626, - "step": 102, - "train_accuracy_first_token_arguments": 0.8421052631578947, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.30434782608695654, + "grad_norm": 2.3337051073106223, + "learning_rate": 4.083658897929425e-06, + "loss": 0.3715, + "step": 154 + }, + { + "epoch": 0.30434782608695654, + "step": 154, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_event": 1.0, "train_accuracy_recipient_token_event_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.30533596837944665, - "grad_norm": 2.4623439109068372, - "learning_rate": 4.080156232714976e-06, - "loss": 0.4204, - "step": 103 - }, - { - "epoch": 0.30533596837944665, - "step": 103, - "train_accuracy_first_token_arguments": 0.95, - "train_accuracy_first_token_arguments_total": 20, - "train_accuracy_recipient_token": 0.9473684210526315, + "epoch": 0.30434782608695654, + "step": 154, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_bike": 1.0, "train_accuracy_recipient_token_bike_total": 2, - "train_accuracy_recipient_token_calculate": 0.0, + "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, @@ -5746,12 +5483,17 @@ "train_total_number_recipient_token": 19 }, { - "epoch": 0.30533596837944665, - "step": 103, - "train_accuracy_first_token_arguments": 0.9, - "train_accuracy_first_token_arguments_total": 10, - "train_accuracy_recipient_token": 0.9375, - "train_accuracy_recipient_token_all": 0.9166666666666666, + "epoch": 0.30632411067193677, + "grad_norm": 1.9972989241470553, + "learning_rate": 4.071223994720309e-06, + "loss": 0.2331, + "step": 155 + }, + { + "epoch": 0.30632411067193677, + "step": 155, + "train_accuracy_recipient_token": 0.875, + "train_accuracy_recipient_token_all": 0.8333333333333334, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_disc": 1.0, "train_accuracy_recipient_token_disc_total": 1, @@ -5760,67 +5502,66 @@ "train_total_number_recipient_token": 16 }, { - "epoch": 0.30533596837944665, - "step": 103, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.30632411067193677, + "step": 155, + "train_accuracy_recipient_token": 0.9047619047619048, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_convert": 1.0, "train_accuracy_recipient_token_convert_total": 2, - "train_accuracy_recipient_token_count": 1.0, + "train_accuracy_recipient_token_count": 0.5, "train_accuracy_recipient_token_count_total": 2, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 21 }, { "epoch": 0.308300395256917, - "grad_norm": 1.9389616807058327, - "learning_rate": 4.061413863734869e-06, - "loss": 0.169, - "step": 104 + "grad_norm": 2.4018437558185077, + "learning_rate": 4.058724504646834e-06, + "loss": 0.1986, + "step": 156 }, { "epoch": 0.308300395256917, - "step": 104, - "train_accuracy_first_token_arguments": 0.7, - "train_accuracy_first_token_arguments_total": 20, - "train_accuracy_recipient_token": 0.9473684210526315, + "step": 156, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 2, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 2, "train_accuracy_recipient_token_interpret": 1.0, "train_accuracy_recipient_token_interpret_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.308300395256917, - "step": 104, - "train_accuracy_first_token_arguments": 0.9523809523809523, - "train_accuracy_first_token_arguments_total": 21, - "train_accuracy_recipient_token": 0.95, + "step": 156, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 5, - "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 20 }, { - "epoch": 0.308300395256917, - "step": 104, - "train_accuracy_first_token_arguments": 0.8421052631578947, - "train_accuracy_first_token_arguments_total": 19, + "epoch": 0.3102766798418972, + "grad_norm": 1.9392056531490096, + "learning_rate": 4.046160941514079e-06, + "loss": 0.1503, + "step": 157 + }, + { + "epoch": 0.3102766798418972, + "step": 157, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, @@ -5830,22 +5571,13 @@ "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 1, - "train_accuracy_recipient_token_villa": 1.0, - "train_accuracy_recipient_token_villa_total": 2, + "train_accuracy_recipient_token_v": 1.0, + "train_accuracy_recipient_token_v_total": 2, "train_total_number_recipient_token": 18 }, { - "epoch": 0.31126482213438733, - "grad_norm": 1.819003861346643, - "learning_rate": 4.0425264910109245e-06, - "loss": 0.1843, - "step": 105 - }, - { - "epoch": 0.31126482213438733, - "step": 105, - "train_accuracy_first_token_arguments": 0.8333333333333334, - "train_accuracy_first_token_arguments_total": 12, + "epoch": 0.3102766798418972, + "step": 157, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, @@ -5853,17 +5585,22 @@ "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_m": 1.0, "train_accuracy_recipient_token_m_total": 1, "train_total_number_recipient_token": 17 }, { - "epoch": 0.31126482213438733, - "step": 105, - "train_accuracy_first_token_arguments": 0.875, - "train_accuracy_first_token_arguments_total": 8, + "epoch": 0.31225296442687744, + "grad_norm": 2.00451491556449, + "learning_rate": 4.033533821760917e-06, + "loss": 0.3046, + "step": 158 + }, + { + "epoch": 0.31225296442687744, + "step": 158, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, @@ -5878,10 +5615,8 @@ "train_total_number_recipient_token": 16 }, { - "epoch": 0.31126482213438733, - "step": 105, - "train_accuracy_first_token_arguments": 0.8888888888888888, - "train_accuracy_first_token_arguments_total": 9, + "epoch": 0.31225296442687744, + "step": 158, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, @@ -5891,39 +5626,14 @@ }, { "epoch": 0.3142292490118577, - "grad_norm": 2.4539744307883553, - "learning_rate": 4.023495868555848e-06, - "loss": 0.4031, - "step": 106 - }, - { - "epoch": 0.3142292490118577, - "eval_accuracy_first_token_arguments": 0.9266727772685609, - "eval_accuracy_first_token_arguments_total": 1091, - "eval_accuracy_recipient_token": 0.9235033259423503, - "eval_accuracy_recipient_token_all": 0.9044834307992202, - "eval_accuracy_recipient_token_all_total": 513, - "eval_accuracy_recipient_token_calculate": 1.0, - "eval_accuracy_recipient_token_calculate_total": 13, - "eval_accuracy_recipient_token_find": 0.9649122807017544, - "eval_accuracy_recipient_token_find_total": 57, - "eval_accuracy_recipient_token_get": 0.9744897959183674, - "eval_accuracy_recipient_token_get_total": 196, - "eval_accuracy_recipient_token_identify": 0.8333333333333334, - "eval_accuracy_recipient_token_identify_total": 6, - "eval_loss": 0.25004321336746216, - "eval_perplexity": 1.0396230889094866, - "eval_runtime": 119.1349, - "eval_samples_per_second": 0.848, - "eval_steps_per_second": 0.428, - "eval_total_number_recipient_token": 902, - "step": 106 + "grad_norm": 3.024122625725331, + "learning_rate": 4.020843664438783e-06, + "loss": 0.3273, + "step": 159 }, { "epoch": 0.3142292490118577, - "step": 106, - "train_accuracy_first_token_arguments": 0.8, - "train_accuracy_first_token_arguments_total": 20, + "step": 159, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, @@ -5937,31 +5647,34 @@ }, { "epoch": 0.3142292490118577, - "step": 106, - "train_accuracy_first_token_arguments": 0.88, - "train_accuracy_first_token_arguments_total": 25, - "train_accuracy_recipient_token": 1.0, + "step": 159, + "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_report": 1.0, + "train_accuracy_recipient_token_report": 0.0, "train_accuracy_recipient_token_report_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.3142292490118577, - "step": 106, - "train_accuracy_first_token_arguments": 0.4166666666666667, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.31620553359683795, + "grad_norm": 1.8881036988631603, + "learning_rate": 4.008090991190341e-06, + "loss": 0.1621, + "step": 160 + }, + { + "epoch": 0.31620553359683795, + "step": 160, + "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, @@ -5970,19 +5683,10 @@ "train_total_number_recipient_token": 18 }, { - "epoch": 0.31719367588932806, - "grad_norm": 2.6219196171389454, - "learning_rate": 4.004323763685511e-06, - "loss": 0.2683, - "step": 107 - }, - { - "epoch": 0.31719367588932806, - "step": 107, - "train_accuracy_first_token_arguments": 0.9130434782608695, - "train_accuracy_first_token_arguments_total": 23, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.8181818181818182, + "epoch": 0.31620553359683795, + "step": 160, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, @@ -5995,11 +5699,16 @@ "train_total_number_recipient_token": 18 }, { - "epoch": 0.31719367588932806, - "step": 107, - "train_accuracy_first_token_arguments": 0.918918918918919, - "train_accuracy_first_token_arguments_total": 37, - "train_accuracy_recipient_token": 0.9411764705882353, + "epoch": 0.3181818181818182, + "grad_norm": 3.3688529657875326, + "learning_rate": 3.99527632622804e-06, + "loss": 0.3867, + "step": 161 + }, + { + "epoch": 0.3181818181818182, + "step": 161, + "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_audit": 1.0, @@ -6008,15 +5717,13 @@ "train_accuracy_recipient_token_environment_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_public": 1.0, + "train_accuracy_recipient_token_public": 0.0, "train_accuracy_recipient_token_public_total": 1, "train_total_number_recipient_token": 17 }, { - "epoch": 0.31719367588932806, - "step": 107, - "train_accuracy_first_token_arguments": 0.8333333333333334, - "train_accuracy_first_token_arguments_total": 18, + "epoch": 0.3181818181818182, + "step": 161, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, @@ -6032,20 +5739,18 @@ }, { "epoch": 0.3201581027667984, - "grad_norm": 2.3979253771216524, - "learning_rate": 3.985011956854826e-06, - "loss": 0.3371, - "step": 108 + "grad_norm": 2.245641056003419, + "learning_rate": 3.982400196312565e-06, + "loss": 0.2937, + "step": 162 }, { "epoch": 0.3201581027667984, - "step": 108, - "train_accuracy_first_token_arguments": 0.9285714285714286, - "train_accuracy_first_token_arguments_total": 14, + "step": 162, "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check": 0.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, @@ -6055,11 +5760,9 @@ }, { "epoch": 0.3201581027667984, - "step": 108, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 18, - "train_accuracy_recipient_token": 0.875, - "train_accuracy_recipient_token_all": 0.8, + "step": 162, + "train_accuracy_recipient_token": 0.9375, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 2, @@ -6070,35 +5773,31 @@ "train_total_number_recipient_token": 16 }, { - "epoch": 0.3201581027667984, - "step": 108, - "train_accuracy_first_token_arguments": 0.8571428571428571, - "train_accuracy_first_token_arguments_total": 28, - "train_accuracy_recipient_token": 0.5, + "epoch": 0.3221343873517787, + "grad_norm": 2.548113126420555, + "learning_rate": 3.969463130731183e-06, + "loss": 0.2783, + "step": 163 + }, + { + "epoch": 0.3221343873517787, + "step": 163, + "train_accuracy_recipient_token": 0.5625, "train_accuracy_recipient_token_all": 0.5454545454545454, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_detect": 0.0, "train_accuracy_recipient_token_detect_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_translate": 0.5, + "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 2, "train_total_number_recipient_token": 16 }, { - "epoch": 0.3231225296442688, - "grad_norm": 2.9561525888835276, - "learning_rate": 3.965562241492401e-06, - "loss": 0.3818, - "step": 109 - }, - { - "epoch": 0.3231225296442688, - "step": 109, - "train_accuracy_first_token_arguments": 0.7142857142857143, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.3221343873517787, + "step": 163, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, @@ -6106,15 +5805,20 @@ "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_plan": 1.0, "train_accuracy_recipient_token_plan_total": 1, - "train_accuracy_recipient_token_suggest": 1.0, - "train_accuracy_recipient_token_suggest_total": 1, + "train_accuracy_recipient_token_s": 1.0, + "train_accuracy_recipient_token_s_total": 1, "train_total_number_recipient_token": 17 }, { - "epoch": 0.3231225296442688, - "step": 109, - "train_accuracy_first_token_arguments": 0.9230769230769231, - "train_accuracy_first_token_arguments_total": 13, + "epoch": 0.3241106719367589, + "grad_norm": 2.7435239695326823, + "learning_rate": 3.9564656612759904e-06, + "loss": 0.2899, + "step": 164 + }, + { + "epoch": 0.3241106719367589, + "step": 164, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, @@ -6123,37 +5827,33 @@ "train_total_number_recipient_token": 17 }, { - "epoch": 0.3231225296442688, - "step": 109, - "train_accuracy_first_token_arguments": 0.96875, - "train_accuracy_first_token_arguments_total": 32, + "epoch": 0.3241106719367589, + "step": 164, "train_accuracy_recipient_token": 0.8421052631578947, - "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, - "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 1, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.32608695652173914, - "grad_norm": 1.9549696262976763, - "learning_rate": 3.945976423833987e-06, - "loss": 0.1918, - "step": 110 + "grad_norm": 2.2167192768597896, + "learning_rate": 3.943408322222049e-06, + "loss": 0.2021, + "step": 165 }, { "epoch": 0.32608695652173914, - "step": 110, - "train_accuracy_first_token_arguments": 0.9333333333333333, - "train_accuracy_first_token_arguments_total": 30, - "train_accuracy_recipient_token": 0.8333333333333334, - "train_accuracy_recipient_token_all": 0.7692307692307693, + "step": 165, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8461538461538461, "train_accuracy_recipient_token_all_total": 13, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, @@ -6167,9 +5867,7 @@ }, { "epoch": 0.32608695652173914, - "step": 110, - "train_accuracy_first_token_arguments": 0.7894736842105263, - "train_accuracy_first_token_arguments_total": 38, + "step": 165, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, @@ -6177,17 +5875,22 @@ "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_race": 1.0, "train_accuracy_recipient_token_race_total": 2, "train_total_number_recipient_token": 16 }, { - "epoch": 0.32608695652173914, - "step": 110, - "train_accuracy_first_token_arguments": 0.9333333333333333, - "train_accuracy_first_token_arguments_total": 15, + "epoch": 0.32806324110671936, + "grad_norm": 2.363046339246385, + "learning_rate": 3.930291650305424e-06, + "loss": 0.3441, + "step": 166 + }, + { + "epoch": 0.32806324110671936, + "step": 166, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9166666666666666, "train_accuracy_recipient_token_all_total": 12, @@ -6198,37 +5901,33 @@ "train_total_number_recipient_token": 20 }, { - "epoch": 0.3290513833992095, - "grad_norm": 2.295435621606061, - "learning_rate": 3.92625632275474e-06, - "loss": 0.378, - "step": 111 - }, - { - "epoch": 0.3290513833992095, - "step": 111, - "train_accuracy_first_token_arguments": 0.8888888888888888, - "train_accuracy_first_token_arguments_total": 18, - "train_accuracy_recipient_token": 0.8333333333333334, + "epoch": 0.32806324110671936, + "step": 166, + "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_add": 1.0, "train_accuracy_recipient_token_add_total": 1, - "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_change": 1.0, + "train_accuracy_recipient_token_change": 0.0, "train_accuracy_recipient_token_change_total": 1, - "train_accuracy_recipient_token_highlight": 0.5, + "train_accuracy_recipient_token_highlight": 1.0, "train_accuracy_recipient_token_highlight_total": 2, "train_accuracy_recipient_token_remove": 1.0, "train_accuracy_recipient_token_remove_total": 2, "train_total_number_recipient_token": 18 }, { - "epoch": 0.3290513833992095, - "step": 111, - "train_accuracy_first_token_arguments": 0.896551724137931, - "train_accuracy_first_token_arguments_total": 29, - "train_accuracy_recipient_token": 0.7647058823529411, - "train_accuracy_recipient_token_all": 0.8181818181818182, + "epoch": 0.3300395256916996, + "grad_norm": 2.6127483795239708, + "learning_rate": 3.917116184701125e-06, + "loss": 0.2668, + "step": 167 + }, + { + "epoch": 0.3300395256916996, + "step": 167, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, @@ -6241,10 +5940,8 @@ "train_total_number_recipient_token": 17 }, { - "epoch": 0.3290513833992095, - "step": 111, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 16, + "epoch": 0.3300395256916996, + "step": 167, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, @@ -6252,4964 +5949,6915 @@ "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 2, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 2, "train_total_number_recipient_token": 20 }, { "epoch": 0.33201581027667987, - "grad_norm": 2.4185983751815114, - "learning_rate": 3.906403769600311e-06, - "loss": 0.2139, - "step": 112 + "grad_norm": 1.9279791632118746, + "learning_rate": 3.903882467000938e-06, + "loss": 0.1514, + "step": 168 }, { "epoch": 0.33201581027667987, - "step": 112, - "train_accuracy_first_token_arguments": 0.9090909090909091, - "train_accuracy_first_token_arguments_total": 22, - "train_accuracy_recipient_token": 0.8947368421052632, + "step": 168, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_dispose": 1.0, - "train_accuracy_recipient_token_dispose_total": 1, - "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_up": 1.0, - "train_accuracy_recipient_token_up_total": 1, + "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_order": 1.0, + "train_accuracy_recipient_token_order_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.33201581027667987, - "step": 112, - "train_accuracy_first_token_arguments": 0.9230769230769231, - "train_accuracy_first_token_arguments_total": 13, - "train_accuracy_recipient_token": 0.9411764705882353, + "step": 168, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 0.8, - "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_all_total": 8, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 14 }, { "epoch": 0.33201581027667987, - "step": 112, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 13, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, + "step": 168, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.8, + "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 8, - "train_accuracy_recipient_token_verify": 1.0, - "train_accuracy_recipient_token_verify_total": 1, + "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 19 }, { - "epoch": 0.3349802371541502, - "grad_norm": 2.321966400855487, - "learning_rate": 3.886420608016767e-06, - "loss": 0.2771, - "step": 113 - }, - { - "epoch": 0.3349802371541502, - "step": 113, - "train_accuracy_first_token_arguments": 0.9666666666666667, - "train_accuracy_first_token_arguments_total": 30, + "epoch": 0.33201581027667987, + "step": 168, "train_accuracy_recipient_token": 0.9, - "train_accuracy_recipient_token_add": 1.0, - "train_accuracy_recipient_token_add_total": 1, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_change": 1.0, - "train_accuracy_recipient_token_change_total": 1, - "train_accuracy_recipient_token_get": 0.8, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_replace": 0.5, - "train_accuracy_recipient_token_replace_total": 2, + "train_accuracy_recipient_token_find": 0.6666666666666666, + "train_accuracy_recipient_token_find_total": 6, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 20 }, { - "epoch": 0.3349802371541502, - "step": 113, - "train_accuracy_first_token_arguments": 0.6, - "train_accuracy_first_token_arguments_total": 10, - "train_accuracy_recipient_token": 0.9411764705882353, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 2, - "train_accuracy_recipient_token_bird": 1.0, - "train_accuracy_recipient_token_bird_total": 4, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_find": 0.6, + "train_accuracy_recipient_token_find_total": 5, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 1, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.3349802371541502, - "step": 113, - "train_accuracy_first_token_arguments": 0.5833333333333334, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.8947368421052632, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 0.7777777777777778, - "train_accuracy_recipient_token_get_total": 9, - "train_total_number_recipient_token": 19 + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.875, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, + "train_accuracy_recipient_token_get": 0.5, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_ident": 0.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 2, + "train_total_number_recipient_token": 16 }, { - "epoch": 0.33794466403162055, - "grad_norm": 2.028221423810436, - "learning_rate": 3.8663086937793845e-06, - "loss": 0.2435, - "step": 114 + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_histor": 1.0, + "train_accuracy_recipient_token_histor_total": 3, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.33794466403162055, - "step": 114, - "train_accuracy_first_token_arguments": 0.9473684210526315, - "train_accuracy_first_token_arguments_total": 114, - "train_accuracy_recipient_token": 0.7368421052631579, + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_arch": 0.0, - "train_accuracy_recipient_token_arch_total": 1, - "train_accuracy_recipient_token_building": 1.0, - "train_accuracy_recipient_token_building_total": 1, - "train_accuracy_recipient_token_get": 0.5, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_histor": 0.0, - "train_accuracy_recipient_token_histor_total": 1, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 1, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.33794466403162055, - "step": 114, - "train_accuracy_first_token_arguments": 0.5833333333333334, - "train_accuracy_first_token_arguments_total": 36, - "train_accuracy_recipient_token": 0.95, + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 3, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_im": 1.0, - "train_accuracy_recipient_token_im_total": 1, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 4, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.33794466403162055, - "step": 114, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 11, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 12, + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 7, - "train_total_number_recipient_token": 19 - }, - { - "epoch": 0.3409090909090909, - "grad_norm": 1.8641084259637979, - "learning_rate": 3.846069894620306e-06, - "loss": 0.2402, - "step": 115 + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.3409090909090909, - "step": 115, - "train_accuracy_first_token_arguments": 0.75, - "train_accuracy_first_token_arguments_total": 24, - "train_accuracy_recipient_token": 0.8947368421052632, + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 0.6, "train_accuracy_recipient_token_find_total": 5, - "train_accuracy_recipient_token_organ": 1.0, - "train_accuracy_recipient_token_organ_total": 1, - "train_accuracy_recipient_token_schedule": 1.0, - "train_accuracy_recipient_token_schedule_total": 1, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_ident": 0.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_interpret": 1.0, + "train_accuracy_recipient_token_interpret_total": 2, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.3409090909090909, - "step": 115, - "train_accuracy_first_token_arguments": 0.8181818181818182, - "train_accuracy_first_token_arguments_total": 22, - "train_accuracy_recipient_token": 0.8235294117647058, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compare": 0.0, - "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_get": 0.8, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_total_number_recipient_token": 17 + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.8666666666666667, + "train_accuracy_recipient_token_all_total": 15, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_find": 0.6666666666666666, + "train_accuracy_recipient_token_find_total": 3, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.3409090909090909, - "step": 115, - "train_accuracy_first_token_arguments": 0.84, - "train_accuracy_first_token_arguments_total": 25, - "train_accuracy_recipient_token": 0.9, + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.75, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_bookmark": 1.0, - "train_accuracy_recipient_token_bookmark_total": 1, - "train_accuracy_recipient_token_get": 0.8, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_refresh": 1.0, - "train_accuracy_recipient_token_refresh_total": 1, - "train_accuracy_recipient_token_switch": 0.5, - "train_accuracy_recipient_token_switch_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.25, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_tour": 0.5, + "train_accuracy_recipient_token_tour_total": 4, "train_total_number_recipient_token": 20 }, { - "epoch": 0.3438735177865613, - "grad_norm": 2.0312557408429788, - "learning_rate": 3.825706090055088e-06, - "loss": 0.2456, - "step": 116 + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_book": 0.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.8333333333333334, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.3438735177865613, - "step": 116, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 0.95, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_remove": 1.0, - "train_accuracy_recipient_token_remove_total": 1, - "train_total_number_recipient_token": 20 + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 3, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.3438735177865613, - "step": 116, - "train_accuracy_first_token_arguments": 0.7391304347826086, - "train_accuracy_first_token_arguments_total": 23, - "train_accuracy_recipient_token": 0.8823529411764706, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.85, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_suggest": 1.0, - "train_accuracy_recipient_token_suggest_total": 1, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_hang": 0.0, + "train_accuracy_recipient_token_hang_total": 1, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, + "train_accuracy_recipient_token_remove": 1.0, + "train_accuracy_recipient_token_remove_total": 1, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.3438735177865613, - "step": 116, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 14, + "epoch": 0.33201581027667987, + "step": 168, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_book": 1.0, - "train_accuracy_recipient_token_book_total": 1, - "train_accuracy_recipient_token_create": 1.0, - "train_accuracy_recipient_token_create_total": 2, - "train_accuracy_recipient_token_design": 1.0, - "train_accuracy_recipient_token_design_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, - "train_total_number_recipient_token": 16 - }, - { - "epoch": 0.3468379446640316, - "grad_norm": 2.9747145067980383, - "learning_rate": 3.80521917120816e-06, - "loss": 0.3586, - "step": 117 + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_invest": 1.0, + "train_accuracy_recipient_token_invest_total": 1, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.3468379446640316, - "step": 117, - "train_accuracy_first_token_arguments": 0.8043478260869565, - "train_accuracy_first_token_arguments_total": 46, - "train_accuracy_recipient_token": 0.9375, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.875, + "train_accuracy_recipient_token_add": 1.0, + "train_accuracy_recipient_token_add_total": 1, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_construct": 0.0, + "train_accuracy_recipient_token_construct_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_select": 1.0, + "train_accuracy_recipient_token_select_total": 2, "train_total_number_recipient_token": 16 }, { - "epoch": 0.3468379446640316, - "step": 117, - "train_accuracy_first_token_arguments": 0.8888888888888888, - "train_accuracy_first_token_arguments_total": 27, - "train_accuracy_recipient_token": 0.9411764705882353, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_count": 1.0, - "train_accuracy_recipient_token_count_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { - "epoch": 0.3468379446640316, - "step": 117, - "train_accuracy_first_token_arguments": 0.7894736842105263, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.8235294117647058, - "train_accuracy_recipient_token_all": 0.7272727272727273, + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_art": 1.0, + "train_accuracy_recipient_token_art_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_setup": 1.0, - "train_accuracy_recipient_token_setup_total": 2, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_histor": 0.0, + "train_accuracy_recipient_token_histor_total": 1, + "train_accuracy_recipient_token_m": 1.0, + "train_accuracy_recipient_token_m_total": 1, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.34980237154150196, - "grad_norm": 2.619218782176094, - "learning_rate": 3.784611040637198e-06, - "loss": 0.386, - "step": 118 - }, - { - "epoch": 0.34980237154150196, - "step": 118, - "train_accuracy_first_token_arguments": 0.9285714285714286, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 0.8888888888888888, + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 0.75, - "train_accuracy_recipient_token_get_total": 8, - "train_total_number_recipient_token": 18 - }, - { - "epoch": 0.34980237154150196, - "step": 118, - "train_accuracy_first_token_arguments": 0.9, - "train_accuracy_first_token_arguments_total": 10, - "train_accuracy_recipient_token": 0.9411764705882353, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_project": 1.0, - "train_accuracy_recipient_token_project_total": 1, - "train_accuracy_recipient_token_sc": 1.0, - "train_accuracy_recipient_token_sc_total": 2, - "train_accuracy_recipient_token_verify": 1.0, - "train_accuracy_recipient_token_verify_total": 1, - "train_total_number_recipient_token": 17 - }, - { - "epoch": 0.34980237154150196, - "step": 118, - "train_accuracy_first_token_arguments": 0.76, - "train_accuracy_first_token_arguments_total": 25, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_calculate": 1.0, - "train_accuracy_recipient_token_calculate_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 6, + "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_total_number_recipient_token": 19 - }, - { - "epoch": 0.35276679841897235, - "grad_norm": 1.783725464528477, - "learning_rate": 3.7638836121564414e-06, - "loss": 0.3088, - "step": 119 + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_gr": 1.0, + "train_accuracy_recipient_token_gr_total": 3, + "train_accuracy_recipient_token_ident": 0.0, + "train_accuracy_recipient_token_ident_total": 2, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.35276679841897235, - "step": 119, - "train_accuracy_first_token_arguments": 0.8695652173913043, - "train_accuracy_first_token_arguments_total": 23, - "train_accuracy_recipient_token": 0.95, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 0.9, - "train_accuracy_recipient_token_get_total": 10, - "train_total_number_recipient_token": 20 + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.8571428571428571, + "train_accuracy_recipient_token_all": 0.8888888888888888, + "train_accuracy_recipient_token_all_total": 9, + "train_accuracy_recipient_token_draw": 1.0, + "train_accuracy_recipient_token_draw_total": 1, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 14 }, { - "epoch": 0.35276679841897235, - "step": 119, - "train_accuracy_first_token_arguments": 0.9090909090909091, - "train_accuracy_first_token_arguments_total": 11, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_financial": 1.0, - "train_accuracy_recipient_token_financial_total": 4, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_invest": 1.0, - "train_accuracy_recipient_token_invest_total": 1, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_review": 0.5, + "train_accuracy_recipient_token_review_total": 2, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.35276679841897235, - "step": 119, - "train_accuracy_first_token_arguments": 0.967741935483871, - "train_accuracy_first_token_arguments_total": 31, - "train_accuracy_recipient_token": 0.9473684210526315, + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 7, - "train_total_number_recipient_token": 19 - }, - { - "epoch": 0.3557312252964427, - "grad_norm": 1.880678138487034, - "learning_rate": 3.7430388106589632e-06, - "loss": 0.2326, - "step": 120 + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.3557312252964427, - "step": 120, - "train_accuracy_first_token_arguments": 0.7894736842105263, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.9411764705882353, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 1, - "train_accuracy_recipient_token_detect": 1.0, - "train_accuracy_recipient_token_detect_total": 1, - "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_create": 0.5, + "train_accuracy_recipient_token_create_total": 2, + "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 17 }, { - "epoch": 0.3557312252964427, - "step": 120, - "train_accuracy_first_token_arguments": 0.7916666666666666, - "train_accuracy_first_token_arguments_total": 24, - "train_accuracy_recipient_token": 0.9411764705882353, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_book": 1.0, - "train_accuracy_recipient_token_book_total": 2, - "train_accuracy_recipient_token_buy": 1.0, - "train_accuracy_recipient_token_buy_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_suggest": 1.0, - "train_accuracy_recipient_token_suggest_total": 2, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_get": 0.875, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.3557312252964427, - "step": 120, - "train_accuracy_first_token_arguments": 0.8947368421052632, - "train_accuracy_first_token_arguments_total": 57, - "train_accuracy_recipient_token": 0.8333333333333334, - "train_accuracy_recipient_token_all": 0.8, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_calculate": 1.0, - "train_accuracy_recipient_token_calculate_total": 3, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 0.75, - "train_accuracy_recipient_token_get_total": 4, - "train_total_number_recipient_token": 18 + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 0.875, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.358695652173913, - "grad_norm": 2.3660230978435117, - "learning_rate": 3.72207857193791e-06, - "loss": 0.2103, - "step": 121 + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.7894736842105263, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 9, + "train_accuracy_recipient_token_custom": 0.5, + "train_accuracy_recipient_token_custom_total": 4, + "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.358695652173913, - "step": 121, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 9, - "train_accuracy_recipient_token": 0.7647058823529411, - "train_accuracy_recipient_token_all": 0.6, + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_eng": 0.5, + "train_accuracy_recipient_token_eng_total": 2, + "train_accuracy_recipient_token_list": 1.0, + "train_accuracy_recipient_token_list_total": 1, + "train_accuracy_recipient_token_lock": 1.0, + "train_accuracy_recipient_token_lock_total": 1, + "train_accuracy_recipient_token_tem": 1.0, + "train_accuracy_recipient_token_tem_total": 3, "train_total_number_recipient_token": 17 }, { - "epoch": 0.358695652173913, - "step": 121, - "train_accuracy_first_token_arguments": 0.7575757575757576, - "train_accuracy_first_token_arguments_total": 33, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_cl": 1.0, - "train_accuracy_recipient_token_cl_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_im": 1.0, - "train_accuracy_recipient_token_im_total": 1, - "train_accuracy_recipient_token_search": 0.5, - "train_accuracy_recipient_token_search_total": 2, - "train_total_number_recipient_token": 18 + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.9375, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 16 }, { - "epoch": 0.358695652173913, - "step": 121, - "train_accuracy_first_token_arguments": 0.8260869565217391, - "train_accuracy_first_token_arguments_total": 23, + "epoch": 0.33201581027667987, + "step": 168, "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_car": 1.0, + "train_accuracy_recipient_token_car_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 6, - "train_accuracy_recipient_token_list": 1.0, - "train_accuracy_recipient_token_list_total": 1, - "train_accuracy_recipient_token_schedule": 1.0, - "train_accuracy_recipient_token_schedule_total": 1, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 0.5, + "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 19 }, { - "epoch": 0.3616600790513834, - "grad_norm": 1.9095006331871722, - "learning_rate": 3.7010048425067317e-06, - "loss": 0.2754, - "step": 122 + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.7222222222222222, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_get": 0.4, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_invest": 1.0, + "train_accuracy_recipient_token_invest_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.3616600790513834, - "step": 122, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 15, + "epoch": 0.33201581027667987, + "step": 168, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 9, + "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { - "epoch": 0.3616600790513834, - "step": 122, - "train_accuracy_first_token_arguments": 0.8571428571428571, - "train_accuracy_first_token_arguments_total": 28, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 1, - "train_accuracy_recipient_token_art": 1.0, - "train_accuracy_recipient_token_art_total": 3, + "train_accuracy_recipient_token_bird": 1.0, + "train_accuracy_recipient_token_bird_total": 2, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_histor": 1.0, + "train_accuracy_recipient_token_histor_total": 1, + "train_accuracy_recipient_token_symbol": 1.0, + "train_accuracy_recipient_token_symbol_total": 1, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.3616600790513834, - "step": 122, - "train_accuracy_first_token_arguments": 0.8666666666666667, - "train_accuracy_first_token_arguments_total": 45, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_decode": 1.0, - "train_accuracy_recipient_token_decode_total": 1, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 3, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_total_number_recipient_token": 19 - }, - { - "epoch": 0.36462450592885376, - "grad_norm": 1.6047057072988866, - "learning_rate": 3.679819579418414e-06, - "loss": 0.1743, - "step": 123 + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_organ": 0.0, + "train_accuracy_recipient_token_organ_total": 1, + "train_accuracy_recipient_token_rent": 1.0, + "train_accuracy_recipient_token_rent_total": 1, + "train_total_number_recipient_token": 16 }, { - "epoch": 0.36462450592885376, - "step": 123, - "train_accuracy_first_token_arguments": 0.71875, - "train_accuracy_first_token_arguments_total": 32, - "train_accuracy_recipient_token": 0.8823529411764706, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_arch": 1.0, - "train_accuracy_recipient_token_arch_total": 2, - "train_accuracy_recipient_token_build": 1.0, - "train_accuracy_recipient_token_build_total": 3, - "train_accuracy_recipient_token_histor": 0.0, - "train_accuracy_recipient_token_histor_total": 1, - "train_accuracy_recipient_token_suggest": 1.0, - "train_accuracy_recipient_token_suggest_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_ident": 0.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_predict": 1.0, + "train_accuracy_recipient_token_predict_total": 1, "train_total_number_recipient_token": 17 }, { - "epoch": 0.36462450592885376, - "step": 123, - "train_accuracy_first_token_arguments": 0.8, - "train_accuracy_first_token_arguments_total": 20, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9166666666666666, - "train_accuracy_recipient_token_all_total": 12, - "train_accuracy_recipient_token_book": 1.0, - "train_accuracy_recipient_token_book_total": 1, - "train_accuracy_recipient_token_check": 1.0, - "train_accuracy_recipient_token_check_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 5, - "train_total_number_recipient_token": 19 + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_evaluate": 0.0, + "train_accuracy_recipient_token_evaluate_total": 1, + "train_accuracy_recipient_token_find": 0.3333333333333333, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.36462450592885376, - "step": 123, - "train_accuracy_first_token_arguments": 0.9375, - "train_accuracy_first_token_arguments_total": 16, + "epoch": 0.33201581027667987, + "step": 168, "train_accuracy_recipient_token": 0.8947368421052632, - "train_accuracy_recipient_token_all": 0.8181818181818182, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_land": 1.0, - "train_accuracy_recipient_token_land_total": 4, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.7777777777777778, + "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { - "epoch": 0.3675889328063241, - "grad_norm": 1.6871352500543018, - "learning_rate": 3.658524750083733e-06, - "loss": 0.3106, - "step": 124 - }, - { - "epoch": 0.3675889328063241, - "step": 124, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 7, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_generate": 1.0, - "train_accuracy_recipient_token_generate_total": 1, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 16 + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_locate": 0.0, + "train_accuracy_recipient_token_locate_total": 1, + "train_accuracy_recipient_token_schedule": 0.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.3675889328063241, - "step": 124, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 7, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_park": 1.0, + "train_accuracy_recipient_token_park_total": 2, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.3675889328063241, - "step": 124, - "train_accuracy_first_token_arguments": 0.9090909090909091, - "train_accuracy_first_token_arguments_total": 22, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.7777777777777778, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_evaluate": 1.0, - "train_accuracy_recipient_token_evaluate_total": 1, - "train_accuracy_recipient_token_phrase": 1.0, - "train_accuracy_recipient_token_phrase_total": 4, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 1, - "train_accuracy_recipient_token_translate": 1.0, - "train_accuracy_recipient_token_translate_total": 1, - "train_total_number_recipient_token": 20 - }, - { - "epoch": 0.3705533596837945, - "grad_norm": 2.016083076720157, - "learning_rate": 3.6371223320885492e-06, - "loss": 0.3026, - "step": 125 + "train_accuracy_recipient_token_art": 0.6666666666666666, + "train_accuracy_recipient_token_art_total": 3, + "train_accuracy_recipient_token_artist": 0.0, + "train_accuracy_recipient_token_artist_total": 1, + "train_accuracy_recipient_token_get": 0.5, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.3705533596837945, - "step": 125, - "train_accuracy_first_token_arguments": 0.6764705882352942, - "train_accuracy_first_token_arguments_total": 34, + "epoch": 0.33201581027667987, + "step": 168, "train_accuracy_recipient_token": 0.8333333333333334, - "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_create": 1.0, - "train_accuracy_recipient_token_create_total": 1, - "train_accuracy_recipient_token_get": 0.8, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_plan": 0.0, - "train_accuracy_recipient_token_plan_total": 1, - "train_accuracy_recipient_token_suggest": 1.0, - "train_accuracy_recipient_token_suggest_total": 1, + "train_accuracy_recipient_token_analy": 0.5, + "train_accuracy_recipient_token_analy_total": 2, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 2, + "train_accuracy_recipient_token_object": 0.6666666666666666, + "train_accuracy_recipient_token_object_total": 3, + "train_accuracy_recipient_token_transfer": 0.0, + "train_accuracy_recipient_token_transfer_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.3705533596837945, - "step": 125, - "train_accuracy_first_token_arguments": 0.9090909090909091, - "train_accuracy_first_token_arguments_total": 11, + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_fruit": 1.0, + "train_accuracy_recipient_token_fruit_total": 4, + "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.33201581027667987, + "step": 168, "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { - "epoch": 0.3705533596837945, - "step": 125, - "train_accuracy_first_token_arguments": 0.9523809523809523, - "train_accuracy_first_token_arguments_total": 21, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 1, - "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_suggest": 1.0, - "train_accuracy_recipient_token_suggest_total": 1, - "train_total_number_recipient_token": 19 - }, - { - "epoch": 0.37351778656126483, - "grad_norm": 2.1413931479434973, - "learning_rate": 3.6156143130101516e-06, - "loss": 0.3208, - "step": 126 - }, - { - "epoch": 0.37351778656126483, - "step": 126, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 13, - "train_accuracy_recipient_token": 0.8666666666666667, - "train_accuracy_recipient_token_all": 0.8, + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.9375, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_organ": 1.0, "train_accuracy_recipient_token_organ_total": 1, - "train_total_number_recipient_token": 15 + "train_total_number_recipient_token": 16 }, { - "epoch": 0.37351778656126483, - "step": 126, - "train_accuracy_first_token_arguments": 0.9705882352941176, - "train_accuracy_first_token_arguments_total": 34, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_check": 1.0, - "train_accuracy_recipient_token_check_total": 1, - "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 2, + "train_accuracy_recipient_token_decode": 1.0, + "train_accuracy_recipient_token_decode_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 1, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_interpret": 1.0, + "train_accuracy_recipient_token_interpret_total": 1, + "train_accuracy_recipient_token_stage": 1.0, + "train_accuracy_recipient_token_stage_total": 4, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.37351778656126483, - "step": 126, - "train_accuracy_first_token_arguments": 0.8, - "train_accuracy_first_token_arguments_total": 15, + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.7647058823529411, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 0.6666666666666666, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 0.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_rent": 1.0, + "train_accuracy_recipient_token_rent_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.33201581027667987, + "step": 168, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_count": 1.0, - "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 6, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { - "epoch": 0.37648221343873517, - "grad_norm": 2.3478599036765084, - "learning_rate": 3.5940026902326825e-06, - "loss": 0.2732, - "step": 127 - }, - { - "epoch": 0.37648221343873517, - "step": 127, - "train_accuracy_first_token_arguments": 0.782608695652174, - "train_accuracy_first_token_arguments_total": 69, + "epoch": 0.33201581027667987, + "step": 168, "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 1, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 4, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { - "epoch": 0.37648221343873517, - "step": 127, - "train_accuracy_first_token_arguments": 0.7631578947368421, - "train_accuracy_first_token_arguments_total": 38, - "train_accuracy_recipient_token": 0.7222222222222222, - "train_accuracy_recipient_token_all": 0.7, + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compare": 0.5, - "train_accuracy_recipient_token_compare_total": 2, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 1, - "train_accuracy_recipient_token_find": 0.6666666666666666, - "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 18 }, { - "epoch": 0.37648221343873517, - "step": 127, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 17, - "train_accuracy_recipient_token": 0.8947368421052632, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 0.8888888888888888, - "train_accuracy_recipient_token_get_total": 9, - "train_total_number_recipient_token": 19 + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.7777777777777778, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 5, + "train_accuracy_recipient_token_f": 0.0, + "train_accuracy_recipient_token_f_total": 1, + "train_accuracy_recipient_token_pest": 0.0, + "train_accuracy_recipient_token_pest_total": 1, + "train_accuracy_recipient_token_plant": 1.0, + "train_accuracy_recipient_token_plant_total": 1, + "train_accuracy_recipient_token_sun": 1.0, + "train_accuracy_recipient_token_sun_total": 1, + "train_total_number_recipient_token": 9 }, { - "epoch": 0.3794466403162055, - "grad_norm": 1.8233289275067917, - "learning_rate": 3.5722894707616417e-06, - "loss": 0.2708, - "step": 128 + "epoch": 0.33201581027667987, + "eval_loss": 0.218434140086174, + "eval_runtime": 425.3868, + "eval_samples_per_second": 0.237, + "eval_steps_per_second": 0.12, + "step": 168 }, { - "epoch": 0.3794466403162055, - "step": 128, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_dis": 1.0, + "train_accuracy_recipient_token_dis_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 9, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_up": 1.0, + "train_accuracy_recipient_token_up_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.3794466403162055, - "step": 128, - "train_accuracy_first_token_arguments": 0.8095238095238095, - "train_accuracy_first_token_arguments_total": 21, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_clean": 1.0, - "train_accuracy_recipient_token_clean_total": 1, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 3, - "train_accuracy_recipient_token_repair": 1.0, - "train_accuracy_recipient_token_repair_total": 2, - "train_accuracy_recipient_token_restore": 1.0, - "train_accuracy_recipient_token_restore_total": 2, - "train_total_number_recipient_token": 18 - }, - { - "epoch": 0.3794466403162055, - "step": 128, - "train_accuracy_first_token_arguments": 0.9722222222222222, - "train_accuracy_first_token_arguments_total": 36, - "train_accuracy_recipient_token": 0.8947368421052632, + "epoch": 0.33201581027667987, + "step": 168, + "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_book": 0.5, - "train_accuracy_recipient_token_book_total": 2, - "train_accuracy_recipient_token_check": 1.0, - "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_send": 1.0, - "train_accuracy_recipient_token_send_total": 1, - "train_total_number_recipient_token": 19 + "train_total_number_recipient_token": 17 }, { - "epoch": 0.3824110671936759, - "grad_norm": 2.7033960410551843, - "learning_rate": 3.550476671037505e-06, - "loss": 0.3598, - "step": 129 + "epoch": 0.3339920948616601, + "grad_norm": 2.3704926233947448, + "learning_rate": 3.890591041191162e-06, + "loss": 0.2582, + "step": 169 }, { - "epoch": 0.3824110671936759, - "step": 129, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 22, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.3339920948616601, + "step": 169, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.875, "train_accuracy_recipient_token_get_total": 8, - "train_total_number_recipient_token": 18 - }, - { - "epoch": 0.3824110671936759, - "step": 129, - "train_accuracy_first_token_arguments": 0.803921568627451, - "train_accuracy_first_token_arguments_total": 51, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 1, + "train_accuracy_recipient_token_verify": 1.0, + "train_accuracy_recipient_token_verify_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.3824110671936759, - "step": 129, - "train_accuracy_first_token_arguments": 0.7142857142857143, - "train_accuracy_first_token_arguments_total": 7, - "train_accuracy_recipient_token": 0.8823529411764706, - "train_accuracy_recipient_token_all": 0.8, + "epoch": 0.3339920948616601, + "step": 169, + "train_accuracy_recipient_token": 0.8, + "train_accuracy_recipient_token_add": 1.0, + "train_accuracy_recipient_token_add_total": 1, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_arch": 1.0, - "train_accuracy_recipient_token_arch_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 6, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_change": 1.0, + "train_accuracy_recipient_token_change_total": 1, + "train_accuracy_recipient_token_get": 0.4, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_replace": 0.5, + "train_accuracy_recipient_token_replace_total": 2, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.38537549407114624, - "grad_norm": 1.817103444485219, - "learning_rate": 3.528566316748462e-06, - "loss": 0.2691, - "step": 130 + "epoch": 0.3359683794466403, + "grad_norm": 2.4827135380795813, + "learning_rate": 3.8772424536302565e-06, + "loss": 0.227, + "step": 170 }, { - "epoch": 0.38537549407114624, - "step": 130, - "train_accuracy_first_token_arguments": 0.875, - "train_accuracy_first_token_arguments_total": 8, + "epoch": 0.3359683794466403, + "step": 170, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 17 - }, - { - "epoch": 0.38537549407114624, - "step": 130, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 18, - "train_accuracy_recipient_token": 0.8235294117647058, - "train_accuracy_recipient_token_all": 0.7, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_log": 1.0, - "train_accuracy_recipient_token_log_total": 4, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 2, + "train_accuracy_recipient_token_bird": 1.0, + "train_accuracy_recipient_token_bird_total": 4, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 17 }, { - "epoch": 0.38537549407114624, - "step": 130, - "train_accuracy_first_token_arguments": 0.9411764705882353, - "train_accuracy_first_token_arguments_total": 17, - "train_accuracy_recipient_token": 0.8421052631578947, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.3359683794466403, + "step": 170, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_book": 0.5, - "train_accuracy_recipient_token_book_total": 2, - "train_accuracy_recipient_token_find": 0.75, - "train_accuracy_recipient_token_find_total": 4, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_get": 0.7777777777777778, + "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { - "epoch": 0.3883399209486166, - "grad_norm": 2.3178394841243724, - "learning_rate": 3.5065604426422995e-06, - "loss": 0.2196, - "step": 131 - }, - { - "epoch": 0.3883399209486166, - "step": 131, - "train_accuracy_first_token_arguments": 0.9166666666666666, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_check": 1.0, - "train_accuracy_recipient_token_check_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 8, - "train_total_number_recipient_token": 20 + "epoch": 0.33794466403162055, + "grad_norm": 1.626312276221932, + "learning_rate": 3.863837253026372e-06, + "loss": 0.2166, + "step": 171 }, { - "epoch": 0.3883399209486166, - "step": 131, - "train_accuracy_first_token_arguments": 0.7580645161290323, - "train_accuracy_first_token_arguments_total": 62, - "train_accuracy_recipient_token": 0.8421052631578947, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 0.3333333333333333, - "train_accuracy_recipient_token_analyze_total": 3, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_log": 1.0, - "train_accuracy_recipient_token_log_total": 1, + "epoch": 0.33794466403162055, + "step": 171, + "train_accuracy_recipient_token": 0.7894736842105263, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_arch": 0.0, + "train_accuracy_recipient_token_arch_total": 1, + "train_accuracy_recipient_token_building": 1.0, + "train_accuracy_recipient_token_building_total": 1, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_histor": 0.0, + "train_accuracy_recipient_token_histor_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.3883399209486166, - "step": 131, - "train_accuracy_first_token_arguments": 0.7368421052631579, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.9444444444444444, + "epoch": 0.33794466403162055, + "step": 171, + "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, - "train_accuracy_recipient_token_list": 1.0, - "train_accuracy_recipient_token_list_total": 1, - "train_accuracy_recipient_token_recommend": 0.5, - "train_accuracy_recipient_token_recommend_total": 2, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_ident": 0.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_im": 1.0, + "train_accuracy_recipient_token_im_total": 1, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.391304347826087, - "grad_norm": 1.7888078327053374, - "learning_rate": 3.484461092337434e-06, - "loss": 0.2452, - "step": 132 + "epoch": 0.33992094861660077, + "grad_norm": 1.5592305643249844, + "learning_rate": 3.8503759904148005e-06, + "loss": 0.183, + "step": 172 }, { - "epoch": 0.391304347826087, - "step": 132, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 13, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, + "epoch": 0.33992094861660077, + "step": 172, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9166666666666666, + "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_trans": 1.0, - "train_accuracy_recipient_token_trans_total": 3, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.391304347826087, - "step": 132, - "train_accuracy_first_token_arguments": 0.5897435897435898, - "train_accuracy_first_token_arguments_total": 39, + "epoch": 0.33992094861660077, + "step": 172, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 0.6666666666666666, - "train_accuracy_recipient_token_analyze_total": 3, - "train_accuracy_recipient_token_c": 1.0, - "train_accuracy_recipient_token_c_total": 1, - "train_accuracy_recipient_token_histor": 1.0, - "train_accuracy_recipient_token_histor_total": 3, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_total_number_recipient_token": 19 - }, - { - "epoch": 0.391304347826087, - "step": 132, - "train_accuracy_first_token_arguments": 0.8947368421052632, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.8571428571428571, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 0.7777777777777778, - "train_accuracy_recipient_token_get_total": 9, - "train_accuracy_recipient_token_order": 0.0, - "train_accuracy_recipient_token_order_total": 1, - "train_total_number_recipient_token": 21 + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 0.8, + "train_accuracy_recipient_token_find_total": 5, + "train_accuracy_recipient_token_organ": 1.0, + "train_accuracy_recipient_token_organ_total": 1, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.3942687747035573, - "grad_norm": 2.3453355040982307, - "learning_rate": 3.462270318133136e-06, - "loss": 0.273, - "step": 133 + "epoch": 0.34189723320158105, + "grad_norm": 2.1789375117907976, + "learning_rate": 3.8368592191353246e-06, + "loss": 0.256, + "step": 173 }, { - "epoch": 0.3942687747035573, - "step": 133, - "train_accuracy_first_token_arguments": 0.75, - "train_accuracy_first_token_arguments_total": 12, + "epoch": 0.34189723320158105, + "step": 173, "train_accuracy_recipient_token": 0.9411764705882353, - "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_identify": 0.5, - "train_accuracy_recipient_token_identify_total": 2, - "train_accuracy_recipient_token_translate": 1.0, - "train_accuracy_recipient_token_translate_total": 1, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 17 }, { - "epoch": 0.3942687747035573, - "step": 133, - "train_accuracy_first_token_arguments": 0.7894736842105263, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_add": 1.0, - "train_accuracy_recipient_token_add_total": 4, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_construct": 1.0, - "train_accuracy_recipient_token_construct_total": 1, - "train_accuracy_recipient_token_remove": 1.0, - "train_accuracy_recipient_token_remove_total": 1, - "train_accuracy_recipient_token_select": 1.0, - "train_accuracy_recipient_token_select_total": 2, - "train_total_number_recipient_token": 19 - }, - { - "epoch": 0.3942687747035573, - "step": 133, - "train_accuracy_first_token_arguments": 0.8333333333333334, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.8421052631578947, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 4, - "train_accuracy_recipient_token_get": 0.5, - "train_accuracy_recipient_token_get_total": 4, - "train_total_number_recipient_token": 19 - }, - { - "epoch": 0.39723320158102765, - "grad_norm": 2.500149372884034, - "learning_rate": 3.4399901808189327e-06, - "loss": 0.2709, - "step": 134 - }, - { - "epoch": 0.39723320158102765, - "step": 134, - "train_accuracy_first_token_arguments": 0.5555555555555556, - "train_accuracy_first_token_arguments_total": 9, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 8, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_total_number_recipient_token": 19 - }, - { - "epoch": 0.39723320158102765, - "step": 134, - "train_accuracy_first_token_arguments": 0.8181818181818182, - "train_accuracy_first_token_arguments_total": 11, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.34189723320158105, + "step": 173, + "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compose": 1.0, - "train_accuracy_recipient_token_compose_total": 1, - "train_accuracy_recipient_token_create": 1.0, - "train_accuracy_recipient_token_create_total": 1, - "train_accuracy_recipient_token_generate": 1.0, - "train_accuracy_recipient_token_generate_total": 2, + "train_accuracy_recipient_token_bookmark": 1.0, + "train_accuracy_recipient_token_bookmark_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_refresh": 1.0, + "train_accuracy_recipient_token_refresh_total": 1, + "train_accuracy_recipient_token_switch": 0.5, + "train_accuracy_recipient_token_switch_total": 2, "train_total_number_recipient_token": 20 }, { - "epoch": 0.39723320158102765, - "step": 134, - "train_accuracy_first_token_arguments": 0.75, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 0.8571428571428571, - "train_accuracy_recipient_token_get_total": 7, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_total_number_recipient_token": 18 - }, - { - "epoch": 0.40019762845849804, - "grad_norm": 2.056275874523913, - "learning_rate": 3.4176227494832305e-06, - "loss": 0.3526, - "step": 135 + "epoch": 0.3438735177865613, + "grad_norm": 2.171729945066953, + "learning_rate": 3.823287494809469e-06, + "loss": 0.2111, + "step": 174 }, { - "epoch": 0.40019762845849804, - "step": 135, - "train_accuracy_first_token_arguments": 0.9230769230769231, - "train_accuracy_first_token_arguments_total": 13, - "train_accuracy_recipient_token": 0.85, + "epoch": 0.3438735177865613, + "step": 174, + "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_get": 0.7777777777777778, - "train_accuracy_recipient_token_get_total": 9, - "train_total_number_recipient_token": 20 - }, - { - "epoch": 0.40019762845849804, - "step": 135, - "train_accuracy_first_token_arguments": 0.8888888888888888, - "train_accuracy_first_token_arguments_total": 18, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_calculate": 1.0, - "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_remove": 1.0, + "train_accuracy_recipient_token_remove_total": 1, "train_total_number_recipient_token": 20 }, { - "epoch": 0.40019762845849804, - "step": 135, - "train_accuracy_first_token_arguments": 0.9090909090909091, - "train_accuracy_first_token_arguments_total": 11, + "epoch": 0.3438735177865613, + "step": 174, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_convert": 1.0, - "train_accuracy_recipient_token_convert_total": 1, - "train_accuracy_recipient_token_get": 0.8333333333333334, - "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_s": 1.0, + "train_accuracy_recipient_token_s_total": 1, "train_total_number_recipient_token": 17 }, { - "epoch": 0.4031620553359684, - "grad_norm": 1.9432341730217852, - "learning_rate": 3.3951701013211665e-06, - "loss": 0.214, - "step": 136 + "epoch": 0.3458498023715415, + "grad_norm": 2.962276041812103, + "learning_rate": 3.8096613753176635e-06, + "loss": 0.3309, + "step": 175 }, { - "epoch": 0.4031620553359684, - "step": 136, - "train_accuracy_first_token_arguments": 0.8, - "train_accuracy_first_token_arguments_total": 15, - "train_accuracy_recipient_token": 0.8421052631578947, + "epoch": 0.3458498023715415, + "step": 175, + "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 0.6666666666666666, - "train_accuracy_recipient_token_find_total": 3, - "train_accuracy_recipient_token_get": 0.8, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_locate": 1.0, - "train_accuracy_recipient_token_locate_total": 1, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_create": 1.0, + "train_accuracy_recipient_token_create_total": 2, + "train_accuracy_recipient_token_design": 1.0, + "train_accuracy_recipient_token_design_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_total_number_recipient_token": 16 }, { - "epoch": 0.4031620553359684, - "step": 136, - "train_accuracy_first_token_arguments": 0.9130434782608695, - "train_accuracy_first_token_arguments_total": 23, - "train_accuracy_recipient_token": 0.95, + "epoch": 0.3458498023715415, + "step": 175, + "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_invest": 1.0, - "train_accuracy_recipient_token_invest_total": 1, - "train_accuracy_recipient_token_tractor": 1.0, - "train_accuracy_recipient_token_tractor_total": 4, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 16 }, { - "epoch": 0.4031620553359684, - "step": 136, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 28, - "train_accuracy_recipient_token": 0.95, - "train_accuracy_recipient_token_all": 0.9166666666666666, - "train_accuracy_recipient_token_all_total": 12, - "train_accuracy_recipient_token_calculate": 1.0, - "train_accuracy_recipient_token_calculate_total": 1, - "train_accuracy_recipient_token_detect": 1.0, - "train_accuracy_recipient_token_detect_total": 1, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 1, + "epoch": 0.34782608695652173, + "grad_norm": 1.9766749939858486, + "learning_rate": 3.7959814207763134e-06, + "loss": 0.3564, + "step": 176 + }, + { + "epoch": 0.34782608695652173, + "step": 176, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_count": 1.0, + "train_accuracy_recipient_token_count_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.4061264822134387, - "grad_norm": 1.6348659932120928, - "learning_rate": 3.3726343214417023e-06, - "loss": 0.1342, - "step": 137 + "epoch": 0.34782608695652173, + "step": 176, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.7272727272727273, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_setup": 1.0, + "train_accuracy_recipient_token_setup_total": 2, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.4061264822134387, - "step": 137, - "train_accuracy_first_token_arguments": 0.8333333333333334, - "train_accuracy_first_token_arguments_total": 24, + "epoch": 0.34980237154150196, + "grad_norm": 3.077796900581793, + "learning_rate": 3.782248193514766e-06, + "loss": 0.2989, + "step": 177 + }, + { + "epoch": 0.34980237154150196, + "step": 177, "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_pay": 0.5, - "train_accuracy_recipient_token_pay_total": 2, + "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { - "epoch": 0.4061264822134387, - "step": 137, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.8235294117647058, - "train_accuracy_recipient_token_all": 0.7, + "epoch": 0.34980237154150196, + "step": 177, + "train_accuracy_recipient_token": 0.7647058823529411, + "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, - "train_accuracy_recipient_token_make": 1.0, - "train_accuracy_recipient_token_make_total": 4, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_project": 0.0, + "train_accuracy_recipient_token_project_total": 1, + "train_accuracy_recipient_token_sc": 0.5, + "train_accuracy_recipient_token_sc_total": 2, + "train_accuracy_recipient_token_verify": 1.0, + "train_accuracy_recipient_token_verify_total": 1, "train_total_number_recipient_token": 17 }, { - "epoch": 0.4061264822134387, - "step": 137, - "train_accuracy_first_token_arguments": 0.9230769230769231, - "train_accuracy_first_token_arguments_total": 13, + "epoch": 0.35177865612648224, + "grad_norm": 2.055425382576303, + "learning_rate": 3.7684622580522057e-06, + "loss": 0.3784, + "step": 178 + }, + { + "epoch": 0.35177865612648224, + "step": 178, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_count": 1.0, - "train_accuracy_recipient_token_count_total": 1, - "train_accuracy_recipient_token_decode": 1.0, - "train_accuracy_recipient_token_decode_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 6, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 19 }, { - "epoch": 0.4090909090909091, - "grad_norm": 2.0536484514852344, - "learning_rate": 3.3500175026739916e-06, - "loss": 0.279, - "step": 138 - }, - { - "epoch": 0.4090909090909091, - "step": 138, - "train_accuracy_first_token_arguments": 0.7142857142857143, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.35177865612648224, + "step": 178, + "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_bird": 1.0, - "train_accuracy_recipient_token_bird_total": 4, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_in": 1.0, - "train_accuracy_recipient_token_in_total": 3, - "train_accuracy_recipient_token_validate": 1.0, - "train_accuracy_recipient_token_validate_total": 1, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_get": 0.9, + "train_accuracy_recipient_token_get_total": 10, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.4090909090909091, - "step": 138, - "train_accuracy_first_token_arguments": 0.8048780487804879, - "train_accuracy_first_token_arguments_total": 41, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.35375494071146246, + "grad_norm": 1.638839258863397, + "learning_rate": 3.7546241810744444e-06, + "loss": 0.1502, + "step": 179 + }, + { + "epoch": 0.35375494071146246, + "step": 179, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_fin": 0.75, + "train_accuracy_recipient_token_fin_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_invest": 1.0, + "train_accuracy_recipient_token_invest_total": 1, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.4090909090909091, - "step": 138, - "train_accuracy_first_token_arguments": 0.9411764705882353, - "train_accuracy_first_token_arguments_total": 17, - "train_accuracy_recipient_token": 0.9411764705882353, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 1, + "epoch": 0.35375494071146246, + "step": 179, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_total_number_recipient_token": 17 - }, - { - "epoch": 0.41205533596837945, - "grad_norm": 2.2289267853293144, - "learning_rate": 3.327321745373021e-06, - "loss": 0.3159, - "step": 139 + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.41205533596837945, - "step": 139, - "train_accuracy_first_token_arguments": 0.9333333333333333, - "train_accuracy_first_token_arguments_total": 15, - "train_accuracy_recipient_token": 0.8333333333333334, - "train_accuracy_recipient_token_all": 0.7272727272727273, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_pest": 1.0, - "train_accuracy_recipient_token_pest_total": 1, - "train_accuracy_recipient_token_sun": 1.0, - "train_accuracy_recipient_token_sun_total": 1, - "train_total_number_recipient_token": 18 + "epoch": 0.3557312252964427, + "grad_norm": 1.6977698069216782, + "learning_rate": 3.740734531410626e-06, + "loss": 0.1859, + "step": 180 }, { - "epoch": 0.41205533596837945, - "step": 139, - "train_accuracy_first_token_arguments": 0.8888888888888888, - "train_accuracy_first_token_arguments_total": 18, - "train_accuracy_recipient_token": 0.8823529411764706, - "train_accuracy_recipient_token_all": 0.8181818181818182, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_evaluate": 1.0, - "train_accuracy_recipient_token_evaluate_total": 1, + "epoch": 0.3557312252964427, + "step": 180, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, + "train_accuracy_recipient_token_detect": 1.0, + "train_accuracy_recipient_token_detect_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_provide": 1.0, - "train_accuracy_recipient_token_provide_total": 1, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 17 }, { - "epoch": 0.41205533596837945, - "step": 139, - "train_accuracy_first_token_arguments": 0.8461538461538461, - "train_accuracy_first_token_arguments_total": 26, - "train_accuracy_recipient_token": 0.95, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.3557312252964427, + "step": 180, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_calculate": 1.0, - "train_accuracy_recipient_token_calculate_total": 1, - "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_find": 0.6666666666666666, - "train_accuracy_recipient_token_find_total": 3, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 2, + "train_accuracy_recipient_token_buy": 1.0, + "train_accuracy_recipient_token_buy_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_s": 1.0, + "train_accuracy_recipient_token_s_total": 2, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.4150197628458498, - "grad_norm": 2.1319285672179005, - "learning_rate": 3.304549157224558e-06, - "loss": 0.3213, - "step": 140 + "epoch": 0.3577075098814229, + "grad_norm": 2.104044600552162, + "learning_rate": 3.7267938800098454e-06, + "loss": 0.1879, + "step": 181 }, { - "epoch": 0.4150197628458498, - "step": 140, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 9, + "epoch": 0.3577075098814229, + "step": 181, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 8, - "train_total_number_recipient_token": 18 - }, - { - "epoch": 0.4150197628458498, - "step": 140, - "train_accuracy_first_token_arguments": 0.8333333333333334, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 9, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.4150197628458498, - "step": 140, - "train_accuracy_first_token_arguments": 0.8, - "train_accuracy_first_token_arguments_total": 60, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.3577075098814229, + "step": 181, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_explain": 0.5, - "train_accuracy_recipient_token_explain_total": 2, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, - "train_accuracy_recipient_token_invest": 1.0, - "train_accuracy_recipient_token_invest_total": 1, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.41798418972332013, - "grad_norm": 2.0722211722052575, - "learning_rate": 3.2817018530494164e-06, - "loss": 0.2898, - "step": 141 + "epoch": 0.35968379446640314, + "grad_norm": 2.3821501005519194, + "learning_rate": 3.71280279991768e-06, + "loss": 0.256, + "step": 182 }, { - "epoch": 0.41798418972332013, - "step": 141, - "train_accuracy_first_token_arguments": 0.7368421052631579, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.35968379446640314, + "step": 182, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 0.6666666666666666, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_locate": 1.0, - "train_accuracy_recipient_token_locate_total": 1, - "train_accuracy_recipient_token_p": 1.0, - "train_accuracy_recipient_token_p_total": 3, + "train_accuracy_recipient_token_cl": 1.0, + "train_accuracy_recipient_token_cl_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_im": 1.0, + "train_accuracy_recipient_token_im_total": 1, + "train_accuracy_recipient_token_search": 0.5, + "train_accuracy_recipient_token_search_total": 2, "train_total_number_recipient_token": 18 }, { - "epoch": 0.41798418972332013, - "step": 141, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.9523809523809523, + "epoch": 0.35968379446640314, + "step": 182, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 2, - "train_accuracy_recipient_token_count": 1.0, - "train_accuracy_recipient_token_count_total": 1, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 1, - "train_accuracy_recipient_token_get": 0.8333333333333334, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, - "train_total_number_recipient_token": 21 + "train_accuracy_recipient_token_list": 1.0, + "train_accuracy_recipient_token_list_total": 1, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.41798418972332013, - "step": 141, - "train_accuracy_first_token_arguments": 0.8656716417910447, - "train_accuracy_first_token_arguments_total": 67, + "epoch": 0.3616600790513834, + "grad_norm": 1.7412853263181982, + "learning_rate": 3.698761866252635e-06, + "loss": 0.2009, + "step": 183 + }, + { + "epoch": 0.3616600790513834, + "step": 183, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_count": 1.0, - "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 3, - "train_accuracy_recipient_token_verify": 1.0, - "train_accuracy_recipient_token_verify_total": 2, + "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { - "epoch": 0.4209486166007905, - "grad_norm": 2.1191079314907766, - "learning_rate": 3.2587819546070596e-06, - "loss": 0.255, - "step": 142 - }, - { - "epoch": 0.4209486166007905, - "step": 142, - "train_accuracy_first_token_arguments": 0.7692307692307693, - "train_accuracy_first_token_arguments_total": 13, - "train_accuracy_recipient_token": 0.8888888888888888, + "epoch": 0.3616600790513834, + "step": 183, + "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, + "train_accuracy_recipient_token_art": 1.0, + "train_accuracy_recipient_token_art_total": 3, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 3, - "train_accuracy_recipient_token_get": 0.6666666666666666, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 1, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_find_total": 2, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.4209486166007905, - "step": 142, - "train_accuracy_first_token_arguments": 0.8333333333333334, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.36363636363636365, + "grad_norm": 1.6453660919523958, + "learning_rate": 3.684671656182497e-06, + "loss": 0.1797, + "step": 184 + }, + { + "epoch": 0.36363636363636365, + "step": 184, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_decode": 1.0, + "train_accuracy_recipient_token_decode_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, - "train_accuracy_recipient_token_l": 1.0, - "train_accuracy_recipient_token_l_total": 3, - "train_accuracy_recipient_token_search": 1.0, - "train_accuracy_recipient_token_search_total": 1, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.4209486166007905, - "step": 142, - "train_accuracy_first_token_arguments": 0.9, - "train_accuracy_first_token_arguments_total": 10, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.36363636363636365, + "step": 184, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_event": 0.6666666666666666, - "train_accuracy_recipient_token_event_total": 3, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_arch": 1.0, + "train_accuracy_recipient_token_arch_total": 2, + "train_accuracy_recipient_token_build": 1.0, + "train_accuracy_recipient_token_build_total": 3, + "train_accuracy_recipient_token_histor": 0.0, + "train_accuracy_recipient_token_histor_total": 1, + "train_accuracy_recipient_token_s": 1.0, + "train_accuracy_recipient_token_s_total": 1, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.42391304347826086, - "grad_norm": 2.5497967401591204, - "learning_rate": 3.2357915903985605e-06, - "loss": 0.3137, - "step": 143 + "epoch": 0.36561264822134387, + "grad_norm": 1.7214718462583107, + "learning_rate": 3.670532748900615e-06, + "loss": 0.251, + "step": 185 }, { - "epoch": 0.42391304347826086, - "step": 143, - "train_accuracy_first_token_arguments": 0.95, - "train_accuracy_first_token_arguments_total": 40, - "train_accuracy_recipient_token": 0.95, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_clean": 1.0, - "train_accuracy_recipient_token_clean_total": 1, + "epoch": 0.36561264822134387, + "step": 185, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9166666666666666, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 3, - "train_accuracy_recipient_token_paint": 1.0, - "train_accuracy_recipient_token_paint_total": 2, - "train_accuracy_recipient_token_remove": 1.0, - "train_accuracy_recipient_token_remove_total": 1, - "train_total_number_recipient_token": 20 - }, - { - "epoch": 0.42391304347826086, - "step": 143, - "train_accuracy_first_token_arguments": 0.7619047619047619, - "train_accuracy_first_token_arguments_total": 21, - "train_accuracy_recipient_token": 0.8, - "train_accuracy_recipient_token_all": 0.7, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_decode": 1.0, - "train_accuracy_recipient_token_decode_total": 2, - "train_accuracy_recipient_token_device": 1.0, - "train_accuracy_recipient_token_device_total": 3, - "train_accuracy_recipient_token_extract": 0.0, - "train_accuracy_recipient_token_extract_total": 1, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 2, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_find_total": 5, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.42391304347826086, - "step": 143, - "train_accuracy_first_token_arguments": 0.7924528301886793, - "train_accuracy_first_token_arguments_total": 53, - "train_accuracy_recipient_token": 0.95, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 1, + "epoch": 0.36561264822134387, + "step": 185, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 6, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_order": 1.0, - "train_accuracy_recipient_token_order_total": 1, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_land": 1.0, + "train_accuracy_recipient_token_land_total": 4, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.4268774703557312, - "grad_norm": 1.9145256420639876, - "learning_rate": 3.2127328954689307e-06, - "loss": 0.192, - "step": 144 + "epoch": 0.3675889328063241, + "grad_norm": 1.9383253652427745, + "learning_rate": 3.656345725602089e-06, + "loss": 0.213, + "step": 186 }, { - "epoch": 0.4268774703557312, - "step": 144, - "train_accuracy_first_token_arguments": 0.7391304347826086, - "train_accuracy_first_token_arguments_total": 46, - "train_accuracy_recipient_token": 0.7894736842105263, - "train_accuracy_recipient_token_all": 0.8, + "epoch": 0.3675889328063241, + "step": 186, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 0.6666666666666666, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_schedule": 1.0, - "train_accuracy_recipient_token_schedule_total": 1, - "train_accuracy_recipient_token_translate": 0.75, - "train_accuracy_recipient_token_translate_total": 4, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_generate": 1.0, + "train_accuracy_recipient_token_generate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 16 }, { - "epoch": 0.4268774703557312, - "step": 144, - "train_accuracy_first_token_arguments": 0.875, - "train_accuracy_first_token_arguments_total": 104, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_calculate": 1.0, - "train_accuracy_recipient_token_calculate_total": 1, + "epoch": 0.3675889328063241, + "step": 186, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, - "train_accuracy_recipient_token_optimize": 1.0, - "train_accuracy_recipient_token_optimize_total": 1, + "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { - "epoch": 0.4268774703557312, - "step": 144, - "train_accuracy_first_token_arguments": 0.9090909090909091, - "train_accuracy_first_token_arguments_total": 22, + "epoch": 0.3695652173913043, + "grad_norm": 2.5997132280472526, + "learning_rate": 3.642111169459879e-06, + "loss": 0.3241, + "step": 187 + }, + { + "epoch": 0.3695652173913043, + "step": 187, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 8, - "train_total_number_recipient_token": 18 - }, - { - "epoch": 0.4298418972332016, - "grad_norm": 1.6421584423874946, - "learning_rate": 3.1896080112088477e-06, - "loss": 0.2441, - "step": 145 + "train_accuracy_recipient_token_evaluate": 1.0, + "train_accuracy_recipient_token_evaluate_total": 1, + "train_accuracy_recipient_token_phrase": 1.0, + "train_accuracy_recipient_token_phrase_total": 4, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 1, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.4298418972332016, - "step": 145, - "train_accuracy_first_token_arguments": 0.8846153846153846, - "train_accuracy_first_token_arguments_total": 26, - "train_accuracy_recipient_token": 0.8823529411764706, + "epoch": 0.3695652173913043, + "step": 187, + "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_deploy": 1.0, - "train_accuracy_recipient_token_deploy_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_create": 1.0, + "train_accuracy_recipient_token_create_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_plan": 0.0, + "train_accuracy_recipient_token_plan_total": 1, + "train_accuracy_recipient_token_s": 1.0, + "train_accuracy_recipient_token_s_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.4298418972332016, - "step": 145, - "train_accuracy_first_token_arguments": 0.8947368421052632, - "train_accuracy_first_token_arguments_total": 38, + "epoch": 0.3715415019762846, + "grad_norm": 1.9124092057739746, + "learning_rate": 3.6278296656008366e-06, + "loss": 0.1795, + "step": 188 + }, + { + "epoch": 0.3715415019762846, + "step": 188, "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_car": 1.0, - "train_accuracy_recipient_token_car_total": 1, - "train_accuracy_recipient_token_evaluate": 1.0, - "train_accuracy_recipient_token_evaluate_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_get": 0.5, - "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { - "epoch": 0.4298418972332016, - "step": 145, - "train_accuracy_first_token_arguments": 0.8235294117647058, - "train_accuracy_first_token_arguments_total": 34, - "train_accuracy_recipient_token": 0.8421052631578947, + "epoch": 0.3715415019762846, + "step": 188, + "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_check": 1.0, - "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, + "train_accuracy_recipient_token_compare": 0.0, + "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_order": 1.0, - "train_accuracy_recipient_token_order_total": 1, - "train_accuracy_recipient_token_schedule": 0.0, - "train_accuracy_recipient_token_schedule_total": 1, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_s": 1.0, + "train_accuracy_recipient_token_s_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.43280632411067194, - "grad_norm": 1.9573866581692538, - "learning_rate": 3.166419085155793e-06, - "loss": 0.2593, - "step": 146 + "epoch": 0.37351778656126483, + "grad_norm": 2.4304939026300847, + "learning_rate": 3.613501801081648e-06, + "loss": 0.3143, + "step": 189 }, { - "epoch": 0.43280632411067194, - "step": 146, - "train_accuracy_first_token_arguments": 0.7857142857142857, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 0.9411764705882353, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_list": 1.0, - "train_accuracy_recipient_token_list_total": 1, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 1, - "train_accuracy_recipient_token_stat": 1.0, - "train_accuracy_recipient_token_stat_total": 1, - "train_total_number_recipient_token": 17 + "epoch": 0.37351778656126483, + "step": 189, + "train_accuracy_recipient_token": 0.9333333333333333, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_organ": 1.0, + "train_accuracy_recipient_token_organ_total": 1, + "train_total_number_recipient_token": 15 }, { - "epoch": 0.43280632411067194, - "step": 146, - "train_accuracy_first_token_arguments": 0.84, - "train_accuracy_first_token_arguments_total": 25, - "train_accuracy_recipient_token": 0.95, + "epoch": 0.37351778656126483, + "step": 189, + "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 0.9, - "train_accuracy_recipient_token_get_total": 10, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_compare": 0.5, + "train_accuracy_recipient_token_compare_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.43280632411067194, - "step": 146, - "train_accuracy_first_token_arguments": 0.7142857142857143, - "train_accuracy_first_token_arguments_total": 7, - "train_accuracy_recipient_token": 0.9375, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.37549407114624506, + "grad_norm": 1.8544268929291403, + "learning_rate": 3.599128164864706e-06, + "loss": 0.1948, + "step": 190 + }, + { + "epoch": 0.37549407114624506, + "step": 190, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_count": 1.0, + "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_total_number_recipient_token": 16 + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.4357707509881423, - "grad_norm": 2.277891831921691, - "learning_rate": 3.143168270794612e-06, - "loss": 0.4509, - "step": 147 + "epoch": 0.37549407114624506, + "step": 190, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.4357707509881423, - "step": 147, - "train_accuracy_first_token_arguments": 0.8888888888888888, - "train_accuracy_first_token_arguments_total": 18, - "train_accuracy_recipient_token": 0.8823529411764706, + "epoch": 0.3774703557312253, + "grad_norm": 2.7234941325542286, + "learning_rate": 3.5847093477938955e-06, + "loss": 0.3302, + "step": 191 + }, + { + "epoch": 0.3774703557312253, + "step": 191, + "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_explain": 1.0, - "train_accuracy_recipient_token_explain_total": 1, + "train_accuracy_recipient_token_compare": 0.5, + "train_accuracy_recipient_token_compare_total": 2, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_traffic": 1.0, - "train_accuracy_recipient_token_traffic_total": 2, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_get_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.4357707509881423, - "step": 147, - "train_accuracy_first_token_arguments": 0.85, - "train_accuracy_first_token_arguments_total": 20, - "train_accuracy_recipient_token": 0.9047619047619048, + "epoch": 0.3774703557312253, + "step": 191, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.8888888888888888, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.3794466403162055, + "grad_norm": 1.7644071316796994, + "learning_rate": 3.5702459425703146e-06, + "loss": 0.198, + "step": 192 + }, + { + "epoch": 0.3794466403162055, + "step": 192, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_calculate": 0.6666666666666666, - "train_accuracy_recipient_token_calculate_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.3794466403162055, + "step": 192, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_clean": 1.0, + "train_accuracy_recipient_token_clean_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 3, + "train_accuracy_recipient_token_rep": 0.5, + "train_accuracy_recipient_token_rep_total": 2, + "train_accuracy_recipient_token_restore": 1.0, + "train_accuracy_recipient_token_restore_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.3814229249011858, + "grad_norm": 2.610658057678565, + "learning_rate": 3.5557385437279e-06, + "loss": 0.3898, + "step": 193 + }, + { + "epoch": 0.3814229249011858, + "step": 193, + "train_accuracy_recipient_token": 0.7894736842105263, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_book": 0.0, + "train_accuracy_recipient_token_book_total": 2, "train_accuracy_recipient_token_check": 1.0, - "train_accuracy_recipient_token_check_total": 2, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_send": 1.0, + "train_accuracy_recipient_token_send_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.3814229249011858, + "step": 193, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.383399209486166, + "grad_norm": 2.4079072059937383, + "learning_rate": 3.541187747608998e-06, + "loss": 0.1893, + "step": 194 + }, + { + "epoch": 0.383399209486166, + "step": 194, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 21 + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.4357707509881423, - "step": 147, - "train_accuracy_first_token_arguments": 0.9125, - "train_accuracy_first_token_arguments_total": 80, - "train_accuracy_recipient_token": 0.7777777777777778, + "epoch": 0.383399209486166, + "step": 194, + "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_arr": 1.0, - "train_accuracy_recipient_token_arr_total": 1, - "train_accuracy_recipient_token_book": 0.5, - "train_accuracy_recipient_token_book_total": 2, - "train_accuracy_recipient_token_hire": 1.0, - "train_accuracy_recipient_token_hire_total": 1, - "train_accuracy_recipient_token_plan": 1.0, - "train_accuracy_recipient_token_plan_total": 1, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_arch": 1.0, + "train_accuracy_recipient_token_arch_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.43873517786561267, - "grad_norm": 2.085474389214059, - "learning_rate": 3.119857727357527e-06, - "loss": 0.2157, - "step": 148 + "epoch": 0.38537549407114624, + "grad_norm": 1.9829071189672327, + "learning_rate": 3.5265941523398455e-06, + "loss": 0.2634, + "step": 195 }, { - "epoch": 0.43873517786561267, - "step": 148, - "train_accuracy_first_token_arguments": 0.782608695652174, - "train_accuracy_first_token_arguments_total": 23, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.38537549407114624, + "step": 195, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_calculate": 0.5, - "train_accuracy_recipient_token_calculate_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 19 + "train_total_number_recipient_token": 17 }, { - "epoch": 0.43873517786561267, - "step": 148, - "train_accuracy_first_token_arguments": 0.75, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.38537549407114624, + "step": 195, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 0.8333333333333334, - "train_accuracy_recipient_token_get_total": 6, - "train_accuracy_recipient_token_stamp": 1.0, - "train_accuracy_recipient_token_stamp_total": 2, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_log": 1.0, + "train_accuracy_recipient_token_log_total": 4, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.43873517786561267, - "step": 148, - "train_accuracy_first_token_arguments": 0.9090909090909091, - "train_accuracy_first_token_arguments_total": 11, - "train_accuracy_recipient_token": 0.8823529411764706, + "epoch": 0.38735177865612647, + "grad_norm": 2.237176244877257, + "learning_rate": 3.5119583578059845e-06, + "loss": 0.2343, + "step": 196 + }, + { + "epoch": 0.38735177865612647, + "step": 196, + "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_display": 1.0, - "train_accuracy_recipient_token_display_total": 1, + "train_accuracy_recipient_token_book": 0.5, + "train_accuracy_recipient_token_book_total": 2, "train_accuracy_recipient_token_find": 0.75, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.38735177865612647, + "step": 196, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.3893280632411067, + "grad_norm": 1.5354145070510907, + "learning_rate": 3.4972809656276047e-06, + "loss": 0.1565, + "step": 197 + }, + { + "epoch": 0.3893280632411067, + "step": 197, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analy": 0.3333333333333333, + "train_accuracy_recipient_token_analy_total": 3, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_log": 1.0, + "train_accuracy_recipient_token_log_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.3893280632411067, + "step": 197, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_list": 1.0, + "train_accuracy_recipient_token_list_total": 1, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.391304347826087, + "grad_norm": 1.8679250394533993, + "learning_rate": 3.4825625791348093e-06, + "loss": 0.2068, + "step": 198 + }, + { + "epoch": 0.391304347826087, + "step": 198, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_trans": 1.0, + "train_accuracy_recipient_token_trans_total": 3, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.391304347826087, + "step": 198, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 3, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_histor": 1.0, + "train_accuracy_recipient_token_histor_total": 3, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.3932806324110672, + "grad_norm": 2.1127643306550565, + "learning_rate": 3.467803803342821e-06, + "loss": 0.2574, + "step": 199 + }, + { + "epoch": 0.3932806324110672, + "step": 199, + "train_accuracy_recipient_token": 0.8571428571428571, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.7777777777777778, + "train_accuracy_recipient_token_get_total": 9, + "train_accuracy_recipient_token_order": 0.0, + "train_accuracy_recipient_token_order_total": 1, + "train_total_number_recipient_token": 21 + }, + { + "epoch": 0.3932806324110672, + "step": 199, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_ident": 0.0, + "train_accuracy_recipient_token_ident_total": 2, + "train_accuracy_recipient_token_translate": 0.0, + "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 17 }, { - "epoch": 0.441699604743083, - "grad_norm": 1.8828482742789217, - "learning_rate": 3.0964896196236217e-06, - "loss": 0.2592, - "step": 149 + "epoch": 0.3952569169960474, + "grad_norm": 2.750136117135516, + "learning_rate": 3.4530052449271044e-06, + "loss": 0.2443, + "step": 200 + }, + { + "epoch": 0.3952569169960474, + "step": 200, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_add": 1.0, + "train_accuracy_recipient_token_add_total": 4, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_construct": 1.0, + "train_accuracy_recipient_token_construct_total": 1, + "train_accuracy_recipient_token_remove": 1.0, + "train_accuracy_recipient_token_remove_total": 1, + "train_accuracy_recipient_token_select": 1.0, + "train_accuracy_recipient_token_select_total": 2, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.3952569169960474, + "step": 200, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 0.5, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.39723320158102765, + "grad_norm": 2.4254009242612904, + "learning_rate": 3.438167512198436e-06, + "loss": 0.2524, + "step": 201 + }, + { + "epoch": 0.39723320158102765, + "step": 201, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.39723320158102765, + "step": 201, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compose": 1.0, + "train_accuracy_recipient_token_compose_total": 1, + "train_accuracy_recipient_token_create": 1.0, + "train_accuracy_recipient_token_create_total": 1, + "train_accuracy_recipient_token_generate": 1.0, + "train_accuracy_recipient_token_generate_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.39920948616600793, + "grad_norm": 1.9697723841835133, + "learning_rate": 3.4232912150778914e-06, + "loss": 0.292, + "step": 202 + }, + { + "epoch": 0.39920948616600793, + "step": 202, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.8571428571428571, + "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.39920948616600793, + "step": 202, + "train_accuracy_recipient_token": 0.85, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 0.7777777777777778, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.40118577075098816, + "grad_norm": 1.9505870591219885, + "learning_rate": 3.408376965071779e-06, + "loss": 0.2301, + "step": 203 + }, + { + "epoch": 0.40118577075098816, + "step": 203, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.40118577075098816, + "step": 203, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_convert": 1.0, + "train_accuracy_recipient_token_convert_total": 1, + "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.4031620553359684, + "grad_norm": 2.259780090042466, + "learning_rate": 3.393425375246503e-06, + "loss": 0.2284, + "step": 204 + }, + { + "epoch": 0.4031620553359684, + "step": 204, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 0.6, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_locate": 1.0, + "train_accuracy_recipient_token_locate_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.4031620553359684, + "step": 204, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.5, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_invest": 1.0, + "train_accuracy_recipient_token_invest_total": 1, + "train_accuracy_recipient_token_tractor": 1.0, + "train_accuracy_recipient_token_tractor_total": 4, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.4051383399209486, + "grad_norm": 1.654459204817194, + "learning_rate": 3.3784370602033572e-06, + "loss": 0.1279, + "step": 205 + }, + { + "epoch": 0.4051383399209486, + "step": 205, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 0.9166666666666666, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_detect": 1.0, + "train_accuracy_recipient_token_detect_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.4051383399209486, + "step": 205, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_pay": 0.5, + "train_accuracy_recipient_token_pay_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.40711462450592883, + "grad_norm": 1.7599338692834208, + "learning_rate": 3.3634126360532694e-06, + "loss": 0.1629, + "step": 206 + }, + { + "epoch": 0.40711462450592883, + "step": 206, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_make": 1.0, + "train_accuracy_recipient_token_make_total": 4, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.40711462450592883, + "step": 206, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_count": 1.0, + "train_accuracy_recipient_token_count_total": 1, + "train_accuracy_recipient_token_decode": 1.0, + "train_accuracy_recipient_token_decode_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.4090909090909091, + "grad_norm": 1.881346517215222, + "learning_rate": 3.3483527203914694e-06, + "loss": 0.2584, + "step": 207 + }, + { + "epoch": 0.4090909090909091, + "step": 207, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_bird": 1.0, + "train_accuracy_recipient_token_bird_total": 4, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_in": 1.0, + "train_accuracy_recipient_token_in_total": 3, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.4090909090909091, + "step": 207, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.41106719367588934, + "grad_norm": 2.0484839728339317, + "learning_rate": 3.333257932272105e-06, + "loss": 0.2313, + "step": 208 + }, + { + "epoch": 0.41106719367588934, + "step": 208, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.41106719367588934, + "step": 208, + "train_accuracy_recipient_token": 0.7777777777777778, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_pest": 0.0, + "train_accuracy_recipient_token_pest_total": 1, + "train_accuracy_recipient_token_sun": 1.0, + "train_accuracy_recipient_token_sun_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.41304347826086957, + "grad_norm": 2.473490384710613, + "learning_rate": 3.3181288921827925e-06, + "loss": 0.439, + "step": 209 + }, + { + "epoch": 0.41304347826086957, + "step": 209, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_evaluate": 1.0, + "train_accuracy_recipient_token_evaluate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_provide": 1.0, + "train_accuracy_recipient_token_provide_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.41304347826086957, + "step": 209, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 0.6666666666666666, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.4150197628458498, + "grad_norm": 1.8750180775014957, + "learning_rate": 3.3029662220191146e-06, + "loss": 0.1752, + "step": 210 + }, + { + "epoch": 0.4150197628458498, + "step": 210, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.4150197628458498, + "step": 210, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.41699604743083, + "grad_norm": 2.2988992335970173, + "learning_rate": 3.2877705450590525e-06, + "loss": 0.2718, + "step": 211 + }, + { + "epoch": 0.41699604743083, + "step": 211, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_ex": 1.0, + "train_accuracy_recipient_token_ex_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_invest": 1.0, + "train_accuracy_recipient_token_invest_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.41699604743083, + "step": 211, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_locate": 1.0, + "train_accuracy_recipient_token_locate_total": 1, + "train_accuracy_recipient_token_p": 1.0, + "train_accuracy_recipient_token_p_total": 3, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.4189723320158103, + "grad_norm": 2.0126487070423162, + "learning_rate": 3.272542485937369e-06, + "loss": 0.2419, + "step": 212 + }, + { + "epoch": 0.4189723320158103, + "step": 212, + "train_accuracy_recipient_token": 0.9523809523809523, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 2, + "train_accuracy_recipient_token_count": 1.0, + "train_accuracy_recipient_token_count_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_get": 0.8333333333333334, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 21 + }, + { + "epoch": 0.4189723320158103, + "step": 212, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_count": 1.0, + "train_accuracy_recipient_token_count_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 3, + "train_accuracy_recipient_token_verify": 1.0, + "train_accuracy_recipient_token_verify_total": 2, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.4209486166007905, + "grad_norm": 2.030818509388425, + "learning_rate": 3.2572826706199304e-06, + "loss": 0.2074, + "step": 213 + }, + { + "epoch": 0.4209486166007905, + "step": 213, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.4209486166007905, + "step": 213, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_l": 1.0, + "train_accuracy_recipient_token_l_total": 3, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.42292490118577075, + "grad_norm": 1.9072544731043124, + "learning_rate": 3.2419917263779765e-06, + "loss": 0.235, + "step": 214 + }, + { + "epoch": 0.42292490118577075, + "step": 214, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_event": 0.6666666666666666, + "train_accuracy_recipient_token_event_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.42292490118577075, + "step": 214, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_clean": 1.0, + "train_accuracy_recipient_token_clean_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_paint": 0.5, + "train_accuracy_recipient_token_paint_total": 2, + "train_accuracy_recipient_token_remove": 1.0, + "train_accuracy_recipient_token_remove_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.424901185770751, + "grad_norm": 2.623452228952221, + "learning_rate": 3.2266702817623348e-06, + "loss": 0.2415, + "step": 215 + }, + { + "epoch": 0.424901185770751, + "step": 215, + "train_accuracy_recipient_token": 0.8, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_decode": 1.0, + "train_accuracy_recipient_token_decode_total": 2, + "train_accuracy_recipient_token_device": 1.0, + "train_accuracy_recipient_token_device_total": 3, + "train_accuracy_recipient_token_extract": 1.0, + "train_accuracy_recipient_token_extract_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 2, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.424901185770751, + "step": 215, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_order": 1.0, + "train_accuracy_recipient_token_order_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.4268774703557312, + "grad_norm": 2.303051674442213, + "learning_rate": 3.2113189665775812e-06, + "loss": 0.2183, + "step": 216 + }, + { + "epoch": 0.4268774703557312, + "step": 216, + "train_accuracy_recipient_token": 0.7894736842105263, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_accuracy_recipient_token_translate": 0.75, + "train_accuracy_recipient_token_translate_total": 4, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.4268774703557312, + "step": 216, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_opt": 1.0, + "train_accuracy_recipient_token_opt_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.4288537549407115, + "grad_norm": 1.626960528946486, + "learning_rate": 3.1959384118561587e-06, + "loss": 0.2717, + "step": 217 + }, + { + "epoch": 0.4288537549407115, + "step": 217, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 8, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.4288537549407115, + "step": 217, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_deploy": 1.0, + "train_accuracy_recipient_token_deploy_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.4308300395256917, + "grad_norm": 1.985201264804771, + "learning_rate": 3.180529249832428e-06, + "loss": 0.1797, + "step": 218 + }, + { + "epoch": 0.4308300395256917, + "step": 218, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 9, + "train_accuracy_recipient_token_car": 1.0, + "train_accuracy_recipient_token_car_total": 1, + "train_accuracy_recipient_token_evaluate": 1.0, + "train_accuracy_recipient_token_evaluate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.5, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.4308300395256917, + "step": 218, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_order": 1.0, + "train_accuracy_recipient_token_order_total": 1, + "train_accuracy_recipient_token_schedule": 0.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.43280632411067194, + "grad_norm": 1.8348360216552162, + "learning_rate": 3.165092113916688e-06, + "loss": 0.2101, + "step": 219 + }, + { + "epoch": 0.43280632411067194, + "step": 219, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_list": 1.0, + "train_accuracy_recipient_token_list_total": 1, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, + "train_accuracy_recipient_token_stat": 1.0, + "train_accuracy_recipient_token_stat_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.43280632411067194, + "step": 219, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 10, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.43478260869565216, + "grad_norm": 1.9783954952399618, + "learning_rate": 3.1496276386691327e-06, + "loss": 0.2739, + "step": 220 + }, + { + "epoch": 0.43478260869565216, + "step": 220, + "train_accuracy_recipient_token": 0.875, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.43478260869565216, + "step": 220, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_ex": 1.0, + "train_accuracy_recipient_token_ex_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_traffic": 1.0, + "train_accuracy_recipient_token_traffic_total": 2, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.4367588932806324, + "grad_norm": 2.4700705976410786, + "learning_rate": 3.1341364597737684e-06, + "loss": 0.4093, + "step": 221 + }, + { + "epoch": 0.4367588932806324, + "step": 221, + "train_accuracy_recipient_token": 0.9523809523809523, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 3, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 21 + }, + { + "epoch": 0.4367588932806324, + "step": 221, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_arr": 1.0, + "train_accuracy_recipient_token_arr_total": 1, + "train_accuracy_recipient_token_book": 0.5, + "train_accuracy_recipient_token_book_total": 2, + "train_accuracy_recipient_token_hire": 1.0, + "train_accuracy_recipient_token_hire_total": 1, + "train_accuracy_recipient_token_plan": 1.0, + "train_accuracy_recipient_token_plan_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.43873517786561267, + "grad_norm": 2.0704854982802368, + "learning_rate": 3.1186192140122863e-06, + "loss": 0.2264, + "step": 222 + }, + { + "epoch": 0.43873517786561267, + "step": 222, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.43873517786561267, + "step": 222, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_stamp": 1.0, + "train_accuracy_recipient_token_stamp_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.4407114624505929, + "grad_norm": 1.654154021035547, + "learning_rate": 3.1030765392378813e-06, + "loss": 0.1653, + "step": 223 + }, + { + "epoch": 0.4407114624505929, + "step": 223, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_display": 1.0, + "train_accuracy_recipient_token_display_total": 1, + "train_accuracy_recipient_token_find": 0.75, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.4407114624505929, + "step": 223, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 3, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.4426877470355731, + "grad_norm": 1.8368273428097381, + "learning_rate": 3.0875090743490383e-06, + "loss": 0.2253, + "step": 224 + }, + { + "epoch": 0.4426877470355731, + "step": 224, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_s": 1.0, + "train_accuracy_recipient_token_s_total": 2, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.4426877470355731, + "step": 224, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.44466403162055335, + "grad_norm": 1.950504550895436, + "learning_rate": 3.071917459263264e-06, + "loss": 0.2345, + "step": 225 + }, + { + "epoch": 0.44466403162055335, + "step": 225, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_dis": 1.0, + "train_accuracy_recipient_token_dis_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.44466403162055335, + "step": 225, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.44664031620553357, + "grad_norm": 2.1454666355505174, + "learning_rate": 3.056302334890786e-06, + "loss": 0.3298, + "step": 226 + }, + { + "epoch": 0.44664031620553357, + "step": 226, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_stamp": 1.0, + "train_accuracy_recipient_token_stamp_total": 3, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.44664031620553357, + "step": 226, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.44861660079051385, + "grad_norm": 2.377176691255679, + "learning_rate": 3.0406643431082088e-06, + "loss": 0.2614, + "step": 227 + }, + { + "epoch": 0.44861660079051385, + "step": 227, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_locate": 1.0, + "train_accuracy_recipient_token_locate_total": 1, + "train_accuracy_recipient_token_provide": 1.0, + "train_accuracy_recipient_token_provide_total": 1, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.44861660079051385, + "step": 227, + "train_accuracy_recipient_token": 0.8125, + "train_accuracy_recipient_token_all": 0.7272727272727273, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_order": 1.0, + "train_accuracy_recipient_token_order_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.4505928853754941, + "grad_norm": 1.5943078704191154, + "learning_rate": 3.0250041267321234e-06, + "loss": 0.148, + "step": 228 + }, + { + "epoch": 0.4505928853754941, + "step": 228, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_configure": 1.0, + "train_accuracy_recipient_token_configure_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_track": 0.5, + "train_accuracy_recipient_token_track_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.4505928853754941, + "step": 228, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.4525691699604743, + "grad_norm": 2.166389899834986, + "learning_rate": 3.009322329492689e-06, + "loss": 0.1702, + "step": 229 + }, + { + "epoch": 0.4525691699604743, + "step": 229, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_download": 1.0, + "train_accuracy_recipient_token_download_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_rec": 1.0, + "train_accuracy_recipient_token_rec_total": 4, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.4525691699604743, + "step": 229, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 3, + "train_accuracy_recipient_token_get": 0.5, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 2, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.45454545454545453, + "grad_norm": 2.2170367848067176, + "learning_rate": 2.993619596007168e-06, + "loss": 0.2933, + "step": 230 + }, + { + "epoch": 0.45454545454545453, + "step": 230, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.8333333333333334, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_describe": 1.0, + "train_accuracy_recipient_token_describe_total": 1, + "train_accuracy_recipient_token_find": 0.75, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_organ": 1.0, + "train_accuracy_recipient_token_organ_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.45454545454545453, + "step": 230, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_generate": 1.0, + "train_accuracy_recipient_token_generate_total": 5, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.45652173913043476, + "grad_norm": 1.4984470237108378, + "learning_rate": 2.9778965717534314e-06, + "loss": 0.1063, + "step": 231 + }, + { + "epoch": 0.45652173913043476, + "step": 231, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_f": 1.0, + "train_accuracy_recipient_token_f_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.45652173913043476, + "step": 231, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 0.9166666666666666, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_m": 0.75, + "train_accuracy_recipient_token_m_total": 4, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.45849802371541504, + "grad_norm": 2.713980767721741, + "learning_rate": 2.9621539030434223e-06, + "loss": 0.3257, + "step": 232 + }, + { + "epoch": 0.45849802371541504, + "step": 232, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.875, + "train_accuracy_recipient_token_all_total": 8, + "train_accuracy_recipient_token_check": 0.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_histor": 1.0, + "train_accuracy_recipient_token_histor_total": 2, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.45849802371541504, + "step": 232, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 0.9166666666666666, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_decode": 1.0, + "train_accuracy_recipient_token_decode_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.46047430830039526, + "grad_norm": 1.9449680707287624, + "learning_rate": 2.946392236996592e-06, + "loss": 0.1973, + "step": 233 + }, + { + "epoch": 0.46047430830039526, + "step": 233, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8888888888888888, + "train_accuracy_recipient_token_all_total": 9, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, + "train_accuracy_recipient_token_estimate": 0.5, + "train_accuracy_recipient_token_estimate_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.46047430830039526, + "step": 233, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9166666666666666, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_translate": 0.0, + "train_accuracy_recipient_token_translate_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.4624505928853755, + "grad_norm": 2.428900846914706, + "learning_rate": 2.9306122215132974e-06, + "loss": 0.2668, + "step": 234 + }, + { + "epoch": 0.4624505928853755, + "step": 234, + "train_accuracy_recipient_token": 0.9375, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.4624505928853755, + "step": 234, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.8333333333333334, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.4644268774703557, + "grad_norm": 2.3084940546349375, + "learning_rate": 2.91481450524817e-06, + "loss": 0.2538, + "step": 235 + }, + { + "epoch": 0.4644268774703557, + "step": 235, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 5, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.4644268774703557, + "step": 235, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_sum": 1.0, + "train_accuracy_recipient_token_sum_total": 1, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.466403162055336, + "grad_norm": 1.7162275334301005, + "learning_rate": 2.8989997375834485e-06, + "loss": 0.2411, + "step": 236 + }, + { + "epoch": 0.466403162055336, + "step": 236, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 5, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_plan": 1.0, + "train_accuracy_recipient_token_plan_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.466403162055336, + "step": 236, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.3333333333333333, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_p": 1.0, + "train_accuracy_recipient_token_p_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.4683794466403162, + "grad_norm": 1.597659173981818, + "learning_rate": 2.8831685686022897e-06, + "loss": 0.1603, + "step": 237 + }, + { + "epoch": 0.4683794466403162, + "step": 237, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 3, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.4683794466403162, + "step": 237, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 9, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_shop": 1.0, + "train_accuracy_recipient_token_shop_total": 2, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.47035573122529645, + "grad_norm": 1.9065156428525225, + "learning_rate": 2.8673216490620453e-06, + "loss": 0.1811, + "step": 238 + }, + { + "epoch": 0.47035573122529645, + "step": 238, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.47035573122529645, + "step": 238, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.4723320158102767, + "grad_norm": 2.176658370568172, + "learning_rate": 2.8514596303675073e-06, + "loss": 0.1601, + "step": 239 + }, + { + "epoch": 0.4723320158102767, + "step": 239, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_invest": 1.0, + "train_accuracy_recipient_token_invest_total": 1, + "train_accuracy_recipient_token_vehicle": 1.0, + "train_accuracy_recipient_token_vehicle_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.4723320158102767, + "step": 239, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.4743083003952569, + "grad_norm": 1.768442218413903, + "learning_rate": 2.835583164544139e-06, + "loss": 0.1682, + "step": 240 + }, + { + "epoch": 0.4743083003952569, + "step": 240, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.4743083003952569, + "step": 240, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_art": 1.0, + "train_accuracy_recipient_token_art_total": 3, + "train_accuracy_recipient_token_artist": 0.0, + "train_accuracy_recipient_token_artist_total": 1, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, + "train_accuracy_recipient_token_water": 1.0, + "train_accuracy_recipient_token_water_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.4762845849802372, + "grad_norm": 1.725662364780488, + "learning_rate": 2.8196929042112652e-06, + "loss": 0.1773, + "step": 241 + }, + { + "epoch": 0.4762845849802372, + "step": 241, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.4762845849802372, + "step": 241, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.4782608695652174, + "grad_norm": 2.278665078686508, + "learning_rate": 2.8037895025552513e-06, + "loss": 0.2275, + "step": 242 + }, + { + "epoch": 0.4782608695652174, + "step": 242, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 0.5, + "train_accuracy_recipient_token_check_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.4782608695652174, + "step": 242, + "train_accuracy_recipient_token": 0.875, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 2, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.48023715415019763, + "grad_norm": 2.4836024942390087, + "learning_rate": 2.787873613302649e-06, + "loss": 0.1997, + "step": 243 + }, + { + "epoch": 0.48023715415019763, + "step": 243, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_detect": 1.0, + "train_accuracy_recipient_token_detect_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 2, + "train_total_number_recipient_token": 21 + }, + { + "epoch": 0.48023715415019763, + "step": 243, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 0.5, + "train_accuracy_recipient_token_calculate_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.48221343873517786, + "grad_norm": 2.3999235961249066, + "learning_rate": 2.7719458906933277e-06, + "loss": 0.1463, + "step": 244 + }, + { + "epoch": 0.48221343873517786, + "step": 244, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_list": 1.0, + "train_accuracy_recipient_token_list_total": 5, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.48221343873517786, + "step": 244, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 3, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.4841897233201581, + "grad_norm": 1.7608203760764525, + "learning_rate": 2.7560069894535783e-06, + "loss": 0.2735, + "step": 245 + }, + { + "epoch": 0.4841897233201581, + "step": 245, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_color": 1.0, + "train_accuracy_recipient_token_color_total": 1, + "train_accuracy_recipient_token_draw": 1.0, + "train_accuracy_recipient_token_draw_total": 1, + "train_accuracy_recipient_token_object": 1.0, + "train_accuracy_recipient_token_object_total": 1, + "train_accuracy_recipient_token_scene": 1.0, + "train_accuracy_recipient_token_scene_total": 2, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.4841897233201581, + "step": 245, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.48616600790513836, + "grad_norm": 1.6831661789344432, + "learning_rate": 2.7400575647692046e-06, + "loss": 0.1882, + "step": 246 + }, + { + "epoch": 0.48616600790513836, + "step": 246, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_ins": 1.0, + "train_accuracy_recipient_token_ins_total": 4, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.48616600790513836, + "step": 246, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_review": 1.0, + "train_accuracy_recipient_token_review_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.4881422924901186, + "grad_norm": 1.8826315438523655, + "learning_rate": 2.724098272258584e-06, + "loss": 0.2473, + "step": 247 + }, + { + "epoch": 0.4881422924901186, + "step": 247, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_learn": 1.0, + "train_accuracy_recipient_token_learn_total": 1, + "train_accuracy_recipient_token_research": 1.0, + "train_accuracy_recipient_token_research_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.4881422924901186, + "step": 247, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_s": 1.0, + "train_accuracy_recipient_token_s_total": 1, + "train_accuracy_recipient_token_stat": 1.0, + "train_accuracy_recipient_token_stat_total": 1, + "train_total_number_recipient_token": 15 + }, + { + "epoch": 0.4901185770750988, + "grad_norm": 1.9713625520958284, + "learning_rate": 2.7081297679457238e-06, + "loss": 0.3234, + "step": 248 + }, + { + "epoch": 0.4901185770750988, + "step": 248, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_st": 1.0, + "train_accuracy_recipient_token_st_total": 3, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.4901185770750988, + "step": 248, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_count": 1.0, + "train_accuracy_recipient_token_count_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 2, + "train_accuracy_recipient_token_flag": 1.0, + "train_accuracy_recipient_token_flag_total": 1, + "train_accuracy_recipient_token_pl": 0.0, + "train_accuracy_recipient_token_pl_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.49209486166007904, + "grad_norm": 2.4921370963085328, + "learning_rate": 2.692152708233292e-06, + "loss": 0.4432, + "step": 249 + }, + { + "epoch": 0.49209486166007904, + "step": 249, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.49209486166007904, + "step": 249, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_create": 1.0, + "train_accuracy_recipient_token_create_total": 2, + "train_accuracy_recipient_token_design": 1.0, + "train_accuracy_recipient_token_design_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.49407114624505927, + "grad_norm": 2.402509089657591, + "learning_rate": 2.676167749875635e-06, + "loss": 0.2032, + "step": 250 + }, + { + "epoch": 0.49407114624505927, + "step": 250, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.49407114624505927, + "step": 250, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.49604743083003955, + "grad_norm": 2.717753245253635, + "learning_rate": 2.6601755499517826e-06, + "loss": 0.315, + "step": 251 + }, + { + "epoch": 0.49604743083003955, + "step": 251, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 2, + "train_accuracy_recipient_token_compile": 1.0, + "train_accuracy_recipient_token_compile_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.49604743083003955, + "step": 251, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.4980237154150198, + "grad_norm": 1.8526560236000496, + "learning_rate": 2.6441767658384363e-06, + "loss": 0.2212, + "step": 252 + }, + { + "epoch": 0.4980237154150198, + "step": 252, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.8888888888888888, + "train_accuracy_recipient_token_all_total": 9, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_setup": 1.0, + "train_accuracy_recipient_token_setup_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.4980237154150198, + "step": 252, + "train_accuracy_recipient_token": 0.75, + "train_accuracy_recipient_token_all": 0.6666666666666666, + "train_accuracy_recipient_token_all_total": 9, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 12 + }, + { + "epoch": 0.5, + "grad_norm": 2.0997848100152594, + "learning_rate": 2.628172055182948e-06, + "loss": 0.188, + "step": 253 + }, + { + "epoch": 0.5, + "step": 253, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5, + "step": 253, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.8571428571428571, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.5019762845849802, + "grad_norm": 2.296699723342699, + "learning_rate": 2.6121620758762877e-06, + "loss": 0.3644, + "step": 254 + }, + { + "epoch": 0.5019762845849802, + "step": 254, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_camera": 1.0, + "train_accuracy_recipient_token_camera_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_invest": 1.0, + "train_accuracy_recipient_token_invest_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5019762845849802, + "step": 254, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.5039525691699605, + "grad_norm": 1.7053263617954735, + "learning_rate": 2.596147486025996e-06, + "loss": 0.2601, + "step": 255 + }, + { + "epoch": 0.5039525691699605, + "step": 255, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_m": 1.0, + "train_accuracy_recipient_token_m_total": 4, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5039525691699605, + "step": 255, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5059288537549407, + "grad_norm": 1.6057303693260692, + "learning_rate": 2.580128943929139e-06, + "loss": 0.1021, + "step": 256 + }, + { + "epoch": 0.5059288537549407, + "step": 256, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, + "train_accuracy_recipient_token_extract": 1.0, + "train_accuracy_recipient_token_extract_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_recogn": 0.0, + "train_accuracy_recipient_token_recogn_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5059288537549407, + "step": 256, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 10, + "train_total_number_recipient_token": 21 + }, + { + "epoch": 0.5079051383399209, + "grad_norm": 2.160182965266891, + "learning_rate": 2.564107108045239e-06, + "loss": 0.3013, + "step": 257 + }, + { + "epoch": 0.5079051383399209, + "step": 257, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5079051383399209, + "step": 257, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 5, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5098814229249012, + "grad_norm": 2.540894830262886, + "learning_rate": 2.5480826369692178e-06, + "loss": 0.2537, + "step": 258 + }, + { + "epoch": 0.5098814229249012, + "step": 258, + "train_accuracy_recipient_token": 0.8666666666666667, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 15 + }, + { + "epoch": 0.5098814229249012, + "step": 258, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, + "train_accuracy_recipient_token_histor": 0.5, + "train_accuracy_recipient_token_histor_total": 2, + "train_accuracy_recipient_token_organ": 1.0, + "train_accuracy_recipient_token_organ_total": 1, + "train_accuracy_recipient_token_public": 1.0, + "train_accuracy_recipient_token_public_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.5118577075098815, + "grad_norm": 1.6078898635903305, + "learning_rate": 2.532056189404318e-06, + "loss": 0.1394, + "step": 259 + }, + { + "epoch": 0.5118577075098815, + "step": 259, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5118577075098815, + "step": 259, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_plant": 1.0, + "train_accuracy_recipient_token_plant_total": 3, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5138339920948617, + "grad_norm": 4.285683109324995, + "learning_rate": 2.516028424135028e-06, + "loss": 0.2524, + "step": 260 + }, + { + "epoch": 0.5138339920948617, + "step": 260, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, + "train_accuracy_recipient_token_evaluate": 1.0, + "train_accuracy_recipient_token_evaluate_total": 1, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_plan": 1.0, + "train_accuracy_recipient_token_plan_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.5138339920948617, + "step": 260, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 5, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_update": 1.0, + "train_accuracy_recipient_token_update_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5158102766798419, + "grad_norm": 1.4223559498348044, + "learning_rate": 2.5e-06, + "loss": 0.1352, + "step": 261 + }, + { + "epoch": 0.5158102766798419, + "step": 261, + "train_accuracy_recipient_token": 0.9333333333333333, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_int": 1.0, + "train_accuracy_recipient_token_int_total": 1, + "train_total_number_recipient_token": 15 + }, + { + "epoch": 0.5158102766798419, + "step": 261, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.5177865612648221, + "grad_norm": 2.2628735436679763, + "learning_rate": 2.4839715758649724e-06, + "loss": 0.3725, + "step": 262 + }, + { + "epoch": 0.5177865612648221, + "step": 262, + "train_accuracy_recipient_token": 0.875, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.5177865612648221, + "step": 262, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_ca": 1.0, + "train_accuracy_recipient_token_ca_total": 4, + "train_accuracy_recipient_token_decode": 1.0, + "train_accuracy_recipient_token_decode_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.5197628458498024, + "grad_norm": 2.2947519367621347, + "learning_rate": 2.467943810595682e-06, + "loss": 0.4472, + "step": 263 + }, + { + "epoch": 0.5197628458498024, + "step": 263, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_art": 1.0, + "train_accuracy_recipient_token_art_total": 2, + "train_accuracy_recipient_token_decode": 1.0, + "train_accuracy_recipient_token_decode_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.5197628458498024, + "step": 263, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.5217391304347826, + "grad_norm": 2.106955012004411, + "learning_rate": 2.4519173630307826e-06, + "loss": 0.3995, + "step": 264 + }, + { + "epoch": 0.5217391304347826, + "step": 264, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 7, + "train_accuracy_recipient_token_count": 1.0, + "train_accuracy_recipient_token_count_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.5217391304347826, + "step": 264, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_create": 1.0, + "train_accuracy_recipient_token_create_total": 1, + "train_accuracy_recipient_token_draw": 1.0, + "train_accuracy_recipient_token_draw_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.5237154150197628, + "grad_norm": 3.6280464608191956, + "learning_rate": 2.4358928919547616e-06, + "loss": 0.3171, + "step": 265 + }, + { + "epoch": 0.5237154150197628, + "step": 265, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_phone": 1.0, + "train_accuracy_recipient_token_phone_total": 2, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.5237154150197628, + "step": 265, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.8333333333333334, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_generate": 0.5, + "train_accuracy_recipient_token_generate_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.525691699604743, + "grad_norm": 1.7300477450689355, + "learning_rate": 2.4198710560708623e-06, + "loss": 0.2653, + "step": 266 + }, + { + "epoch": 0.525691699604743, + "step": 266, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 0.5, + "train_accuracy_recipient_token_check_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_sum": 1.0, + "train_accuracy_recipient_token_sum_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.525691699604743, + "step": 266, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compress": 1.0, + "train_accuracy_recipient_token_compress_total": 1, + "train_accuracy_recipient_token_crop": 1.0, + "train_accuracy_recipient_token_crop_total": 1, + "train_accuracy_recipient_token_d": 1.0, + "train_accuracy_recipient_token_d_total": 2, + "train_accuracy_recipient_token_enh": 1.0, + "train_accuracy_recipient_token_enh_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.5276679841897233, + "grad_norm": 2.345904530746849, + "learning_rate": 2.403852513974004e-06, + "loss": 0.2374, + "step": 267 + }, + { + "epoch": 0.5276679841897233, + "step": 267, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_custom": 0.75, + "train_accuracy_recipient_token_custom_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.5276679841897233, + "step": 267, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 0.7142857142857143, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5296442687747036, + "grad_norm": 2.7558657907586137, + "learning_rate": 2.3878379241237136e-06, + "loss": 0.3184, + "step": 268 + }, + { + "epoch": 0.5296442687747036, + "step": 268, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.5, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.5296442687747036, + "step": 268, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 5, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 10 + }, + { + "epoch": 0.5316205533596838, + "grad_norm": 7.208775244573762, + "learning_rate": 2.3718279448170527e-06, + "loss": 0.8561, + "step": 269 + }, + { + "epoch": 0.5316205533596838, + "step": 269, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 0.7142857142857143, + "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.5316205533596838, + "step": 269, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_describe": 1.0, + "train_accuracy_recipient_token_describe_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.5335968379446641, + "grad_norm": 2.6145328310474136, + "learning_rate": 2.3558232341615645e-06, + "loss": 0.2586, + "step": 270 + }, + { + "epoch": 0.5335968379446641, + "step": 270, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.441699604743083, - "step": 149, - "train_accuracy_first_token_arguments": 0.8095238095238095, - "train_accuracy_first_token_arguments_total": 21, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.5335968379446641, + "step": 270, + "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 3, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_ex": 1.0, + "train_accuracy_recipient_token_ex_total": 3, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.441699604743083, - "step": 149, - "train_accuracy_first_token_arguments": 0.6666666666666666, - "train_accuracy_first_token_arguments_total": 21, - "train_accuracy_recipient_token": 0.95, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 1, + "epoch": 0.5355731225296443, + "grad_norm": 2.3232906112218323, + "learning_rate": 2.339824450048218e-06, + "loss": 0.2427, + "step": 271 + }, + { + "epoch": 0.5355731225296443, + "step": 271, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_get": 0.8, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_suggest": 1.0, - "train_accuracy_recipient_token_suggest_total": 2, + "train_accuracy_recipient_token_fish": 1.0, + "train_accuracy_recipient_token_fish_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_purchase": 1.0, + "train_accuracy_recipient_token_purchase_total": 1, "train_total_number_recipient_token": 20 }, { - "epoch": 0.441699604743083, - "step": 149, - "train_accuracy_first_token_arguments": 0.9285714285714286, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.5355731225296443, + "step": 271, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_get": 0.9, + "train_accuracy_recipient_token_get_total": 10, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.44466403162055335, - "grad_norm": 1.6792776522615755, - "learning_rate": 3.0730661177177957e-06, - "loss": 0.2082, - "step": 150 + "epoch": 0.5375494071146245, + "grad_norm": 2.208719311446786, + "learning_rate": 2.323832250124365e-06, + "loss": 0.2572, + "step": 272 }, { - "epoch": 0.44466403162055335, - "step": 150, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 10, + "epoch": 0.5375494071146245, + "step": 272, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_discover": 1.0, - "train_accuracy_recipient_token_discover_total": 3, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.44466403162055335, - "step": 150, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 17, - "train_accuracy_recipient_token": 0.8823529411764706, - "train_accuracy_recipient_token_all": 0.8, + "epoch": 0.5375494071146245, + "step": 272, + "train_accuracy_recipient_token": 0.9375, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 2, + "train_accuracy_recipient_token_extract": 1.0, + "train_accuracy_recipient_token_extract_total": 1, + "train_accuracy_recipient_token_monitor": 1.0, + "train_accuracy_recipient_token_monitor_total": 1, + "train_accuracy_recipient_token_recogn": 0.0, + "train_accuracy_recipient_token_recogn_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.5395256916996047, + "grad_norm": 1.8998607337146658, + "learning_rate": 2.307847291766709e-06, + "loss": 0.335, + "step": 273 + }, + { + "epoch": 0.5395256916996047, + "step": 273, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_notify": 1.0, + "train_accuracy_recipient_token_notify_total": 1, + "train_accuracy_recipient_token_set": 1.0, + "train_accuracy_recipient_token_set_total": 1, "train_total_number_recipient_token": 17 }, { - "epoch": 0.44466403162055335, - "step": 150, - "train_accuracy_first_token_arguments": 0.9473684210526315, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 6, - "train_accuracy_recipient_token_stamp": 1.0, - "train_accuracy_recipient_token_stamp_total": 3, - "train_total_number_recipient_token": 19 + "epoch": 0.5395256916996047, + "step": 273, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.9166666666666666, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_s": 0.0, + "train_accuracy_recipient_token_s_total": 1, + "train_accuracy_recipient_token_vol": 1.0, + "train_accuracy_recipient_token_vol_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.44762845849802374, - "grad_norm": 2.0438916010331187, - "learning_rate": 3.0495893969092395e-06, - "loss": 0.341, - "step": 151 + "epoch": 0.541501976284585, + "grad_norm": 2.3240212915101277, + "learning_rate": 2.291870232054277e-06, + "loss": 0.2411, + "step": 274 }, { - "epoch": 0.44762845849802374, - "step": 151, - "train_accuracy_first_token_arguments": 0.8888888888888888, - "train_accuracy_first_token_arguments_total": 18, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.8181818181818182, + "epoch": 0.541501976284585, + "step": 274, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_book": 0.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 7, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 1, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.44762845849802374, - "step": 151, - "train_accuracy_first_token_arguments": 0.9166666666666666, - "train_accuracy_first_token_arguments_total": 24, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.541501976284585, + "step": 274, + "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 4, - "train_accuracy_recipient_token_locate": 1.0, - "train_accuracy_recipient_token_locate_total": 1, - "train_accuracy_recipient_token_provide": 1.0, - "train_accuracy_recipient_token_provide_total": 1, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 1, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_order": 1.0, + "train_accuracy_recipient_token_order_total": 1, + "train_accuracy_recipient_token_schedule": 0.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.44762845849802374, - "step": 151, - "train_accuracy_first_token_arguments": 0.8333333333333334, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.8125, - "train_accuracy_recipient_token_all": 0.7272727272727273, + "epoch": 0.5434782608695652, + "grad_norm": 2.1736866226516485, + "learning_rate": 2.2759017277414165e-06, + "loss": 0.2978, + "step": 275 + }, + { + "epoch": 0.5434782608695652, + "step": 275, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_order": 1.0, - "train_accuracy_recipient_token_order_total": 1, - "train_total_number_recipient_token": 16 + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.4505928853754941, - "grad_norm": 1.870929622393286, - "learning_rate": 3.0260616374094208e-06, - "loss": 0.2028, - "step": 152 + "epoch": 0.5434782608695652, + "step": 275, + "train_accuracy_recipient_token": 0.8, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 0.6666666666666666, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 0.0, + "train_accuracy_recipient_token_get_total": 1, + "train_total_number_recipient_token": 15 }, { - "epoch": 0.4505928853754941, - "step": 152, - "train_accuracy_first_token_arguments": 0.7894736842105263, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.8333333333333334, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.5454545454545454, + "grad_norm": 2.2088882652148123, + "learning_rate": 2.2599424352307958e-06, + "loss": 0.2148, + "step": 276 + }, + { + "epoch": 0.5454545454545454, + "step": 276, + "train_accuracy_recipient_token": 0.9375, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_configure": 0.0, - "train_accuracy_recipient_token_configure_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_track": 0.0, - "train_accuracy_recipient_token_track_total": 2, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_log": 1.0, + "train_accuracy_recipient_token_log_total": 1, + "train_total_number_recipient_token": 16 }, { - "epoch": 0.4505928853754941, - "step": 152, - "train_accuracy_first_token_arguments": 0.8636363636363636, - "train_accuracy_first_token_arguments_total": 44, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.5454545454545454, + "step": 276, + "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 4, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_interpret": 1.0, + "train_accuracy_recipient_token_interpret_total": 1, + "train_accuracy_recipient_token_stage": 1.0, + "train_accuracy_recipient_token_stage_total": 4, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.4505928853754941, - "step": 152, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 17, + "epoch": 0.5474308300395256, + "grad_norm": 1.9068759419051124, + "learning_rate": 2.243993010546422e-06, + "loss": 0.2597, + "step": 277 + }, + { + "epoch": 0.5474308300395256, + "step": 277, "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_download": 1.0, - "train_accuracy_recipient_token_download_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, - "train_accuracy_recipient_token_rec": 1.0, - "train_accuracy_recipient_token_rec_total": 4, + "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.4535573122529644, - "grad_norm": 2.1374043090533816, - "learning_rate": 3.0024850241696128e-06, - "loss": 0.1932, - "step": 153 - }, - { - "epoch": 0.4535573122529644, - "step": 153, - "train_accuracy_first_token_arguments": 0.7916666666666666, - "train_accuracy_first_token_arguments_total": 24, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 3, + "epoch": 0.5474308300395256, + "step": 277, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.8333333333333334, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_count": 1.0, + "train_accuracy_recipient_token_count_total": 3, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 2, + "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 19 }, { - "epoch": 0.4535573122529644, - "step": 153, - "train_accuracy_first_token_arguments": 0.8, - "train_accuracy_first_token_arguments_total": 35, - "train_accuracy_recipient_token": 0.7894736842105263, - "train_accuracy_recipient_token_all": 0.8333333333333334, - "train_accuracy_recipient_token_all_total": 12, - "train_accuracy_recipient_token_book": 1.0, - "train_accuracy_recipient_token_book_total": 1, - "train_accuracy_recipient_token_describe": 0.0, - "train_accuracy_recipient_token_describe_total": 1, - "train_accuracy_recipient_token_find": 0.75, - "train_accuracy_recipient_token_find_total": 4, - "train_accuracy_recipient_token_organ": 1.0, - "train_accuracy_recipient_token_organ_total": 1, + "epoch": 0.549407114624506, + "grad_norm": 1.8438582147731286, + "learning_rate": 2.228054109306673e-06, + "loss": 0.196, + "step": 278 + }, + { + "epoch": 0.549407114624506, + "step": 278, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_update": 0.0, + "train_accuracy_recipient_token_update_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.4535573122529644, - "step": 153, - "train_accuracy_first_token_arguments": 0.9555555555555556, - "train_accuracy_first_token_arguments_total": 90, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.549407114624506, + "step": 278, + "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_generate": 1.0, - "train_accuracy_recipient_token_generate_total": 5, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_ident": 0.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.45652173913043476, - "grad_norm": 1.934815353317823, - "learning_rate": 2.9788617466779884e-06, - "loss": 0.2336, - "step": 154 + "epoch": 0.5513833992094862, + "grad_norm": 1.7961818288907425, + "learning_rate": 2.212126386697352e-06, + "loss": 0.1863, + "step": 279 }, { - "epoch": 0.45652173913043476, - "step": 154, - "train_accuracy_first_token_arguments": 0.8205128205128205, - "train_accuracy_first_token_arguments_total": 39, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.5513833992094862, + "step": 279, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_book": 1.0, - "train_accuracy_recipient_token_book_total": 1, - "train_accuracy_recipient_token_f": 1.0, - "train_accuracy_recipient_token_f_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 1, + "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { - "epoch": 0.45652173913043476, - "step": 154, - "train_accuracy_first_token_arguments": 0.8095238095238095, - "train_accuracy_first_token_arguments_total": 21, - "train_accuracy_recipient_token": 0.9, + "epoch": 0.5513833992094862, + "step": 279, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_art": 1.0, + "train_accuracy_recipient_token_art_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_multip": 0.5, - "train_accuracy_recipient_token_multip_total": 4, - "train_total_number_recipient_token": 20 - }, - { - "epoch": 0.45652173913043476, - "step": 154, - "train_accuracy_first_token_arguments": 0.9130434782608695, - "train_accuracy_first_token_arguments_total": 23, - "train_accuracy_recipient_token": 0.8235294117647058, - "train_accuracy_recipient_token_all": 0.875, - "train_accuracy_recipient_token_all_total": 8, - "train_accuracy_recipient_token_check": 0.0, - "train_accuracy_recipient_token_check_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 0.8, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_histor": 1.0, - "train_accuracy_recipient_token_histor_total": 2, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 17 }, { - "epoch": 0.45948616600790515, - "grad_norm": 2.3253564073005224, - "learning_rate": 2.9551939987562866e-06, - "loss": 0.3042, - "step": 155 + "epoch": 0.5533596837944664, + "grad_norm": 2.0438685228831366, + "learning_rate": 2.196210497444749e-06, + "loss": 0.3187, + "step": 280 }, { - "epoch": 0.45948616600790515, - "step": 155, - "train_accuracy_first_token_arguments": 0.95, - "train_accuracy_first_token_arguments_total": 20, - "train_accuracy_recipient_token": 0.95, - "train_accuracy_recipient_token_all": 0.9166666666666666, - "train_accuracy_recipient_token_all_total": 12, - "train_accuracy_recipient_token_decode": 1.0, - "train_accuracy_recipient_token_decode_total": 2, + "epoch": 0.5533596837944664, + "step": 280, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_describe": 1.0, + "train_accuracy_recipient_token_describe_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_schedule": 1.0, - "train_accuracy_recipient_token_schedule_total": 1, - "train_total_number_recipient_token": 20 - }, - { - "epoch": 0.45948616600790515, - "step": 155, - "train_accuracy_first_token_arguments": 0.84, - "train_accuracy_first_token_arguments_total": 75, - "train_accuracy_recipient_token": 0.8947368421052632, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 1, - "train_accuracy_recipient_token_estimate": 0.5, - "train_accuracy_recipient_token_estimate_total": 2, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 3, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.45948616600790515, - "step": 155, - "train_accuracy_first_token_arguments": 0.8, - "train_accuracy_first_token_arguments_total": 10, - "train_accuracy_recipient_token": 0.9444444444444444, + "epoch": 0.5533596837944664, + "step": 280, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_translate": 0.0, - "train_accuracy_recipient_token_translate_total": 1, + "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { - "epoch": 0.4624505928853755, - "grad_norm": 2.2711295887254184, - "learning_rate": 2.93148397835608e-06, - "loss": 0.2997, - "step": 156 + "epoch": 0.5553359683794467, + "grad_norm": 2.026195406795956, + "learning_rate": 2.1803070957887348e-06, + "loss": 0.219, + "step": 281 }, { - "epoch": 0.4624505928853755, - "step": 156, - "train_accuracy_first_token_arguments": 0.5, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.9375, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 1, + "epoch": 0.5553359683794467, + "step": 281, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_convert": 1.0, + "train_accuracy_recipient_token_convert_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_validate": 1.0, - "train_accuracy_recipient_token_validate_total": 1, - "train_total_number_recipient_token": 16 + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.4624505928853755, - "step": 156, - "train_accuracy_first_token_arguments": 0.8333333333333334, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.5553359683794467, + "step": 281, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 2, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_decode": 1.0, + "train_accuracy_recipient_token_decode_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 6, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.4624505928853755, - "step": 156, - "train_accuracy_first_token_arguments": 0.8333333333333334, - "train_accuracy_first_token_arguments_total": 30, + "epoch": 0.5573122529644269, + "grad_norm": 2.4977490050823454, + "learning_rate": 2.1644168354558623e-06, + "loss": 0.1723, + "step": 282 + }, + { + "epoch": 0.5573122529644269, + "step": 282, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_c": 1.0, - "train_accuracy_recipient_token_c_total": 5, + "train_accuracy_recipient_token_ex": 1.0, + "train_accuracy_recipient_token_ex_total": 1, + "train_accuracy_recipient_token_gest": 1.0, + "train_accuracy_recipient_token_gest_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 20 }, { - "epoch": 0.4654150197628458, - "grad_norm": 2.0895540217432282, - "learning_rate": 2.907733887354657e-06, - "loss": 0.2419, - "step": 157 + "epoch": 0.5573122529644269, + "step": 282, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.7692307692307693, + "train_accuracy_recipient_token_all_total": 13, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.4654150197628458, - "step": 157, - "train_accuracy_first_token_arguments": 0.7837837837837838, - "train_accuracy_first_token_arguments_total": 37, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.5592885375494071, + "grad_norm": 2.362592143856781, + "learning_rate": 2.1485403696324935e-06, + "loss": 0.3041, + "step": 283 + }, + { + "epoch": 0.5592885375494071, + "step": 283, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_add": 1.0, + "train_accuracy_recipient_token_add_total": 1, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_sum": 0.0, - "train_accuracy_recipient_token_sum_total": 1, - "train_accuracy_recipient_token_translate": 0.5, - "train_accuracy_recipient_token_translate_total": 2, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_list": 1.0, + "train_accuracy_recipient_token_list_total": 1, + "train_accuracy_recipient_token_update": 1.0, + "train_accuracy_recipient_token_update_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.4654150197628458, - "step": 157, - "train_accuracy_first_token_arguments": 0.7804878048780488, - "train_accuracy_first_token_arguments_total": 41, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.5592885375494071, + "step": 283, + "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 9, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_product": 0.75, + "train_accuracy_recipient_token_product_total": 4, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.5612648221343873, + "grad_norm": 2.234906036510673, + "learning_rate": 2.1326783509379555e-06, + "loss": 0.1854, + "step": 284 + }, + { + "epoch": 0.5612648221343873, + "step": 284, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 4, + "train_accuracy_recipient_token_hang": 1.0, + "train_accuracy_recipient_token_hang_total": 1, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, + "train_accuracy_recipient_token_remove": 1.0, + "train_accuracy_recipient_token_remove_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.5612648221343873, + "step": 284, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 5, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_plan": 1.0, - "train_accuracy_recipient_token_plan_total": 1, - "train_total_number_recipient_token": 19 - }, - { - "epoch": 0.4654150197628458, - "step": 157, - "train_accuracy_first_token_arguments": 0.6896551724137931, - "train_accuracy_first_token_arguments_total": 58, - "train_accuracy_recipient_token": 0.9, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 0.3333333333333333, - "train_accuracy_recipient_token_find_total": 3, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 6, - "train_accuracy_recipient_token_prior": 1.0, - "train_accuracy_recipient_token_prior_total": 1, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_schedule": 0.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.4683794466403162, - "grad_norm": 2.203763069427427, - "learning_rate": 2.8839459313505407e-06, - "loss": 0.2426, - "step": 158 + "epoch": 0.5632411067193676, + "grad_norm": 2.0687918980973863, + "learning_rate": 2.11683143139771e-06, + "loss": 0.1854, + "step": 285 }, { - "epoch": 0.4683794466403162, - "step": 158, - "train_accuracy_first_token_arguments": 0.9285714285714286, - "train_accuracy_first_token_arguments_total": 14, + "epoch": 0.5632411067193676, + "step": 285, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 3, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_recommend": 0.6666666666666666, - "train_accuracy_recipient_token_recommend_total": 3, + "train_accuracy_recipient_token_get": 0.875, + "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { - "epoch": 0.4683794466403162, - "step": 158, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 20, + "epoch": 0.5632411067193676, + "step": 285, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, - "train_accuracy_recipient_token_shop": 1.0, - "train_accuracy_recipient_token_shop_total": 2, - "train_total_number_recipient_token": 19 - }, - { - "epoch": 0.4683794466403162, - "step": 158, - "train_accuracy_first_token_arguments": 0.9411764705882353, - "train_accuracy_first_token_arguments_total": 17, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 8, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.47134387351778656, - "grad_norm": 1.9347965319589504, - "learning_rate": 2.8601223194586613e-06, - "loss": 0.2107, - "step": 159 + "epoch": 0.5652173913043478, + "grad_norm": 3.474374141128837, + "learning_rate": 2.1010002624165528e-06, + "loss": 0.4115, + "step": 286 }, { - "epoch": 0.47134387351778656, - "step": 159, - "train_accuracy_first_token_arguments": 0.7777777777777778, - "train_accuracy_first_token_arguments_total": 27, + "epoch": 0.5652173913043478, + "step": 286, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_c": 1.0, - "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, - "train_total_number_recipient_token": 19 + "train_total_number_recipient_token": 18 }, { - "epoch": 0.47134387351778656, - "step": 159, - "train_accuracy_first_token_arguments": 0.9375, - "train_accuracy_first_token_arguments_total": 16, - "train_accuracy_recipient_token": 0.9411764705882353, + "epoch": 0.5652173913043478, + "step": 286, + "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_check": 1.0, - "train_accuracy_recipient_token_check_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_invest": 1.0, - "train_accuracy_recipient_token_invest_total": 1, - "train_accuracy_recipient_token_vehicle": 1.0, - "train_accuracy_recipient_token_vehicle_total": 1, - "train_total_number_recipient_token": 17 - }, - { - "epoch": 0.47134387351778656, - "step": 159, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_check": 1.0, - "train_accuracy_recipient_token_check_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_compose": 1.0, + "train_accuracy_recipient_token_compose_total": 1, + "train_accuracy_recipient_token_create": 1.0, + "train_accuracy_recipient_token_create_total": 2, + "train_accuracy_recipient_token_generate": 1.0, + "train_accuracy_recipient_token_generate_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 19 }, { - "epoch": 0.4743083003952569, - "grad_norm": 1.9897698233644936, - "learning_rate": 2.8362652641052024e-06, - "loss": 0.1806, - "step": 160 + "epoch": 0.567193675889328, + "grad_norm": 2.803107135984778, + "learning_rate": 2.085185494751831e-06, + "loss": 0.2326, + "step": 287 }, { - "epoch": 0.4743083003952569, - "step": 160, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 13, - "train_accuracy_recipient_token": 0.9444444444444444, + "epoch": 0.567193675889328, + "step": 287, + "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 3, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_schedule": 1.0, - "train_accuracy_recipient_token_schedule_total": 1, - "train_total_number_recipient_token": 18 - }, - { - "epoch": 0.4743083003952569, - "step": 160, - "train_accuracy_first_token_arguments": 0.875, - "train_accuracy_first_token_arguments_total": 16, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_art": 1.0, - "train_accuracy_recipient_token_art_total": 3, - "train_accuracy_recipient_token_artist": 0.0, - "train_accuracy_recipient_token_artist_total": 1, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 1, - "train_accuracy_recipient_token_water": 1.0, - "train_accuracy_recipient_token_water_total": 1, - "train_total_number_recipient_token": 19 - }, - { - "epoch": 0.4743083003952569, - "step": 160, - "train_accuracy_first_token_arguments": 0.9130434782608695, - "train_accuracy_first_token_arguments_total": 23, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 2, + "train_accuracy_recipient_token_generate": 1.0, + "train_accuracy_recipient_token_generate_total": 3, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_schedule": 0.0, - "train_accuracy_recipient_token_schedule_total": 1, - "train_total_number_recipient_token": 18 - }, - { - "epoch": 0.4772727272727273, - "grad_norm": 1.869456628968754, - "learning_rate": 2.8123769808221407e-06, - "loss": 0.1735, - "step": 161 + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.4772727272727273, - "step": 161, - "train_accuracy_first_token_arguments": 0.6521739130434783, - "train_accuracy_first_token_arguments_total": 23, + "epoch": 0.567193675889328, + "step": 287, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 2, + "train_accuracy_recipient_token_draw": 1.0, + "train_accuracy_recipient_token_draw_total": 4, + "train_accuracy_recipient_token_s": 1.0, + "train_accuracy_recipient_token_s_total": 1, "train_total_number_recipient_token": 17 }, { - "epoch": 0.4772727272727273, - "step": 161, - "train_accuracy_first_token_arguments": 0.9166666666666666, - "train_accuracy_first_token_arguments_total": 36, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_check": 1.0, - "train_accuracy_recipient_token_check_total": 2, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 19 + "epoch": 0.5691699604743083, + "grad_norm": 2.1007745780069653, + "learning_rate": 2.069387778486703e-06, + "loss": 0.2901, + "step": 288 }, { - "epoch": 0.4772727272727273, - "step": 161, - "train_accuracy_first_token_arguments": 0.8333333333333334, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.9375, + "epoch": 0.5691699604743083, + "step": 288, + "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 2, + "train_accuracy_recipient_token_get": 0.8333333333333334, + "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 16 }, { - "epoch": 0.48023715415019763, - "grad_norm": 2.502634515773549, - "learning_rate": 2.788459688041495e-06, - "loss": 0.2973, - "step": 162 + "epoch": 0.5691699604743083, + "step": 288, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 5, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.48023715415019763, - "step": 162, - "train_accuracy_first_token_arguments": 0.9375, - "train_accuracy_first_token_arguments_total": 16, - "train_accuracy_recipient_token": 0.8571428571428571, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_detect": 1.0, - "train_accuracy_recipient_token_detect_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, + "epoch": 0.5711462450592886, + "grad_norm": 2.5164590329084304, + "learning_rate": 2.053607763003409e-06, + "loss": 0.1953, + "step": 289 + }, + { + "epoch": 0.5711462450592886, + "step": 289, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8333333333333334, + "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_translate": 0.5, - "train_accuracy_recipient_token_translate_total": 2, - "train_total_number_recipient_token": 21 + "train_total_number_recipient_token": 17 }, { - "epoch": 0.48023715415019763, - "step": 162, - "train_accuracy_first_token_arguments": 0.9473684210526315, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.8888888888888888, + "epoch": 0.5711462450592886, + "step": 289, + "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_calculate": 0.5, - "train_accuracy_recipient_token_calculate_total": 2, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_total_number_recipient_token": 18 - }, - { - "epoch": 0.48023715415019763, - "step": 162, - "train_accuracy_first_token_arguments": 0.9285714285714286, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_list": 1.0, - "train_accuracy_recipient_token_list_total": 5, + "train_accuracy_recipient_token_g": 1.0, + "train_accuracy_recipient_token_g_total": 3, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.48320158102766797, - "grad_norm": 2.341420807222338, - "learning_rate": 2.7645156068893075e-06, - "loss": 0.2014, - "step": 163 + "epoch": 0.5731225296442688, + "grad_norm": 2.1664077194429154, + "learning_rate": 2.037846096956578e-06, + "loss": 0.2971, + "step": 290 }, { - "epoch": 0.48320158102766797, - "step": 163, - "train_accuracy_first_token_arguments": 0.8333333333333334, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.95, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.5731225296442688, + "step": 290, + "train_accuracy_recipient_token": 0.9333333333333333, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 0.8333333333333334, - "train_accuracy_recipient_token_get_total": 6, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 3, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_create": 1.0, + "train_accuracy_recipient_token_create_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_total_number_recipient_token": 15 }, { - "epoch": 0.48320158102766797, - "step": 163, - "train_accuracy_first_token_arguments": 0.4782608695652174, - "train_accuracy_first_token_arguments_total": 46, + "epoch": 0.5731225296442688, + "step": 290, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_color": 1.0, - "train_accuracy_recipient_token_color_total": 1, - "train_accuracy_recipient_token_draw": 1.0, - "train_accuracy_recipient_token_draw_total": 1, - "train_accuracy_recipient_token_object": 1.0, - "train_accuracy_recipient_token_object_total": 1, - "train_accuracy_recipient_token_scene": 1.0, - "train_accuracy_recipient_token_scene_total": 2, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 2, + "train_accuracy_recipient_token_energy": 1.0, + "train_accuracy_recipient_token_energy_total": 1, + "train_accuracy_recipient_token_lands": 1.0, + "train_accuracy_recipient_token_lands_total": 1, + "train_accuracy_recipient_token_s": 1.0, + "train_accuracy_recipient_token_s_total": 1, + "train_total_number_recipient_token": 16 }, { - "epoch": 0.48320158102766797, - "step": 163, - "train_accuracy_first_token_arguments": 0.9090909090909091, - "train_accuracy_first_token_arguments_total": 11, + "epoch": 0.575098814229249, + "grad_norm": 2.7683955116401697, + "learning_rate": 2.02210342824657e-06, + "loss": 0.3624, + "step": 291 + }, + { + "epoch": 0.575098814229249, + "step": 291, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_accuracy_recipient_token_sort": 1.0, + "train_accuracy_recipient_token_sort_total": 1, "train_total_number_recipient_token": 17 }, { - "epoch": 0.48616600790513836, - "grad_norm": 1.6883872990934619, - "learning_rate": 2.7405469609793746e-06, - "loss": 0.2554, - "step": 164 - }, - { - "epoch": 0.48616600790513836, - "step": 164, - "train_accuracy_first_token_arguments": 0.7407407407407407, - "train_accuracy_first_token_arguments_total": 81, + "epoch": 0.575098814229249, + "step": 291, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_inspect": 1.0, - "train_accuracy_recipient_token_inspect_total": 4, + "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_order": 1.0, + "train_accuracy_recipient_token_order_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.48616600790513836, - "step": 164, - "train_accuracy_first_token_arguments": 0.8947368421052632, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.8947368421052632, + "epoch": 0.5770750988142292, + "grad_norm": 1.8345301561689498, + "learning_rate": 2.0063804039928324e-06, + "loss": 0.1363, + "step": 292 + }, + { + "epoch": 0.5770750988142292, + "step": 292, + "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_check": 1.0, - "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, - "train_accuracy_recipient_token_get": 0.8333333333333334, - "train_accuracy_recipient_token_get_total": 6, - "train_accuracy_recipient_token_review": 1.0, - "train_accuracy_recipient_token_review_total": 1, - "train_total_number_recipient_token": 19 - }, - { - "epoch": 0.48616600790513836, - "step": 164, - "train_accuracy_first_token_arguments": 0.7142857142857143, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_learn": 1.0, - "train_accuracy_recipient_token_learn_total": 1, - "train_accuracy_recipient_token_research": 1.0, - "train_accuracy_recipient_token_research_total": 1, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_get_total": 1, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.4891304347826087, - "grad_norm": 1.7224600166430244, - "learning_rate": 2.716555976206748e-06, - "loss": 0.2725, - "step": 165 + "epoch": 0.5770750988142292, + "step": 292, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.4891304347826087, - "step": 165, - "train_accuracy_first_token_arguments": 0.6, - "train_accuracy_first_token_arguments_total": 20, - "train_accuracy_recipient_token": 0.9333333333333333, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, - "train_accuracy_recipient_token_stat": 1.0, - "train_accuracy_recipient_token_stat_total": 1, - "train_accuracy_recipient_token_suggest": 1.0, - "train_accuracy_recipient_token_suggest_total": 1, - "train_total_number_recipient_token": 15 + "epoch": 0.5790513833992095, + "grad_norm": 1.893006363064422, + "learning_rate": 1.9906776705073113e-06, + "loss": 0.1679, + "step": 293 }, { - "epoch": 0.4891304347826087, - "step": 165, - "train_accuracy_first_token_arguments": 0.9166666666666666, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.95, + "epoch": 0.5790513833992095, + "step": 293, + "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_st": 1.0, - "train_accuracy_recipient_token_st_total": 3, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.4891304347826087, - "step": 165, - "train_accuracy_first_token_arguments": 0.6029411764705882, - "train_accuracy_first_token_arguments_total": 68, + "epoch": 0.5790513833992095, + "step": 293, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_count": 1.0, - "train_accuracy_recipient_token_count_total": 1, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 2, - "train_accuracy_recipient_token_flag": 1.0, - "train_accuracy_recipient_token_flag_total": 1, - "train_accuracy_recipient_token_pl": 1.0, - "train_accuracy_recipient_token_pl_total": 1, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_list": 1.0, + "train_accuracy_recipient_token_list_total": 1, + "train_accuracy_recipient_token_stat": 1.0, + "train_accuracy_recipient_token_stat_total": 2, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.49209486166007904, - "grad_norm": 2.488668771406232, - "learning_rate": 2.6925448805410197e-06, - "loss": 0.4772, - "step": 166 + "epoch": 0.5810276679841897, + "grad_norm": 1.8414367691181166, + "learning_rate": 1.9749958732678766e-06, + "loss": 0.197, + "step": 294 }, { - "epoch": 0.49209486166007904, - "step": 166, - "train_accuracy_first_token_arguments": 0.6153846153846154, - "train_accuracy_first_token_arguments_total": 26, + "epoch": 0.5810276679841897, + "step": 294, "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_c": 1.0, - "train_accuracy_recipient_token_c_total": 1, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { - "epoch": 0.49209486166007904, - "step": 166, - "train_accuracy_first_token_arguments": 0.96875, - "train_accuracy_first_token_arguments_total": 32, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_calculate": 1.0, - "train_accuracy_recipient_token_calculate_total": 1, - "train_accuracy_recipient_token_create": 1.0, - "train_accuracy_recipient_token_create_total": 2, - "train_accuracy_recipient_token_design": 1.0, - "train_accuracy_recipient_token_design_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_total_number_recipient_token": 18 - }, - { - "epoch": 0.49209486166007904, - "step": 166, - "train_accuracy_first_token_arguments": 0.8333333333333334, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.9411764705882353, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.5810276679841897, + "step": 294, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_check": 1.0, - "train_accuracy_recipient_token_check_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_schedule": 1.0, - "train_accuracy_recipient_token_schedule_total": 1, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_get": 0.8888888888888888, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.4950592885375494, - "grad_norm": 2.6791883381737107, - "learning_rate": 2.6685159038194202e-06, - "loss": 0.2803, - "step": 167 + "epoch": 0.5830039525691699, + "grad_norm": 2.06959947325046, + "learning_rate": 1.9593356568917916e-06, + "loss": 0.1392, + "step": 295 }, { - "epoch": 0.4950592885375494, - "step": 167, - "train_accuracy_first_token_arguments": 0.8333333333333334, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.5830039525691699, + "step": 295, + "train_accuracy_recipient_token": 0.8571428571428571, + "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 6, - "train_total_number_recipient_token": 18 - }, - { - "epoch": 0.4950592885375494, - "step": 167, - "train_accuracy_first_token_arguments": 0.7083333333333334, - "train_accuracy_first_token_arguments_total": 24, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_check": 1.0, - "train_accuracy_recipient_token_check_total": 2, - "train_accuracy_recipient_token_compile": 1.0, - "train_accuracy_recipient_token_compile_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_monitor": 1.0, + "train_accuracy_recipient_token_monitor_total": 1, + "train_total_number_recipient_token": 14 }, { - "epoch": 0.4950592885375494, - "step": 167, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 9, + "epoch": 0.5830039525691699, + "step": 295, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_construction": 1.0, + "train_accuracy_recipient_token_construction_total": 2, + "train_accuracy_recipient_token_count": 1.0, + "train_accuracy_recipient_token_count_total": 5, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 19 - }, - { - "epoch": 0.4980237154150198, - "grad_norm": 2.068138976829196, - "learning_rate": 2.6444712775397397e-06, - "loss": 0.3013, - "step": 168 + "train_accuracy_recipient_token_get_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.4980237154150198, - "step": 168, - "train_accuracy_first_token_arguments": 0.6559139784946236, - "train_accuracy_first_token_arguments_total": 93, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 1, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_total_number_recipient_token": 19 + "epoch": 0.5849802371541502, + "grad_norm": 2.4681273751326116, + "learning_rate": 1.9436976651092143e-06, + "loss": 0.2056, + "step": 296 }, { - "epoch": 0.4980237154150198, - "step": 168, - "train_accuracy_first_token_arguments": 0.9782608695652174, - "train_accuracy_first_token_arguments_total": 92, - "train_accuracy_recipient_token": 0.75, - "train_accuracy_recipient_token_all": 0.6666666666666666, - "train_accuracy_recipient_token_all_total": 9, + "epoch": 0.5849802371541502, + "step": 296, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_total_number_recipient_token": 12 + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.4980237154150198, - "step": 168, - "train_accuracy_first_token_arguments": 0.75, - "train_accuracy_first_token_arguments_total": 8, + "epoch": 0.5849802371541502, + "step": 296, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_total_number_recipient_token": 18 + "train_total_number_recipient_token": 19 }, { - "epoch": 0.5009881422924901, - "grad_norm": 2.097242382871384, - "learning_rate": 2.6204132346530936e-06, - "loss": 0.3079, - "step": 169 + "epoch": 0.5869565217391305, + "grad_norm": 1.6965670348460065, + "learning_rate": 1.928082540736737e-06, + "loss": 0.1315, + "step": 297 }, { - "epoch": 0.5009881422924901, - "step": 169, - "train_accuracy_first_token_arguments": 0.875, - "train_accuracy_first_token_arguments_total": 32, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.5869565217391305, + "step": 297, + "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 7, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_find": 0.8, + "train_accuracy_recipient_token_find_total": 5, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_make": 1.0, + "train_accuracy_recipient_token_make_total": 1, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.5009881422924901, - "step": 169, - "train_accuracy_first_token_arguments": 0.6764705882352942, - "train_accuracy_first_token_arguments_total": 34, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.5869565217391305, + "step": 297, + "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_camera": 1.0, - "train_accuracy_recipient_token_camera_total": 1, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_estimate": 0.5, + "train_accuracy_recipient_token_estimate_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_invest": 1.0, - "train_accuracy_recipient_token_invest_total": 1, + "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.5009881422924901, - "step": 169, - "train_accuracy_first_token_arguments": 0.7692307692307693, - "train_accuracy_first_token_arguments_total": 13, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_validate": 1.0, - "train_accuracy_recipient_token_validate_total": 1, - "train_total_number_recipient_token": 19 + "epoch": 0.5889328063241107, + "grad_norm": 2.4962220449528836, + "learning_rate": 1.912490925650962e-06, + "loss": 0.1743, + "step": 298 }, { - "epoch": 0.5039525691699605, - "grad_norm": 1.961726304062012, - "learning_rate": 2.5963440093565567e-06, - "loss": 0.3133, - "step": 170 + "epoch": 0.5889328063241107, + "step": 298, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 0.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_compile": 1.0, + "train_accuracy_recipient_token_compile_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.5039525691699605, - "step": 170, - "train_accuracy_first_token_arguments": 0.8461538461538461, - "train_accuracy_first_token_arguments_total": 26, - "train_accuracy_recipient_token": 0.6666666666666666, - "train_accuracy_recipient_token_all": 0.6, + "epoch": 0.5889328063241107, + "step": 298, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_calculate": 0.0, - "train_accuracy_recipient_token_calculate_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_multiply": 0.75, - "train_accuracy_recipient_token_multiply_total": 4, + "train_accuracy_recipient_token_get": 0.875, + "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { - "epoch": 0.5039525691699605, - "step": 170, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 19, + "epoch": 0.5909090909090909, + "grad_norm": 1.7917091655168156, + "learning_rate": 1.8969234607621187e-06, + "loss": 0.1669, + "step": 299 + }, + { + "epoch": 0.5909090909090909, + "step": 299, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 7, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, + "train_accuracy_recipient_token_s": 1.0, + "train_accuracy_recipient_token_s_total": 1, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.5039525691699605, - "step": 170, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 11, + "epoch": 0.5909090909090909, + "step": 299, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 1, - "train_accuracy_recipient_token_extract": 1.0, - "train_accuracy_recipient_token_extract_total": 1, + "train_accuracy_recipient_token_building": 1.0, + "train_accuracy_recipient_token_building_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_recogn": 1.0, - "train_accuracy_recipient_token_recogn_total": 1, - "train_total_number_recipient_token": 18 - }, - { - "epoch": 0.5069169960474308, - "grad_norm": 2.0506442694884104, - "learning_rate": 2.572265836885682e-06, - "loss": 0.2265, - "step": 171 + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.5069169960474308, - "step": 171, - "train_accuracy_first_token_arguments": 0.9, - "train_accuracy_first_token_arguments_total": 30, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 10, - "train_total_number_recipient_token": 21 + "epoch": 0.5928853754940712, + "grad_norm": 1.5595949160601696, + "learning_rate": 1.8813807859877147e-06, + "loss": 0.1331, + "step": 300 }, { - "epoch": 0.5069169960474308, - "step": 171, - "train_accuracy_first_token_arguments": 0.8125, - "train_accuracy_first_token_arguments_total": 16, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.5928853754940712, + "step": 300, + "train_accuracy_recipient_token": 0.9375, + "train_accuracy_recipient_token_air": 1.0, + "train_accuracy_recipient_token_air_total": 1, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 1, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 16 }, { - "epoch": 0.5069169960474308, - "step": 171, - "train_accuracy_first_token_arguments": 0.8888888888888888, - "train_accuracy_first_token_arguments_total": 9, + "epoch": 0.5928853754940712, + "step": 300, "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 9, + "train_accuracy_recipient_token_average": 1.0, + "train_accuracy_recipient_token_average_total": 1, + "train_accuracy_recipient_token_compare": 0.0, + "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 5, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 18 }, { - "epoch": 0.5098814229249012, - "grad_norm": 2.4101209365627443, - "learning_rate": 2.548180953306918e-06, - "loss": 0.2901, - "step": 172 - }, - { - "epoch": 0.5098814229249012, - "step": 172, - "train_accuracy_first_token_arguments": 0.9, - "train_accuracy_first_token_arguments_total": 20, - "train_accuracy_recipient_token": 0.8666666666666667, - "train_accuracy_recipient_token_all": 0.8, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 15 + "epoch": 0.5948616600790514, + "grad_norm": 2.413658770037087, + "learning_rate": 1.865863540226232e-06, + "loss": 0.249, + "step": 301 }, { - "epoch": 0.5098814229249012, - "step": 172, - "train_accuracy_first_token_arguments": 0.7678571428571429, - "train_accuracy_first_token_arguments_total": 56, - "train_accuracy_recipient_token": 0.9473684210526315, + "epoch": 0.5948616600790514, + "step": 301, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 1, - "train_accuracy_recipient_token_histor": 0.5, - "train_accuracy_recipient_token_histor_total": 2, - "train_accuracy_recipient_token_organ": 1.0, - "train_accuracy_recipient_token_organ_total": 1, - "train_accuracy_recipient_token_public": 1.0, - "train_accuracy_recipient_token_public_total": 1, + "train_accuracy_recipient_token_attach": 1.0, + "train_accuracy_recipient_token_attach_total": 1, + "train_accuracy_recipient_token_clean": 1.0, + "train_accuracy_recipient_token_clean_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_remove": 1.0, + "train_accuracy_recipient_token_remove_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.5098814229249012, - "step": 172, - "train_accuracy_first_token_arguments": 0.5714285714285714, - "train_accuracy_first_token_arguments_total": 21, + "epoch": 0.5948616600790514, + "step": 301, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_generate": 1.0, + "train_accuracy_recipient_token_generate_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 1, - "train_accuracy_recipient_token_search": 1.0, - "train_accuracy_recipient_token_search_total": 1, + "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { - "epoch": 0.5128458498023716, - "grad_norm": 2.403125205150021, - "learning_rate": 2.524091595309952e-06, - "loss": 0.2281, - "step": 173 + "epoch": 0.5968379446640316, + "grad_norm": 1.981013163078487, + "learning_rate": 1.8503723613308683e-06, + "loss": 0.2832, + "step": 302 }, { - "epoch": 0.5128458498023716, - "step": 173, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 11, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.5968379446640316, + "step": 302, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_ext": 0.0, + "train_accuracy_recipient_token_ext_total": 1, + "train_accuracy_recipient_token_fire": 1.0, + "train_accuracy_recipient_token_fire_total": 4, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_plant": 1.0, - "train_accuracy_recipient_token_plant_total": 3, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.5968379446640316, + "step": 302, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_d": 1.0, + "train_accuracy_recipient_token_d_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.5128458498023716, - "step": 173, - "train_accuracy_first_token_arguments": 0.9285714285714286, - "train_accuracy_first_token_arguments_total": 28, - "train_accuracy_recipient_token": 0.9, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 1, - "train_accuracy_recipient_token_evaluate": 1.0, - "train_accuracy_recipient_token_evaluate_total": 1, - "train_accuracy_recipient_token_get": 0.8, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_plan": 1.0, - "train_accuracy_recipient_token_plan_total": 1, - "train_total_number_recipient_token": 20 + "epoch": 0.5988142292490118, + "grad_norm": 1.8168717362636846, + "learning_rate": 1.8349078860833125e-06, + "loss": 0.1964, + "step": 303 }, { - "epoch": 0.5128458498023716, - "step": 173, - "train_accuracy_first_token_arguments": 0.8947368421052632, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.8181818181818182, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 5, + "epoch": 0.5988142292490118, + "step": 303, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, - "train_accuracy_recipient_token_update": 1.0, - "train_accuracy_recipient_token_update_total": 1, - "train_total_number_recipient_token": 18 - }, - { - "epoch": 0.5158102766798419, - "grad_norm": 1.7135004510273728, - "learning_rate": 2.5e-06, - "loss": 0.1725, - "step": 174 + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 2, + "train_accuracy_recipient_token_plant": 1.0, + "train_accuracy_recipient_token_plant_total": 3, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.5158102766798419, - "step": 174, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 7, - "train_accuracy_recipient_token": 0.9333333333333333, + "epoch": 0.5988142292490118, + "step": 303, + "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_convert": 1.0, + "train_accuracy_recipient_token_convert_total": 1, + "train_accuracy_recipient_token_create": 0.0, + "train_accuracy_recipient_token_create_total": 1, + "train_accuracy_recipient_token_generate": 1.0, + "train_accuracy_recipient_token_generate_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_int": 1.0, - "train_accuracy_recipient_token_int_total": 1, - "train_total_number_recipient_token": 15 + "train_total_number_recipient_token": 18 }, { - "epoch": 0.5158102766798419, - "step": 174, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 13, + "epoch": 0.6007905138339921, + "grad_norm": 1.9313706363255394, + "learning_rate": 1.8194707501675724e-06, + "loss": 0.2054, + "step": 304 + }, + { + "epoch": 0.6007905138339921, + "step": 304, "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_measure": 1.0, + "train_accuracy_recipient_token_measure_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.6007905138339921, + "step": 304, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 6, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { - "epoch": 0.5158102766798419, - "step": 174, - "train_accuracy_first_token_arguments": 0.75, - "train_accuracy_first_token_arguments_total": 16, - "train_accuracy_recipient_token": 0.875, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.6027667984189723, + "grad_norm": 2.503698222924907, + "learning_rate": 1.8040615881438425e-06, + "loss": 0.4178, + "step": 305 + }, + { + "epoch": 0.6027667984189723, + "step": 305, + "train_accuracy_recipient_token": 0.8125, + "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 0.75, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_d": 1.0, + "train_accuracy_recipient_token_d_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 16 }, { - "epoch": 0.5187747035573123, - "grad_norm": 2.193026807831911, - "learning_rate": 2.475908404690049e-06, - "loss": 0.4577, - "step": 175 + "epoch": 0.6027667984189723, + "step": 305, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 0.9166666666666666, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_bridge": 1.0, + "train_accuracy_recipient_token_bridge_total": 3, + "train_accuracy_recipient_token_find": 0.75, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_invest": 1.0, + "train_accuracy_recipient_token_invest_total": 1, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.5187747035573123, - "step": 175, - "train_accuracy_first_token_arguments": 0.9090909090909091, - "train_accuracy_first_token_arguments_total": 22, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.6047430830039525, + "grad_norm": 2.0299144209716946, + "learning_rate": 1.7886810334224192e-06, + "loss": 0.2902, + "step": 306 + }, + { + "epoch": 0.6047430830039525, + "step": 306, + "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_ca": 1.0, - "train_accuracy_recipient_token_ca_total": 4, - "train_accuracy_recipient_token_decode": 1.0, - "train_accuracy_recipient_token_decode_total": 1, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 3, + "train_accuracy_recipient_token_track": 1.0, + "train_accuracy_recipient_token_track_total": 1, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.5187747035573123, - "step": 175, - "train_accuracy_first_token_arguments": 0.7058823529411765, - "train_accuracy_first_token_arguments_total": 17, + "epoch": 0.6047430830039525, + "step": 306, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.6067193675889329, + "grad_norm": 1.5819437624234434, + "learning_rate": 1.7733297182376663e-06, + "loss": 0.1652, + "step": 307 + }, + { + "epoch": 0.6067193675889329, + "step": 307, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_art": 1.0, - "train_accuracy_recipient_token_art_total": 2, - "train_accuracy_recipient_token_decode": 1.0, - "train_accuracy_recipient_token_decode_total": 2, + "train_accuracy_recipient_token_book": 0.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.5187747035573123, - "step": 175, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 24, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.6067193675889329, + "step": 307, + "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_organ": 1.0, + "train_accuracy_recipient_token_organ_total": 1, + "train_accuracy_recipient_token_review": 0.5, + "train_accuracy_recipient_token_review_total": 2, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 16 }, { - "epoch": 0.5217391304347826, - "grad_norm": 2.046983156954581, - "learning_rate": 2.4518190466930837e-06, - "loss": 0.4614, - "step": 176 + "epoch": 0.6086956521739131, + "grad_norm": 1.805001545693983, + "learning_rate": 1.7580082736220235e-06, + "loss": 0.1703, + "step": 308 }, { - "epoch": 0.5217391304347826, - "step": 176, - "train_accuracy_first_token_arguments": 0.8235294117647058, - "train_accuracy_first_token_arguments_total": 17, + "epoch": 0.6086956521739131, + "step": 308, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_calculate": 0.8571428571428571, - "train_accuracy_recipient_token_calculate_total": 7, - "train_accuracy_recipient_token_count": 1.0, - "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, - "train_total_number_recipient_token": 19 - }, - { - "epoch": 0.5217391304347826, - "step": 176, - "train_accuracy_first_token_arguments": 0.8666666666666667, - "train_accuracy_first_token_arguments_total": 15, - "train_accuracy_recipient_token": 0.8823529411764706, - "train_accuracy_recipient_token_all": 0.8, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_create": 1.0, - "train_accuracy_recipient_token_create_total": 1, - "train_accuracy_recipient_token_draw": 1.0, - "train_accuracy_recipient_token_draw_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_parse": 0.5, + "train_accuracy_recipient_token_parse_total": 2, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.5217391304347826, - "step": 176, - "train_accuracy_first_token_arguments": 0.9444444444444444, - "train_accuracy_first_token_arguments_total": 18, + "epoch": 0.6086956521739131, + "step": 308, "train_accuracy_recipient_token": 0.9411764705882353, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_phone": 1.0, - "train_accuracy_recipient_token_phone_total": 2, - "train_total_number_recipient_token": 17 - }, - { - "epoch": 0.5247035573122529, - "grad_norm": 2.7763963997279046, - "learning_rate": 2.427734163114319e-06, - "loss": 0.3222, - "step": 177 - }, - { - "epoch": 0.5247035573122529, - "step": 177, - "train_accuracy_first_token_arguments": 0.9473684210526315, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.8235294117647058, - "train_accuracy_recipient_token_all": 0.8333333333333334, - "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_convert": 1.0, + "train_accuracy_recipient_token_convert_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_generate": 0.5, - "train_accuracy_recipient_token_generate_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 17 }, { - "epoch": 0.5247035573122529, - "step": 177, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 16, - "train_accuracy_recipient_token": 0.7647058823529411, - "train_accuracy_recipient_token_all": 0.8, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_check": 0.5, - "train_accuracy_recipient_token_check_total": 2, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 0.6666666666666666, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_sum": 1.0, - "train_accuracy_recipient_token_sum_total": 1, - "train_total_number_recipient_token": 17 + "epoch": 0.6106719367588933, + "grad_norm": 2.3514653723969947, + "learning_rate": 1.74271732938007e-06, + "loss": 0.266, + "step": 309 }, { - "epoch": 0.5247035573122529, - "step": 177, - "train_accuracy_first_token_arguments": 0.8823529411764706, - "train_accuracy_first_token_arguments_total": 17, - "train_accuracy_recipient_token": 0.9411764705882353, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.6106719367588933, + "step": 309, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compress": 1.0, - "train_accuracy_recipient_token_compress_total": 1, - "train_accuracy_recipient_token_crop": 1.0, - "train_accuracy_recipient_token_crop_total": 1, - "train_accuracy_recipient_token_d": 1.0, - "train_accuracy_recipient_token_d_total": 2, - "train_accuracy_recipient_token_enh": 0.0, - "train_accuracy_recipient_token_enh_total": 1, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_get": 0.875, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.6106719367588933, + "step": 309, + "train_accuracy_recipient_token": 0.9047619047619048, + "train_accuracy_recipient_token_all": 0.9285714285714286, + "train_accuracy_recipient_token_all_total": 14, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_find": 0.6666666666666666, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 21 }, { - "epoch": 0.5276679841897233, - "grad_norm": 2.245421899066183, - "learning_rate": 2.403655990643444e-06, - "loss": 0.3072, - "step": 178 + "epoch": 0.6126482213438735, + "grad_norm": 1.995728274275021, + "learning_rate": 1.7274575140626318e-06, + "loss": 0.1586, + "step": 310 }, { - "epoch": 0.5276679841897233, - "step": 178, - "train_accuracy_first_token_arguments": 0.875, - "train_accuracy_first_token_arguments_total": 24, - "train_accuracy_recipient_token": 0.7894736842105263, + "epoch": 0.6126482213438735, + "step": 310, + "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_customize": 0.5, - "train_accuracy_recipient_token_customize_total": 4, - "train_accuracy_recipient_token_get": 0.6, - "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_average": 1.0, + "train_accuracy_recipient_token_average_total": 1, + "train_accuracy_recipient_token_cat": 0.0, + "train_accuracy_recipient_token_cat_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 19 }, { - "epoch": 0.5276679841897233, - "step": 178, - "train_accuracy_first_token_arguments": 0.8333333333333334, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.7777777777777778, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_get": 0.5714285714285714, - "train_accuracy_recipient_token_get_total": 7, + "epoch": 0.6126482213438735, + "step": 310, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.7857142857142857, + "train_accuracy_recipient_token_all_total": 14, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { - "epoch": 0.5276679841897233, - "step": 178, - "train_accuracy_first_token_arguments": 0.8823529411764706, - "train_accuracy_first_token_arguments_total": 17, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_check": 1.0, - "train_accuracy_recipient_token_check_total": 1, - "train_accuracy_recipient_token_find": 0.5, - "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_get": 0.75, - "train_accuracy_recipient_token_get_total": 4, - "train_total_number_recipient_token": 18 + "epoch": 0.6146245059288538, + "grad_norm": 2.072925082708649, + "learning_rate": 1.7122294549409486e-06, + "loss": 0.2363, + "step": 311 }, { - "epoch": 0.5306324110671937, - "grad_norm": 2.405264876175068, - "learning_rate": 2.3795867653469072e-06, - "loss": 0.344, - "step": 179 + "epoch": 0.6146245059288538, + "step": 311, + "train_accuracy_recipient_token": 0.8, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 15 }, { - "epoch": 0.5306324110671937, - "step": 179, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 9, + "epoch": 0.6146245059288538, + "step": 311, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 5, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_design": 1.0, + "train_accuracy_recipient_token_design_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_total_number_recipient_token": 10 + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_monitor": 1.0, + "train_accuracy_recipient_token_monitor_total": 1, + "train_accuracy_recipient_token_order": 1.0, + "train_accuracy_recipient_token_order_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.5306324110671937, - "step": 179, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 15, - "train_accuracy_recipient_token": 0.8421052631578947, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_get": 0.7142857142857143, - "train_accuracy_recipient_token_get_total": 7, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_total_number_recipient_token": 19 + "epoch": 0.616600790513834, + "grad_norm": 2.2383854312267015, + "learning_rate": 1.6970337779808862e-06, + "loss": 0.2371, + "step": 312 }, { - "epoch": 0.5306324110671937, - "step": 179, - "train_accuracy_first_token_arguments": 0.7954545454545454, - "train_accuracy_first_token_arguments_total": 44, + "epoch": 0.616600790513834, + "step": 312, "train_accuracy_recipient_token": 0.8947368421052632, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_calculate": 1.0, - "train_accuracy_recipient_token_calculate_total": 1, - "train_accuracy_recipient_token_describe": 1.0, - "train_accuracy_recipient_token_describe_total": 1, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 0.5, + "train_accuracy_recipient_token_calculate_total": 2, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_schedule": 0.0, + "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.5335968379446641, - "grad_norm": 5.342437189948287, - "learning_rate": 2.355528722460261e-06, - "loss": 0.7124, - "step": 180 + "epoch": 0.616600790513834, + "step": 312, + "train_accuracy_recipient_token": 0.8125, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.8333333333333334, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 16 }, { - "epoch": 0.5335968379446641, - "step": 180, - "train_accuracy_first_token_arguments": 0.8666666666666667, - "train_accuracy_first_token_arguments_total": 15, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_get": 0.8571428571428571, - "train_accuracy_recipient_token_get_total": 7, - "train_total_number_recipient_token": 18 + "epoch": 0.6185770750988142, + "grad_norm": 1.9179139741219748, + "learning_rate": 1.681871107817208e-06, + "loss": 0.2351, + "step": 313 }, { - "epoch": 0.5335968379446641, - "step": 180, - "train_accuracy_first_token_arguments": 0.8461538461538461, - "train_accuracy_first_token_arguments_total": 13, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_explain": 1.0, - "train_accuracy_recipient_token_explain_total": 3, - "train_accuracy_recipient_token_find": 0.5, - "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_get": 0.6666666666666666, - "train_accuracy_recipient_token_get_total": 3, - "train_total_number_recipient_token": 18 + "epoch": 0.6185770750988142, + "step": 313, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 15, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.5335968379446641, - "step": 180, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 22, - "train_accuracy_recipient_token": 0.9, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_fish": 1.0, - "train_accuracy_recipient_token_fish_total": 3, + "epoch": 0.6185770750988142, + "step": 313, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_purchase": 1.0, - "train_accuracy_recipient_token_purchase_total": 1, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_p": 1.0, + "train_accuracy_recipient_token_p_total": 3, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.5365612648221344, - "grad_norm": 2.1777035417905934, - "learning_rate": 2.3314840961805806e-06, - "loss": 0.2801, - "step": 181 + "epoch": 0.6205533596837944, + "grad_norm": 2.832295500440934, + "learning_rate": 1.6667420677278962e-06, + "loss": 0.2569, + "step": 314 }, { - "epoch": 0.5365612648221344, - "step": 181, - "train_accuracy_first_token_arguments": 0.75, - "train_accuracy_first_token_arguments_total": 36, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.6205533596837944, + "step": 314, + "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.8571428571428571, + "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, "train_total_number_recipient_token": 20 }, { - "epoch": 0.5365612648221344, - "step": 181, - "train_accuracy_first_token_arguments": 0.8, - "train_accuracy_first_token_arguments_total": 15, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.6205533596837944, + "step": 314, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 9, - "train_total_number_recipient_token": 19 - }, - { - "epoch": 0.5365612648221344, - "step": 181, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 8, - "train_accuracy_recipient_token": 0.875, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 2, - "train_accuracy_recipient_token_extract": 1.0, - "train_accuracy_recipient_token_extract_total": 1, - "train_accuracy_recipient_token_monitor": 0.0, - "train_accuracy_recipient_token_monitor_total": 1, - "train_accuracy_recipient_token_recogn": 0.0, - "train_accuracy_recipient_token_recogn_total": 1, - "train_total_number_recipient_token": 16 + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_locate": 1.0, + "train_accuracy_recipient_token_locate_total": 1, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.5395256916996047, - "grad_norm": 1.9844826167678842, - "learning_rate": 2.3074551194589816e-06, - "loss": 0.3629, - "step": 182 + "epoch": 0.6225296442687747, + "grad_norm": 1.4578028303646424, + "learning_rate": 1.6516472796085314e-06, + "loss": 0.1809, + "step": 315 }, { - "epoch": 0.5395256916996047, - "step": 182, - "train_accuracy_first_token_arguments": 0.9019607843137255, - "train_accuracy_first_token_arguments_total": 51, - "train_accuracy_recipient_token": 0.8823529411764706, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.6225296442687747, + "step": 315, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 0.8, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_notify": 0.0, - "train_accuracy_recipient_token_notify_total": 1, - "train_accuracy_recipient_token_set": 1.0, - "train_accuracy_recipient_token_set_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { - "epoch": 0.5395256916996047, - "step": 182, - "train_accuracy_first_token_arguments": 0.9090909090909091, - "train_accuracy_first_token_arguments_total": 22, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 12, - "train_accuracy_recipient_token_get": 0.75, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_sponsor": 1.0, - "train_accuracy_recipient_token_sponsor_total": 1, - "train_accuracy_recipient_token_vol": 1.0, - "train_accuracy_recipient_token_vol_total": 1, - "train_total_number_recipient_token": 18 - }, - { - "epoch": 0.5395256916996047, - "step": 182, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 10, - "train_accuracy_recipient_token": 0.8823529411764706, + "epoch": 0.6225296442687747, + "step": 315, + "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_book": 0.0, - "train_accuracy_recipient_token_book_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_validate": 1.0, - "train_accuracy_recipient_token_validate_total": 1, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_validate_total": 3, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.5424901185770751, - "grad_norm": 2.429551580616199, - "learning_rate": 2.2834440237932537e-06, - "loss": 0.3087, - "step": 183 + "epoch": 0.6245059288537549, + "grad_norm": 2.1140362979737986, + "learning_rate": 1.6365873639467314e-06, + "loss": 0.2706, + "step": 316 + }, + { + "epoch": 0.6245059288537549, + "step": 316, + "train_accuracy_recipient_token": 0.875, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_total_number_recipient_token": 16 }, { - "epoch": 0.5424901185770751, - "step": 183, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 17, - "train_accuracy_recipient_token": 0.8947368421052632, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.6245059288537549, + "step": 316, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 0.5, - "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_order": 0.0, - "train_accuracy_recipient_token_order_total": 1, - "train_accuracy_recipient_token_schedule": 1.0, - "train_accuracy_recipient_token_schedule_total": 1, + "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { - "epoch": 0.5424901185770751, - "step": 183, - "train_accuracy_first_token_arguments": 0.9354838709677419, - "train_accuracy_first_token_arguments_total": 31, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.6264822134387352, + "grad_norm": 2.2633287960322597, + "learning_rate": 1.6215629397966432e-06, + "loss": 0.31, + "step": 317 + }, + { + "epoch": 0.6264822134387352, + "step": 317, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_calculate": 1.0, - "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_interpret": 1.0, + "train_accuracy_recipient_token_interpret_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.5424901185770751, - "step": 183, - "train_accuracy_first_token_arguments": 0.875, - "train_accuracy_first_token_arguments_total": 8, - "train_accuracy_recipient_token": 0.8, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_find": 0.6666666666666666, - "train_accuracy_recipient_token_find_total": 3, - "train_accuracy_recipient_token_get": 0.0, - "train_accuracy_recipient_token_get_total": 1, - "train_total_number_recipient_token": 15 + "epoch": 0.6264822134387352, + "step": 317, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.5454545454545454, - "grad_norm": 2.062437887240769, - "learning_rate": 2.259453039020626e-06, - "loss": 0.2448, - "step": 184 + "epoch": 0.6284584980237155, + "grad_norm": 2.5410331978041145, + "learning_rate": 1.6065746247534986e-06, + "loss": 0.23, + "step": 318 }, { - "epoch": 0.5454545454545454, - "step": 184, - "train_accuracy_first_token_arguments": 0.6976744186046512, - "train_accuracy_first_token_arguments_total": 43, - "train_accuracy_recipient_token": 0.9375, + "epoch": 0.6284584980237155, + "step": 318, + "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_log": 1.0, - "train_accuracy_recipient_token_log_total": 1, - "train_total_number_recipient_token": 16 + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.5454545454545454, - "step": 184, - "train_accuracy_first_token_arguments": 0.6153846153846154, - "train_accuracy_first_token_arguments_total": 13, + "epoch": 0.6284584980237155, + "step": 318, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_interpret": 1.0, - "train_accuracy_recipient_token_interpret_total": 1, - "train_accuracy_recipient_token_stage": 1.0, - "train_accuracy_recipient_token_stage_total": 4, + "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 19 }, { - "epoch": 0.5454545454545454, - "step": 184, - "train_accuracy_first_token_arguments": 0.8846153846153846, - "train_accuracy_first_token_arguments_total": 26, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 7, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_total_number_recipient_token": 18 - }, - { - "epoch": 0.5484189723320159, - "grad_norm": 1.7733764406317996, - "learning_rate": 2.2354843931106933e-06, - "loss": 0.2303, - "step": 185 - }, - { - "epoch": 0.5484189723320159, - "step": 185, - "train_accuracy_first_token_arguments": 0.8918918918918919, - "train_accuracy_first_token_arguments_total": 37, - "train_accuracy_recipient_token": 0.8947368421052632, - "train_accuracy_recipient_token_all": 0.9166666666666666, - "train_accuracy_recipient_token_all_total": 12, - "train_accuracy_recipient_token_count": 1.0, - "train_accuracy_recipient_token_count_total": 3, - "train_accuracy_recipient_token_estimate": 0.5, - "train_accuracy_recipient_token_estimate_total": 2, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_total_number_recipient_token": 19 + "epoch": 0.6304347826086957, + "grad_norm": 1.6183376573109411, + "learning_rate": 1.5916230349282215e-06, + "loss": 0.1793, + "step": 319 }, { - "epoch": 0.5484189723320159, - "step": 185, - "train_accuracy_first_token_arguments": 0.6, - "train_accuracy_first_token_arguments_total": 55, + "epoch": 0.6304347826086957, + "step": 319, "train_accuracy_recipient_token": 0.8421052631578947, - "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_check": 0.5, - "train_accuracy_recipient_token_check_total": 2, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_compare": 0.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 0.75, + "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_update": 1.0, - "train_accuracy_recipient_token_update_total": 1, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_locate": 0.0, + "train_accuracy_recipient_token_locate_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.5484189723320159, - "step": 185, - "train_accuracy_first_token_arguments": 0.7352941176470589, - "train_accuracy_first_token_arguments_total": 34, - "train_accuracy_recipient_token": 0.9444444444444444, + "epoch": 0.6304347826086957, + "step": 319, + "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 1, - "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_sum": 0.5, + "train_accuracy_recipient_token_sum_total": 2, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 2, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.5513833992094862, - "grad_norm": 1.7231295422156472, - "learning_rate": 2.211540311958506e-06, - "loss": 0.2471, - "step": 186 + "epoch": 0.6324110671936759, + "grad_norm": 1.526229611696862, + "learning_rate": 1.5767087849221097e-06, + "loss": 0.2264, + "step": 320 }, { - "epoch": 0.5513833992094862, - "step": 186, - "train_accuracy_first_token_arguments": 0.8571428571428571, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 0.9411764705882353, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.6324110671936759, + "step": 320, + "train_accuracy_recipient_token": 0.9375, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_register": 1.0, + "train_accuracy_recipient_token_register_total": 1, + "train_accuracy_recipient_token_verify": 1.0, + "train_accuracy_recipient_token_verify_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.6324110671936759, + "step": 320, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_filter": 1.0, + "train_accuracy_recipient_token_filter_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_ident": 0.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.5513833992094862, - "step": 186, - "train_accuracy_first_token_arguments": 0.9230769230769231, - "train_accuracy_first_token_arguments_total": 13, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.6343873517786561, + "grad_norm": 1.6433946920394795, + "learning_rate": 1.561832487801565e-06, + "loss": 0.199, + "step": 321 + }, + { + "epoch": 0.6343873517786561, + "step": 321, + "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_art": 1.0, - "train_accuracy_recipient_token_art_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_validate": 1.0, - "train_accuracy_recipient_token_validate_total": 1, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_get": 0.8333333333333334, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.5513833992094862, - "step": 186, - "train_accuracy_first_token_arguments": 0.6712328767123288, - "train_accuracy_first_token_arguments_total": 73, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_describe": 1.0, - "train_accuracy_recipient_token_describe_total": 1, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 1, + "epoch": 0.6343873517786561, + "step": 321, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_boat": 1.0, + "train_accuracy_recipient_token_boat_total": 3, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 3, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_s": 1.0, + "train_accuracy_recipient_token_s_total": 2, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.5543478260869565, - "grad_norm": 1.9589966620887675, - "learning_rate": 2.1876230191778598e-06, - "loss": 0.3246, - "step": 187 + "epoch": 0.6363636363636364, + "grad_norm": 1.9498159113092581, + "learning_rate": 1.546994755072896e-06, + "loss": 0.1708, + "step": 322 }, { - "epoch": 0.5543478260869565, - "step": 187, - "train_accuracy_first_token_arguments": 0.8260869565217391, - "train_accuracy_first_token_arguments_total": 23, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.6363636363636364, + "step": 322, + "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 8, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_h": 0.6666666666666666, + "train_accuracy_recipient_token_h_total": 3, + "train_accuracy_recipient_token_invest": 1.0, + "train_accuracy_recipient_token_invest_total": 1, + "train_accuracy_recipient_token_plane": 1.0, + "train_accuracy_recipient_token_plane_total": 1, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.5543478260869565, - "step": 187, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 32, - "train_accuracy_recipient_token": 0.8235294117647058, + "epoch": 0.6363636363636364, + "step": 322, + "train_accuracy_recipient_token": 0.6875, "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_convert": 1.0, - "train_accuracy_recipient_token_convert_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_update": 0.0, + "train_accuracy_recipient_token_update_total": 1, + "train_total_number_recipient_token": 16 }, { - "epoch": 0.5543478260869565, - "step": 187, - "train_accuracy_first_token_arguments": 0.75, - "train_accuracy_first_token_arguments_total": 8, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.6383399209486166, + "grad_norm": 2.3482135262198, + "learning_rate": 1.5321961966571801e-06, + "loss": 0.221, + "step": 323 + }, + { + "epoch": 0.6383399209486166, + "step": 323, + "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_decode": 1.0, - "train_accuracy_recipient_token_decode_total": 1, + "train_accuracy_recipient_token_evaluate": 0.0, + "train_accuracy_recipient_token_evaluate_total": 1, + "train_accuracy_recipient_token_find": 0.3333333333333333, + "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_pack": 1.0, + "train_accuracy_recipient_token_pack_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.5573122529644269, - "grad_norm": 2.2200460526334824, - "learning_rate": 2.1637347358947984e-06, - "loss": 0.216, - "step": 188 - }, - { - "epoch": 0.5573122529644269, - "step": 188, - "train_accuracy_first_token_arguments": 0.95, - "train_accuracy_first_token_arguments_total": 20, + "epoch": 0.6383399209486166, + "step": 323, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_explain": 1.0, - "train_accuracy_recipient_token_explain_total": 1, - "train_accuracy_recipient_token_gesture": 1.0, - "train_accuracy_recipient_token_gesture_total": 4, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_ind": 1.0, + "train_accuracy_recipient_token_ind_total": 4, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 20 }, { - "epoch": 0.5573122529644269, - "step": 188, - "train_accuracy_first_token_arguments": 0.9473684210526315, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.8461538461538461, - "train_accuracy_recipient_token_all_total": 13, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 18 + "epoch": 0.6403162055335968, + "grad_norm": 2.265690935530158, + "learning_rate": 1.5174374208651913e-06, + "loss": 0.2203, + "step": 324 }, { - "epoch": 0.5573122529644269, - "step": 188, - "train_accuracy_first_token_arguments": 0.75, - "train_accuracy_first_token_arguments_total": 24, + "epoch": 0.6403162055335968, + "step": 324, "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_add": 1.0, - "train_accuracy_recipient_token_add_total": 1, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_arch": 1.0, + "train_accuracy_recipient_token_arch_total": 1, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_list": 1.0, - "train_accuracy_recipient_token_list_total": 1, - "train_accuracy_recipient_token_update": 1.0, - "train_accuracy_recipient_token_update_total": 1, - "train_total_number_recipient_token": 18 - }, - { - "epoch": 0.5602766798418972, - "grad_norm": 2.440285593927014, - "learning_rate": 2.13987768054134e-06, - "loss": 0.3342, - "step": 189 + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.5602766798418972, - "step": 189, - "train_accuracy_first_token_arguments": 0.9130434782608695, - "train_accuracy_first_token_arguments_total": 23, - "train_accuracy_recipient_token": 0.9, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 1, + "epoch": 0.6403162055335968, + "step": 324, + "train_accuracy_recipient_token": 0.8, + "train_accuracy_recipient_token_all": 0.7692307692307693, + "train_accuracy_recipient_token_all_total": 13, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 3, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_product": 0.75, - "train_accuracy_recipient_token_product_total": 4, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.8333333333333334, + "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 20 }, { - "epoch": 0.5602766798418972, - "step": 189, - "train_accuracy_first_token_arguments": 0.9886363636363636, - "train_accuracy_first_token_arguments_total": 88, + "epoch": 0.642292490118577, + "grad_norm": 1.6543940069908076, + "learning_rate": 1.502719034372396e-06, + "loss": 0.2158, + "step": 325 + }, + { + "epoch": 0.642292490118577, + "step": 325, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_calculate": 1.0, - "train_accuracy_recipient_token_calculate_total": 4, - "train_accuracy_recipient_token_hang": 1.0, - "train_accuracy_recipient_token_hang_total": 1, - "train_accuracy_recipient_token_re": 0.0, - "train_accuracy_recipient_token_re_total": 1, - "train_accuracy_recipient_token_remove": 1.0, - "train_accuracy_recipient_token_remove_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.875, + "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 20 }, { - "epoch": 0.5602766798418972, - "step": 189, - "train_accuracy_first_token_arguments": 0.7391304347826086, - "train_accuracy_first_token_arguments_total": 23, - "train_accuracy_recipient_token": 0.8333333333333334, - "train_accuracy_recipient_token_all": 0.8, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_book": 1.0, - "train_accuracy_recipient_token_book_total": 1, + "epoch": 0.642292490118577, + "step": 325, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 9, + "train_accuracy_recipient_token_compare": 0.0, + "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.75, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_schedule": 1.0, - "train_accuracy_recipient_token_schedule_total": 1, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.5632411067193676, - "grad_norm": 2.2639594853684795, - "learning_rate": 2.1160540686494597e-06, - "loss": 0.2142, - "step": 190 + "epoch": 0.6442687747035574, + "grad_norm": 1.915721888237991, + "learning_rate": 1.4880416421940155e-06, + "loss": 0.3389, + "step": 326 }, { - "epoch": 0.5632411067193676, - "step": 190, - "train_accuracy_first_token_arguments": 0.7857142857142857, - "train_accuracy_first_token_arguments_total": 14, + "epoch": 0.6442687747035574, + "step": 326, "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.5632411067193676, - "step": 190, - "train_accuracy_first_token_arguments": 0.9166666666666666, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, + "epoch": 0.6442687747035574, + "step": 326, + "train_accuracy_recipient_token": 0.875, + "train_accuracy_recipient_token_all": 0.8888888888888888, + "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_ins": 0.0, + "train_accuracy_recipient_token_ins_total": 1, + "train_total_number_recipient_token": 16 }, { - "epoch": 0.5632411067193676, - "step": 190, - "train_accuracy_first_token_arguments": 0.7777777777777778, - "train_accuracy_first_token_arguments_total": 36, + "epoch": 0.6462450592885376, + "grad_norm": 2.057668318543242, + "learning_rate": 1.4734058476601553e-06, + "loss": 0.1014, + "step": 327 + }, + { + "epoch": 0.6462450592885376, + "step": 327, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_calculate": 1.0, - "train_accuracy_recipient_token_calculate_total": 3, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_language": 1.0, + "train_accuracy_recipient_token_language_total": 1, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 4, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.5662055335968379, - "grad_norm": 3.2075927432234157, - "learning_rate": 2.0922661126453436e-06, - "loss": 0.4145, - "step": 191 + "epoch": 0.6462450592885376, + "step": 327, + "train_accuracy_recipient_token": 0.7647058823529411, + "train_accuracy_recipient_token_all": 0.7272727272727273, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 0.6666666666666666, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.5662055335968379, - "step": 191, - "train_accuracy_first_token_arguments": 0.7166666666666667, - "train_accuracy_first_token_arguments_total": 60, - "train_accuracy_recipient_token": 0.8947368421052632, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compose": 1.0, - "train_accuracy_recipient_token_compose_total": 1, - "train_accuracy_recipient_token_create": 1.0, - "train_accuracy_recipient_token_create_total": 2, - "train_accuracy_recipient_token_generate": 1.0, - "train_accuracy_recipient_token_generate_total": 1, - "train_accuracy_recipient_token_get": 0.75, - "train_accuracy_recipient_token_get_total": 4, - "train_total_number_recipient_token": 19 + "epoch": 0.6482213438735178, + "grad_norm": 1.286134009298201, + "learning_rate": 1.4588122523910032e-06, + "loss": 0.1612, + "step": 328 }, { - "epoch": 0.5662055335968379, - "step": 191, - "train_accuracy_first_token_arguments": 0.6385542168674698, - "train_accuracy_first_token_arguments_total": 83, - "train_accuracy_recipient_token": 0.9411764705882353, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 2, - "train_accuracy_recipient_token_generate": 1.0, - "train_accuracy_recipient_token_generate_total": 3, + "epoch": 0.6482213438735178, + "step": 328, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 9, + "train_accuracy_recipient_token_art": 1.0, + "train_accuracy_recipient_token_art_total": 4, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 17 }, { - "epoch": 0.5662055335968379, - "step": 191, - "train_accuracy_first_token_arguments": 0.7121212121212122, - "train_accuracy_first_token_arguments_total": 66, + "epoch": 0.6482213438735178, + "step": 328, "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.6666666666666666, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.650197628458498, + "grad_norm": 1.7384058205276167, + "learning_rate": 1.4442614562721011e-06, + "loss": 0.175, + "step": 329 + }, + { + "epoch": 0.650197628458498, + "step": 329, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_ex": 1.0, + "train_accuracy_recipient_token_ex_total": 5, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.650197628458498, + "step": 329, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_b": 1.0, + "train_accuracy_recipient_token_b_total": 1, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.6521739130434783, + "grad_norm": 2.0702750470665947, + "learning_rate": 1.429754057429687e-06, + "loss": 0.1972, + "step": 330 + }, + { + "epoch": 0.6521739130434783, + "step": 330, + "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 2, - "train_accuracy_recipient_token_draw": 1.0, - "train_accuracy_recipient_token_draw_total": 4, - "train_accuracy_recipient_token_suggest": 1.0, - "train_accuracy_recipient_token_suggest_total": 1, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_ins": 1.0, + "train_accuracy_recipient_token_ins_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.6521739130434783, + "step": 330, + "train_accuracy_recipient_token": 0.9375, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 16 }, { - "epoch": 0.5691699604743083, - "grad_norm": 1.9745905408889657, - "learning_rate": 2.0685160216439205e-06, - "loss": 0.2894, - "step": 192 + "epoch": 0.6541501976284585, + "grad_norm": 3.148476858942437, + "learning_rate": 1.415290652206105e-06, + "loss": 0.3255, + "step": 331 }, { - "epoch": 0.5691699604743083, - "step": 192, - "train_accuracy_first_token_arguments": 0.7647058823529411, - "train_accuracy_first_token_arguments_total": 17, - "train_accuracy_recipient_token": 0.875, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.6541501976284585, + "step": 331, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, - "train_total_number_recipient_token": 16 + "train_total_number_recipient_token": 19 }, { - "epoch": 0.5691699604743083, - "step": 192, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 53, + "epoch": 0.6541501976284585, + "step": 331, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_calculate": 1.0, - "train_accuracy_recipient_token_calculate_total": 5, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, - "train_total_number_recipient_token": 19 - }, - { - "epoch": 0.5691699604743083, - "step": 192, - "train_accuracy_first_token_arguments": 0.9545454545454546, - "train_accuracy_first_token_arguments_total": 22, - "train_accuracy_recipient_token": 0.8823529411764706, - "train_accuracy_recipient_token_all": 0.8333333333333334, - "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_invest": 1.0, + "train_accuracy_recipient_token_invest_total": 1, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.5721343873517787, - "grad_norm": 2.517506728021728, - "learning_rate": 2.044806001243714e-06, - "loss": 0.2701, - "step": 193 + "epoch": 0.6561264822134387, + "grad_norm": 1.4492697108896657, + "learning_rate": 1.400871835135295e-06, + "loss": 0.1334, + "step": 332 }, { - "epoch": 0.5721343873517787, - "step": 193, - "train_accuracy_first_token_arguments": 0.6296296296296297, - "train_accuracy_first_token_arguments_total": 27, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_check": 1.0, - "train_accuracy_recipient_token_check_total": 1, + "epoch": 0.6561264822134387, + "step": 332, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_g": 1.0, - "train_accuracy_recipient_token_g_total": 3, - "train_accuracy_recipient_token_search": 1.0, - "train_accuracy_recipient_token_search_total": 1, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_load": 1.0, + "train_accuracy_recipient_token_load_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.5721343873517787, - "step": 193, - "train_accuracy_first_token_arguments": 0.75, - "train_accuracy_first_token_arguments_total": 8, - "train_accuracy_recipient_token": 0.9333333333333333, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 1, - "train_accuracy_recipient_token_check": 1.0, - "train_accuracy_recipient_token_check_total": 1, - "train_accuracy_recipient_token_create": 1.0, - "train_accuracy_recipient_token_create_total": 1, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_total_number_recipient_token": 15 - }, - { - "epoch": 0.5721343873517787, - "step": 193, - "train_accuracy_first_token_arguments": 0.5471698113207547, - "train_accuracy_first_token_arguments_total": 53, - "train_accuracy_recipient_token": 0.875, - "train_accuracy_recipient_token_all": 0.8, + "epoch": 0.6561264822134387, + "step": 332, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_check": 1.0, - "train_accuracy_recipient_token_check_total": 2, - "train_accuracy_recipient_token_energy": 1.0, - "train_accuracy_recipient_token_energy_total": 1, - "train_accuracy_recipient_token_landscape": 1.0, - "train_accuracy_recipient_token_landscape_total": 1, - "train_accuracy_recipient_token_suggest": 1.0, - "train_accuracy_recipient_token_suggest_total": 1, - "train_total_number_recipient_token": 16 + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.575098814229249, - "grad_norm": 2.477613864495461, - "learning_rate": 2.021138253322012e-06, - "loss": 0.3818, - "step": 194 + "epoch": 0.658102766798419, + "grad_norm": 1.8406241806364723, + "learning_rate": 1.386498198918352e-06, + "loss": 0.2765, + "step": 333 }, { - "epoch": 0.575098814229249, - "step": 194, - "train_accuracy_first_token_arguments": 0.967741935483871, - "train_accuracy_first_token_arguments_total": 31, + "epoch": 0.658102766798419, + "step": 333, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_device": 1.0, + "train_accuracy_recipient_token_device_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_schedule": 1.0, - "train_accuracy_recipient_token_schedule_total": 1, - "train_accuracy_recipient_token_sort": 1.0, - "train_accuracy_recipient_token_sort_total": 1, + "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 17 }, { - "epoch": 0.575098814229249, - "step": 194, - "train_accuracy_first_token_arguments": 0.9705882352941176, - "train_accuracy_first_token_arguments_total": 34, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.658102766798419, + "step": 333, + "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 7, - "train_accuracy_recipient_token_order": 1.0, - "train_accuracy_recipient_token_order_total": 1, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_get": 0.9, + "train_accuracy_recipient_token_get_total": 10, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.575098814229249, - "step": 194, - "train_accuracy_first_token_arguments": 0.7368421052631579, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.8823529411764706, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.6600790513833992, + "grad_norm": 1.716219108289217, + "learning_rate": 1.3721703343991634e-06, + "loss": 0.0999, + "step": 334 + }, + { + "epoch": 0.6600790513833992, + "step": 334, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_calculate": 1.0, - "train_accuracy_recipient_token_calculate_total": 1, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_park": 1.0, + "train_accuracy_recipient_token_park_total": 2, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 2, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.5780632411067194, - "grad_norm": 1.7625337382807924, - "learning_rate": 1.9975149758303885e-06, - "loss": 0.148, - "step": 195 + "epoch": 0.6600790513833992, + "step": 334, + "train_accuracy_recipient_token": 0.8666666666666667, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 0.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_total_number_recipient_token": 15 }, { - "epoch": 0.5780632411067194, - "step": 195, - "train_accuracy_first_token_arguments": 0.8055555555555556, - "train_accuracy_first_token_arguments_total": 36, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, + "epoch": 0.6620553359683794, + "grad_norm": 2.70419172920364, + "learning_rate": 1.3578888305401208e-06, + "loss": 0.2241, + "step": 335 + }, + { + "epoch": 0.6620553359683794, + "step": 335, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_ex": 1.0, + "train_accuracy_recipient_token_ex_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 6, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_ident": 0.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_translate": 0.0, + "train_accuracy_recipient_token_translate_total": 1, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.5780632411067194, - "step": 195, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 23, - "train_accuracy_recipient_token": 0.9473684210526315, + "epoch": 0.6620553359683794, + "step": 335, + "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_local": 1.0, + "train_accuracy_recipient_token_local_total": 1, + "train_accuracy_recipient_token_plan": 1.0, + "train_accuracy_recipient_token_plan_total": 1, + "train_accuracy_recipient_token_restaurant": 0.5, + "train_accuracy_recipient_token_restaurant_total": 2, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.6640316205533597, + "grad_norm": 2.4216736231197733, + "learning_rate": 1.3436542743979125e-06, + "loss": 0.3036, + "step": 336 + }, + { + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_order": 1.0, + "train_accuracy_recipient_token_order_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.5780632411067194, - "step": 195, - "train_accuracy_first_token_arguments": 0.9523809523809523, - "train_accuracy_first_token_arguments_total": 21, - "train_accuracy_recipient_token": 0.9411764705882353, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 0.9285714285714286, + "train_accuracy_recipient_token_all": 0.875, + "train_accuracy_recipient_token_all_total": 8, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_list": 1.0, - "train_accuracy_recipient_token_list_total": 1, - "train_accuracy_recipient_token_stat": 1.0, - "train_accuracy_recipient_token_stat_total": 2, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 14 }, { - "epoch": 0.5810276679841897, - "grad_norm": 1.8753430945988403, - "learning_rate": 1.97393836259058e-06, - "loss": 0.2221, - "step": 196 + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.8, + "train_accuracy_recipient_token_find_total": 5, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.5810276679841897, - "step": 196, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 28, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 8, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.5810276679841897, - "step": 196, - "train_accuracy_first_token_arguments": 0.8367346938775511, - "train_accuracy_first_token_arguments_total": 49, + "epoch": 0.6640316205533597, + "step": 336, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 9, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 1, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.5810276679841897, - "step": 196, - "train_accuracy_first_token_arguments": 0.9230769230769231, - "train_accuracy_first_token_arguments_total": 13, - "train_accuracy_recipient_token": 0.8571428571428571, - "train_accuracy_recipient_token_all": 0.8, + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 0.9375, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_calculate": 1.0, - "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_monitor": 1.0, - "train_accuracy_recipient_token_monitor_total": 1, - "train_total_number_recipient_token": 14 - }, - { - "epoch": 0.58399209486166, - "grad_norm": 2.1633104244151853, - "learning_rate": 1.9504106030907605e-06, - "loss": 0.2133, - "step": 197 + "train_accuracy_recipient_token_ident": 0.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 2, + "train_total_number_recipient_token": 16 }, { - "epoch": 0.58399209486166, - "step": 197, - "train_accuracy_first_token_arguments": 0.9090909090909091, - "train_accuracy_first_token_arguments_total": 22, + "epoch": 0.6640316205533597, + "step": 336, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_construction": 1.0, - "train_accuracy_recipient_token_construction_total": 2, - "train_accuracy_recipient_token_count": 1.0, - "train_accuracy_recipient_token_count_total": 5, + "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_histor": 1.0, + "train_accuracy_recipient_token_histor_total": 3, "train_total_number_recipient_token": 18 }, { - "epoch": 0.58399209486166, - "step": 197, - "train_accuracy_first_token_arguments": 0.9411764705882353, - "train_accuracy_first_token_arguments_total": 17, + "epoch": 0.6640316205533597, + "step": 336, "train_accuracy_recipient_token": 0.8823529411764706, - "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 17 }, { - "epoch": 0.58399209486166, - "step": 197, - "train_accuracy_first_token_arguments": 0.6527777777777778, - "train_accuracy_first_token_arguments_total": 72, + "epoch": 0.6640316205533597, + "step": 336, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 1, - "train_accuracy_recipient_token_c": 1.0, - "train_accuracy_recipient_token_c_total": 1, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 4, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.5869565217391305, - "grad_norm": 2.0288710387741813, - "learning_rate": 1.9269338822822047e-06, - "loss": 0.1518, - "step": 198 + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.5869565217391305, - "step": 198, - "train_accuracy_first_token_arguments": 0.8235294117647058, - "train_accuracy_first_token_arguments_total": 17, + "epoch": 0.6640316205533597, + "step": 336, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_make": 1.0, - "train_accuracy_recipient_token_make_total": 1, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_interpret": 1.0, + "train_accuracy_recipient_token_interpret_total": 2, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.5869565217391305, - "step": 198, - "train_accuracy_first_token_arguments": 0.7777777777777778, - "train_accuracy_first_token_arguments_total": 27, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 2, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, - "train_total_number_recipient_token": 18 + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.8666666666666667, + "train_accuracy_recipient_token_all_total": 15, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_find": 0.6666666666666666, + "train_accuracy_recipient_token_find_total": 3, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.5869565217391305, - "step": 198, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 12, + "epoch": 0.6640316205533597, + "step": 336, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_check": 0.0, - "train_accuracy_recipient_token_check_total": 1, - "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_compile": 1.0, - "train_accuracy_recipient_token_compile_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_tour": 0.75, + "train_accuracy_recipient_token_tour_total": 4, "train_total_number_recipient_token": 20 }, { - "epoch": 0.5899209486166008, - "grad_norm": 2.090710210779781, - "learning_rate": 1.9035103803763793e-06, - "loss": 0.1552, - "step": 199 - }, - { - "epoch": 0.5899209486166008, - "step": 199, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 21, - "train_accuracy_recipient_token": 0.9444444444444444, + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_book": 0.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 8, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.5899209486166008, - "step": 199, - "train_accuracy_first_token_arguments": 0.8571428571428571, - "train_accuracy_first_token_arguments_total": 14, + "epoch": 0.6640316205533597, + "step": 336, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 1, - "train_accuracy_recipient_token_suggest": 1.0, - "train_accuracy_recipient_token_suggest_total": 1, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.5899209486166008, - "step": 199, - "train_accuracy_first_token_arguments": 0.8125, - "train_accuracy_first_token_arguments_total": 16, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_building": 1.0, - "train_accuracy_recipient_token_building_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_hang": 1.0, + "train_accuracy_recipient_token_hang_total": 1, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, + "train_accuracy_recipient_token_remove": 1.0, + "train_accuracy_recipient_token_remove_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_invest": 1.0, + "train_accuracy_recipient_token_invest_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.5928853754940712, - "grad_norm": 1.7471771012162318, - "learning_rate": 1.8801422726424735e-06, - "loss": 0.1951, - "step": 200 - }, - { - "epoch": 0.5928853754940712, - "step": 200, - "train_accuracy_first_token_arguments": 0.9230769230769231, - "train_accuracy_first_token_arguments_total": 13, + "epoch": 0.6640316205533597, + "step": 336, "train_accuracy_recipient_token": 0.9375, - "train_accuracy_recipient_token_air": 1.0, - "train_accuracy_recipient_token_air_total": 1, - "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_add": 1.0, + "train_accuracy_recipient_token_add_total": 1, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_construct": 1.0, + "train_accuracy_recipient_token_construct_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_select": 1.0, + "train_accuracy_recipient_token_select_total": 2, "train_total_number_recipient_token": 16 }, { - "epoch": 0.5928853754940712, - "step": 200, - "train_accuracy_first_token_arguments": 0.8888888888888888, - "train_accuracy_first_token_arguments_total": 18, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_average": 1.0, - "train_accuracy_recipient_token_average_total": 1, - "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_get": 0.5, - "train_accuracy_recipient_token_get_total": 2, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.5928853754940712, - "step": 200, - "train_accuracy_first_token_arguments": 0.8823529411764706, - "train_accuracy_first_token_arguments_total": 17, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_art": 1.0, + "train_accuracy_recipient_token_art_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_histor": 0.0, + "train_accuracy_recipient_token_histor_total": 1, + "train_accuracy_recipient_token_m": 1.0, + "train_accuracy_recipient_token_m_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_attach": 1.0, - "train_accuracy_recipient_token_attach_total": 1, - "train_accuracy_recipient_token_clean": 1.0, - "train_accuracy_recipient_token_clean_total": 1, + "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_remove": 1.0, - "train_accuracy_recipient_token_remove_total": 1, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_gr": 1.0, + "train_accuracy_recipient_token_gr_total": 3, + "train_accuracy_recipient_token_ident": 0.5, + "train_accuracy_recipient_token_ident_total": 2, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.5958498023715415, - "grad_norm": 2.1505922703967726, - "learning_rate": 1.8568317292053894e-06, - "loss": 0.2902, - "step": 201 + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 0.9285714285714286, + "train_accuracy_recipient_token_all": 0.8888888888888888, + "train_accuracy_recipient_token_all_total": 9, + "train_accuracy_recipient_token_draw": 1.0, + "train_accuracy_recipient_token_draw_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 14 }, { - "epoch": 0.5958498023715415, - "step": 201, - "train_accuracy_first_token_arguments": 0.8235294117647058, - "train_accuracy_first_token_arguments_total": 17, + "epoch": 0.6640316205533597, + "step": 336, "train_accuracy_recipient_token": 0.8333333333333334, - "train_accuracy_recipient_token_all": 0.7272727272727273, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_generate": 1.0, - "train_accuracy_recipient_token_generate_total": 1, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_review": 0.5, + "train_accuracy_recipient_token_review_total": 2, "train_total_number_recipient_token": 18 }, { - "epoch": 0.5958498023715415, - "step": 201, - "train_accuracy_first_token_arguments": 0.5925925925925926, - "train_accuracy_first_token_arguments_total": 27, - "train_accuracy_recipient_token": 0.95, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_ext": 1.0, - "train_accuracy_recipient_token_ext_total": 1, - "train_accuracy_recipient_token_fire": 1.0, - "train_accuracy_recipient_token_fire_total": 4, - "train_accuracy_recipient_token_get": 0.8, - "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 20 - }, - { - "epoch": 0.5958498023715415, - "step": 201, - "train_accuracy_first_token_arguments": 0.9032258064516129, - "train_accuracy_first_token_arguments_total": 31, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_d": 1.0, - "train_accuracy_recipient_token_d_total": 1, + "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_schedule": 1.0, @@ -11217,4063 +12865,3877 @@ "train_total_number_recipient_token": 18 }, { - "epoch": 0.5988142292490118, - "grad_norm": 1.8633933589741076, - "learning_rate": 1.8335809148442074e-06, - "loss": 0.2595, - "step": 202 + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_create": 0.5, + "train_accuracy_recipient_token_create_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.5988142292490118, - "step": 202, - "train_accuracy_first_token_arguments": 0.8888888888888888, - "train_accuracy_first_token_arguments_total": 18, + "epoch": 0.6640316205533597, + "step": 336, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 2, - "train_accuracy_recipient_token_plant": 1.0, - "train_accuracy_recipient_token_plant_total": 3, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.5988142292490118, - "step": 202, - "train_accuracy_first_token_arguments": 0.8214285714285714, - "train_accuracy_first_token_arguments_total": 28, - "train_accuracy_recipient_token": 0.8888888888888888, + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 0.875, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.8888888888888888, + "train_accuracy_recipient_token_all_total": 9, + "train_accuracy_recipient_token_custom": 1.0, + "train_accuracy_recipient_token_custom_total": 4, + "train_accuracy_recipient_token_get": 0.8333333333333334, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_convert": 1.0, - "train_accuracy_recipient_token_convert_total": 1, - "train_accuracy_recipient_token_create": 0.0, - "train_accuracy_recipient_token_create_total": 1, - "train_accuracy_recipient_token_generate": 1.0, - "train_accuracy_recipient_token_generate_total": 2, + "train_accuracy_recipient_token_eng": 1.0, + "train_accuracy_recipient_token_eng_total": 2, + "train_accuracy_recipient_token_list": 1.0, + "train_accuracy_recipient_token_list_total": 1, + "train_accuracy_recipient_token_lock": 1.0, + "train_accuracy_recipient_token_lock_total": 1, + "train_accuracy_recipient_token_tem": 1.0, + "train_accuracy_recipient_token_tem_total": 3, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 0.9375, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, - "train_total_number_recipient_token": 18 + "train_total_number_recipient_token": 16 }, { - "epoch": 0.5988142292490118, - "step": 202, - "train_accuracy_first_token_arguments": 0.52, - "train_accuracy_first_token_arguments_total": 25, - "train_accuracy_recipient_token": 0.8947368421052632, + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 12, - "train_accuracy_recipient_token_calculate": 0.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, - "train_accuracy_recipient_token_get": 0.8, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_measure": 1.0, - "train_accuracy_recipient_token_measure_total": 1, + "train_accuracy_recipient_token_car": 1.0, + "train_accuracy_recipient_token_car_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 19 }, { - "epoch": 0.6017786561264822, - "grad_norm": 2.32947277265756, - "learning_rate": 1.8103919887911525e-06, - "loss": 0.3225, - "step": 203 + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_invest": 1.0, + "train_accuracy_recipient_token_invest_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.6017786561264822, - "step": 203, - "train_accuracy_first_token_arguments": 0.8888888888888888, - "train_accuracy_first_token_arguments_total": 9, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 8, + "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { - "epoch": 0.6017786561264822, - "step": 203, - "train_accuracy_first_token_arguments": 0.9, - "train_accuracy_first_token_arguments_total": 10, - "train_accuracy_recipient_token": 0.8125, + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_bird": 1.0, + "train_accuracy_recipient_token_bird_total": 2, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_histor": 1.0, + "train_accuracy_recipient_token_histor_total": 1, + "train_accuracy_recipient_token_symbol": 1.0, + "train_accuracy_recipient_token_symbol_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 0.75, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_d": 1.0, - "train_accuracy_recipient_token_d_total": 3, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_organ": 1.0, + "train_accuracy_recipient_token_organ_total": 1, + "train_accuracy_recipient_token_rent": 1.0, + "train_accuracy_recipient_token_rent_total": 1, "train_total_number_recipient_token": 16 }, { - "epoch": 0.6017786561264822, - "step": 203, - "train_accuracy_first_token_arguments": 0.9411764705882353, - "train_accuracy_first_token_arguments_total": 17, - "train_accuracy_recipient_token": 0.9, - "train_accuracy_recipient_token_all": 0.9166666666666666, - "train_accuracy_recipient_token_all_total": 12, - "train_accuracy_recipient_token_bridge": 1.0, - "train_accuracy_recipient_token_bridge_total": 3, - "train_accuracy_recipient_token_find": 0.75, - "train_accuracy_recipient_token_find_total": 4, - "train_accuracy_recipient_token_invest": 1.0, - "train_accuracy_recipient_token_invest_total": 1, - "train_total_number_recipient_token": 20 + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_predict": 1.0, + "train_accuracy_recipient_token_predict_total": 1, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.6047430830039525, - "grad_norm": 1.8953594708836015, - "learning_rate": 1.7872671045310703e-06, - "loss": 0.33, - "step": 204 + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_evaluate": 0.0, + "train_accuracy_recipient_token_evaluate_total": 1, + "train_accuracy_recipient_token_find": 0.6666666666666666, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.6047430830039525, - "step": 204, - "train_accuracy_first_token_arguments": 0.9, - "train_accuracy_first_token_arguments_total": 10, + "epoch": 0.6640316205533597, + "step": 336, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_search": 1.0, - "train_accuracy_recipient_token_search_total": 3, - "train_accuracy_recipient_token_track": 1.0, - "train_accuracy_recipient_token_track_total": 1, + "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { - "epoch": 0.6047430830039525, - "step": 204, - "train_accuracy_first_token_arguments": 0.8235294117647058, - "train_accuracy_first_token_arguments_total": 17, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_c": 1.0, - "train_accuracy_recipient_token_c_total": 1, - "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_locate": 0.0, + "train_accuracy_recipient_token_locate_total": 1, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.6047430830039525, - "step": 204, - "train_accuracy_first_token_arguments": 0.8, - "train_accuracy_first_token_arguments_total": 20, - "train_accuracy_recipient_token": 0.8947368421052632, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_book": 0.0, - "train_accuracy_recipient_token_book_total": 1, - "train_accuracy_recipient_token_check": 1.0, - "train_accuracy_recipient_token_check_total": 1, + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 19 - }, - { - "epoch": 0.607707509881423, - "grad_norm": 1.4151900097248626, - "learning_rate": 1.7642084096014405e-06, - "loss": 0.1614, - "step": 205 + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_park": 1.0, + "train_accuracy_recipient_token_park_total": 2, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.607707509881423, - "step": 205, - "train_accuracy_first_token_arguments": 0.9705882352941176, - "train_accuracy_first_token_arguments_total": 34, - "train_accuracy_recipient_token": 0.875, + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_organ": 0.0, - "train_accuracy_recipient_token_organ_total": 1, - "train_accuracy_recipient_token_review": 0.5, - "train_accuracy_recipient_token_review_total": 2, - "train_accuracy_recipient_token_schedule": 1.0, - "train_accuracy_recipient_token_schedule_total": 1, - "train_total_number_recipient_token": 16 + "train_accuracy_recipient_token_art": 1.0, + "train_accuracy_recipient_token_art_total": 3, + "train_accuracy_recipient_token_artist": 0.0, + "train_accuracy_recipient_token_artist_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.607707509881423, - "step": 205, - "train_accuracy_first_token_arguments": 0.9166666666666666, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.9473684210526315, + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_parse": 0.5, - "train_accuracy_recipient_token_parse_total": 2, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 2, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 2, + "train_accuracy_recipient_token_object": 0.6666666666666666, + "train_accuracy_recipient_token_object_total": 3, + "train_accuracy_recipient_token_transfer": 1.0, + "train_accuracy_recipient_token_transfer_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.607707509881423, - "step": 205, - "train_accuracy_first_token_arguments": 0.9230769230769231, - "train_accuracy_first_token_arguments_total": 13, - "train_accuracy_recipient_token": 0.8823529411764706, - "train_accuracy_recipient_token_all": 0.8181818181818182, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_convert": 1.0, - "train_accuracy_recipient_token_convert_total": 2, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_fruit": 1.0, + "train_accuracy_recipient_token_fruit_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 17 }, { - "epoch": 0.6106719367588933, - "grad_norm": 2.1839121635772756, - "learning_rate": 1.7412180453929412e-06, - "loss": 0.2944, - "step": 206 - }, - { - "epoch": 0.6106719367588933, - "step": 206, - "train_accuracy_first_token_arguments": 0.8387096774193549, - "train_accuracy_first_token_arguments_total": 31, - "train_accuracy_recipient_token": 0.8421052631578947, + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compare": 0.0, - "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, - "train_total_number_recipient_token": 19 + "train_total_number_recipient_token": 18 }, { - "epoch": 0.6106719367588933, - "step": 206, - "train_accuracy_first_token_arguments": 0.9180327868852459, - "train_accuracy_first_token_arguments_total": 61, - "train_accuracy_recipient_token": 0.8571428571428571, - "train_accuracy_recipient_token_all": 0.9285714285714286, - "train_accuracy_recipient_token_all_total": 14, - "train_accuracy_recipient_token_calculate": 0.0, - "train_accuracy_recipient_token_calculate_total": 1, - "train_accuracy_recipient_token_find": 0.6666666666666666, - "train_accuracy_recipient_token_find_total": 3, + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 0.875, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, - "train_total_number_recipient_token": 21 + "train_accuracy_recipient_token_organ": 1.0, + "train_accuracy_recipient_token_organ_total": 1, + "train_total_number_recipient_token": 16 }, { - "epoch": 0.6106719367588933, - "step": 206, - "train_accuracy_first_token_arguments": 0.8695652173913043, - "train_accuracy_first_token_arguments_total": 23, - "train_accuracy_recipient_token": 0.8947368421052632, + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_average": 1.0, - "train_accuracy_recipient_token_average_total": 1, - "train_accuracy_recipient_token_cat": 0.0, - "train_accuracy_recipient_token_cat_total": 1, + "train_accuracy_recipient_token_decode": 1.0, + "train_accuracy_recipient_token_decode_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_interpret": 1.0, + "train_accuracy_recipient_token_interpret_total": 1, + "train_accuracy_recipient_token_stage": 1.0, + "train_accuracy_recipient_token_stage_total": 4, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, - "train_total_number_recipient_token": 19 - }, - { - "epoch": 0.6136363636363636, - "grad_norm": 1.7289424673499263, - "learning_rate": 1.718298146950585e-06, - "loss": 0.1488, - "step": 207 + "train_accuracy_recipient_token_rent": 1.0, + "train_accuracy_recipient_token_rent_total": 1, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.6136363636363636, - "step": 207, - "train_accuracy_first_token_arguments": 0.7272727272727273, - "train_accuracy_first_token_arguments_total": 11, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.8571428571428571, - "train_accuracy_recipient_token_all_total": 14, + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { - "epoch": 0.6136363636363636, - "step": 207, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 0.8666666666666667, - "train_accuracy_recipient_token_all": 0.8, + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, - "train_total_number_recipient_token": 15 + "train_total_number_recipient_token": 18 }, { - "epoch": 0.6136363636363636, - "step": 207, - "train_accuracy_first_token_arguments": 0.8181818181818182, - "train_accuracy_first_token_arguments_total": 11, + "epoch": 0.6640316205533597, + "step": 336, "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_design": 1.0, - "train_accuracy_recipient_token_design_total": 1, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_monitor": 1.0, - "train_accuracy_recipient_token_monitor_total": 1, - "train_accuracy_recipient_token_order": 0.0, - "train_accuracy_recipient_token_order_total": 1, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 18 }, { - "epoch": 0.616600790513834, - "grad_norm": 2.1828846473287093, - "learning_rate": 1.6954508427754435e-06, - "loss": 0.2959, - "step": 208 - }, - { - "epoch": 0.616600790513834, - "step": 208, - "train_accuracy_first_token_arguments": 0.7608695652173914, - "train_accuracy_first_token_arguments_total": 46, - "train_accuracy_recipient_token": 0.9473684210526315, + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_calculate": 0.5, - "train_accuracy_recipient_token_calculate_total": 2, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_schedule": 1.0, - "train_accuracy_recipient_token_schedule_total": 1, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_all_total": 5, + "train_accuracy_recipient_token_f": 0.0, + "train_accuracy_recipient_token_f_total": 1, + "train_accuracy_recipient_token_pest": 1.0, + "train_accuracy_recipient_token_pest_total": 1, + "train_accuracy_recipient_token_plant": 1.0, + "train_accuracy_recipient_token_plant_total": 1, + "train_accuracy_recipient_token_sun": 1.0, + "train_accuracy_recipient_token_sun_total": 1, + "train_total_number_recipient_token": 9 }, { - "epoch": 0.616600790513834, - "step": 208, - "train_accuracy_first_token_arguments": 0.8888888888888888, - "train_accuracy_first_token_arguments_total": 9, - "train_accuracy_recipient_token": 0.75, - "train_accuracy_recipient_token_all": 0.8, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 0.6666666666666666, - "train_accuracy_recipient_token_get_total": 6, - "train_total_number_recipient_token": 16 + "epoch": 0.6640316205533597, + "eval_loss": 0.1947811394929886, + "eval_runtime": 411.4932, + "eval_samples_per_second": 0.245, + "eval_steps_per_second": 0.124, + "step": 336 }, { - "epoch": 0.616600790513834, - "step": 208, - "train_accuracy_first_token_arguments": 0.95, - "train_accuracy_first_token_arguments_total": 20, - "train_accuracy_recipient_token": 0.7894736842105263, - "train_accuracy_recipient_token_all": 0.7333333333333333, - "train_accuracy_recipient_token_all_total": 15, + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 2, + "train_accuracy_recipient_token_display": 1.0, + "train_accuracy_recipient_token_display_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_s": 1.0, + "train_accuracy_recipient_token_s_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.6195652173913043, - "grad_norm": 2.3743297354062536, - "learning_rate": 1.6726782546269793e-06, - "loss": 0.253, - "step": 209 - }, - { - "epoch": 0.6195652173913043, - "step": 209, - "train_accuracy_first_token_arguments": 0.8571428571428571, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.6640316205533597, + "step": 336, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_p": 1.0, - "train_accuracy_recipient_token_p_total": 3, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, + "train_accuracy_recipient_token_class": 0.0, + "train_accuracy_recipient_token_class_total": 1, + "train_accuracy_recipient_token_decode": 1.0, + "train_accuracy_recipient_token_decode_total": 1, + "train_accuracy_recipient_token_g": 1.0, + "train_accuracy_recipient_token_g_total": 2, "train_total_number_recipient_token": 19 }, { - "epoch": 0.6195652173913043, - "step": 209, - "train_accuracy_first_token_arguments": 0.8571428571428571, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 0.9, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 0.7142857142857143, - "train_accuracy_recipient_token_get_total": 7, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 1, - "train_total_number_recipient_token": 20 + "epoch": 0.66600790513834, + "grad_norm": 1.8225855613190758, + "learning_rate": 1.3294672510993862e-06, + "loss": 0.173, + "step": 337 }, { - "epoch": 0.6195652173913043, - "step": 209, - "train_accuracy_first_token_arguments": 0.8, - "train_accuracy_first_token_arguments_total": 15, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.66600790513834, + "step": 337, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 0.8, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_locate": 1.0, - "train_accuracy_recipient_token_locate_total": 1, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 1, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { - "epoch": 0.6225296442687747, - "grad_norm": 1.6892603326816198, - "learning_rate": 1.6499824973260086e-06, - "loss": 0.2491, - "step": 210 + "epoch": 0.66600790513834, + "step": 337, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_art": 1.0, + "train_accuracy_recipient_token_art_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.8333333333333334, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.6225296442687747, - "step": 210, - "train_accuracy_first_token_arguments": 0.9230769230769231, - "train_accuracy_first_token_arguments_total": 13, - "train_accuracy_recipient_token": 0.9411764705882353, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 7, - "train_total_number_recipient_token": 17 + "epoch": 0.6679841897233202, + "grad_norm": 1.5905272243219477, + "learning_rate": 1.3153283438175036e-06, + "loss": 0.2181, + "step": 338 }, { - "epoch": 0.6225296442687747, - "step": 210, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 26, - "train_accuracy_recipient_token": 0.9473684210526315, + "epoch": 0.6679841897233202, + "step": 338, + "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_ch": 1.0, + "train_accuracy_recipient_token_ch_total": 1, + "train_accuracy_recipient_token_ex": 0.0, + "train_accuracy_recipient_token_ex_total": 1, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_validate": 1.0, - "train_accuracy_recipient_token_validate_total": 3, + "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 19 }, { - "epoch": 0.6225296442687747, - "step": 210, - "train_accuracy_first_token_arguments": 0.75, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.875, - "train_accuracy_recipient_token_all": 0.8, + "epoch": 0.6679841897233202, + "step": 338, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_compose": 1.0, + "train_accuracy_recipient_token_compose_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_total_number_recipient_token": 16 + "train_total_number_recipient_token": 19 }, { - "epoch": 0.6254940711462451, - "grad_norm": 2.3949353891643197, - "learning_rate": 1.6273656785582986e-06, - "loss": 0.3576, - "step": 211 + "epoch": 0.6699604743083004, + "grad_norm": 1.7999227337628305, + "learning_rate": 1.3012381337473656e-06, + "loss": 0.3063, + "step": 339 }, { - "epoch": 0.6254940711462451, - "step": 211, - "train_accuracy_first_token_arguments": 0.6111111111111112, - "train_accuracy_first_token_arguments_total": 18, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.6699604743083004, + "step": 339, + "train_accuracy_recipient_token": 0.8, + "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_c": 1.0, - "train_accuracy_recipient_token_c_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 8, - "train_total_number_recipient_token": 19 - }, - { - "epoch": 0.6254940711462451, - "step": 211, - "train_accuracy_first_token_arguments": 0.9166666666666666, - "train_accuracy_first_token_arguments_total": 24, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_check": 0.0, - "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_configure": 0.0, + "train_accuracy_recipient_token_configure_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_interpret": 1.0, - "train_accuracy_recipient_token_interpret_total": 1, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 15 }, { - "epoch": 0.6254940711462451, - "step": 211, - "train_accuracy_first_token_arguments": 0.75, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.6699604743083004, + "step": 339, + "train_accuracy_recipient_token": 0.8571428571428571, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 1, - "train_accuracy_recipient_token_translate": 1.0, - "train_accuracy_recipient_token_translate_total": 1, - "train_total_number_recipient_token": 18 + "train_total_number_recipient_token": 21 }, { - "epoch": 0.6284584980237155, - "grad_norm": 2.3847289223183123, - "learning_rate": 1.6048298986788345e-06, - "loss": 0.2644, - "step": 212 + "epoch": 0.6719367588932806, + "grad_norm": 1.6464971197164584, + "learning_rate": 1.2871972000823197e-06, + "loss": 0.1449, + "step": 340 }, { - "epoch": 0.6284584980237155, - "eval_accuracy_first_token_arguments": 0.923923006416132, - "eval_accuracy_first_token_arguments_total": 1091, - "eval_accuracy_recipient_token": 0.9356984478935698, - "eval_accuracy_recipient_token_all": 0.9551656920077972, - "eval_accuracy_recipient_token_all_total": 513, - "eval_accuracy_recipient_token_calculate": 1.0, - "eval_accuracy_recipient_token_calculate_total": 13, - "eval_accuracy_recipient_token_find": 0.9473684210526315, - "eval_accuracy_recipient_token_find_total": 57, - "eval_accuracy_recipient_token_get": 0.9387755102040817, - "eval_accuracy_recipient_token_get_total": 196, - "eval_accuracy_recipient_token_identify": 0.6666666666666666, - "eval_accuracy_recipient_token_identify_total": 6, - "eval_loss": 0.22125016152858734, - "eval_perplexity": 1.034604144934006, - "eval_runtime": 113.1216, - "eval_samples_per_second": 0.893, - "eval_steps_per_second": 0.451, - "eval_total_number_recipient_token": 902, - "step": 212 + "epoch": 0.6719367588932806, + "step": 340, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.6284584980237155, - "step": 212, - "train_accuracy_first_token_arguments": 0.9473684210526315, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, + "epoch": 0.6719367588932806, + "step": 340, + "train_accuracy_recipient_token": 0.7777777777777778, + "train_accuracy_recipient_token_all": 0.6666666666666666, + "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 18 }, { - "epoch": 0.6284584980237155, - "step": 212, - "train_accuracy_first_token_arguments": 0.9411764705882353, - "train_accuracy_first_token_arguments_total": 17, - "train_accuracy_recipient_token": 0.8947368421052632, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_get": 0.8333333333333334, - "train_accuracy_recipient_token_get_total": 6, - "train_total_number_recipient_token": 19 + "epoch": 0.6739130434782609, + "grad_norm": 2.125692840238993, + "learning_rate": 1.2732061199901563e-06, + "loss": 0.2194, + "step": 341 }, { - "epoch": 0.6284584980237155, - "step": 212, - "train_accuracy_first_token_arguments": 0.8333333333333334, - "train_accuracy_first_token_arguments_total": 42, - "train_accuracy_recipient_token": 0.7894736842105263, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.6739130434782609, + "step": 341, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compare": 0.0, - "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_find": 0.75, - "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_d": 1.0, + "train_accuracy_recipient_token_d_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_locate": 0.0, - "train_accuracy_recipient_token_locate_total": 1, - "train_total_number_recipient_token": 19 + "train_total_number_recipient_token": 14 }, { - "epoch": 0.6314229249011858, - "grad_norm": 1.4624962550695582, - "learning_rate": 1.58237725051677e-06, - "loss": 0.1802, - "step": 213 + "epoch": 0.6739130434782609, + "step": 341, + "train_accuracy_recipient_token": 0.875, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_custom": 1.0, + "train_accuracy_recipient_token_custom_total": 1, + "train_accuracy_recipient_token_draw": 1.0, + "train_accuracy_recipient_token_draw_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_total_number_recipient_token": 16 }, { - "epoch": 0.6314229249011858, - "step": 213, - "train_accuracy_first_token_arguments": 0.8, - "train_accuracy_first_token_arguments_total": 15, - "train_accuracy_recipient_token": 0.8823529411764706, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, + "epoch": 0.6758893280632411, + "grad_norm": 2.5717224194827053, + "learning_rate": 1.2592654685893757e-06, + "loss": 0.4011, + "step": 342 + }, + { + "epoch": 0.6758893280632411, + "step": 342, + "train_accuracy_recipient_token": 0.7894736842105263, + "train_accuracy_recipient_token_all": 0.7857142857142857, + "train_accuracy_recipient_token_all_total": 14, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_event": 0.0, + "train_accuracy_recipient_token_event_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 0.5, - "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_sum": 1.0, - "train_accuracy_recipient_token_sum_total": 2, - "train_accuracy_recipient_token_translate": 0.5, - "train_accuracy_recipient_token_translate_total": 2, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.6314229249011858, - "step": 213, - "train_accuracy_first_token_arguments": 0.9230769230769231, - "train_accuracy_first_token_arguments_total": 13, - "train_accuracy_recipient_token": 0.875, - "train_accuracy_recipient_token_all": 0.8181818181818182, - "train_accuracy_recipient_token_all_total": 11, + "epoch": 0.6758893280632411, + "step": 342, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 9, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, - "train_accuracy_recipient_token_register": 1.0, - "train_accuracy_recipient_token_register_total": 1, - "train_accuracy_recipient_token_verify": 1.0, - "train_accuracy_recipient_token_verify_total": 1, - "train_total_number_recipient_token": 16 + "train_accuracy_recipient_token_generate": 1.0, + "train_accuracy_recipient_token_generate_total": 4, + "train_accuracy_recipient_token_get": 0.5, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.6314229249011858, - "step": 213, - "train_accuracy_first_token_arguments": 0.8918918918918919, - "train_accuracy_first_token_arguments_total": 37, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_filter": 1.0, - "train_accuracy_recipient_token_filter_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 7, - "train_accuracy_recipient_token_identify": 0.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_total_number_recipient_token": 19 + "epoch": 0.6778656126482213, + "grad_norm": 1.944985236425837, + "learning_rate": 1.2453758189255568e-06, + "loss": 0.2952, + "step": 343 }, { - "epoch": 0.6343873517786561, - "grad_norm": 1.6957397755168004, - "learning_rate": 1.5600098191810682e-06, - "loss": 0.275, - "step": 214 + "epoch": 0.6778656126482213, + "step": 343, + "train_accuracy_recipient_token": 0.8571428571428571, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_custom": 1.0, + "train_accuracy_recipient_token_custom_total": 1, + "train_accuracy_recipient_token_remove": 1.0, + "train_accuracy_recipient_token_remove_total": 1, + "train_accuracy_recipient_token_translate": 0.0, + "train_accuracy_recipient_token_translate_total": 1, + "train_total_number_recipient_token": 14 }, { - "epoch": 0.6343873517786561, - "step": 214, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 28, - "train_accuracy_recipient_token": 0.9473684210526315, + "epoch": 0.6778656126482213, + "step": 343, + "train_accuracy_recipient_token": 0.8666666666666667, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_c": 0.0, + "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 6, - "train_accuracy_recipient_token_schedule": 1.0, - "train_accuracy_recipient_token_schedule_total": 1, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 1, + "train_total_number_recipient_token": 15 }, { - "epoch": 0.6343873517786561, - "step": 214, - "train_accuracy_first_token_arguments": 0.8873239436619719, - "train_accuracy_first_token_arguments_total": 71, + "epoch": 0.6798418972332015, + "grad_norm": 2.939817997079809, + "learning_rate": 1.231537741947795e-06, + "loss": 0.5102, + "step": 344 + }, + { + "epoch": 0.6798418972332015, + "step": 344, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_boat": 1.0, - "train_accuracy_recipient_token_boat_total": 3, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_suggest": 1.0, - "train_accuracy_recipient_token_suggest_total": 2, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 10, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.6343873517786561, - "step": 214, - "train_accuracy_first_token_arguments": 0.8214285714285714, - "train_accuracy_first_token_arguments_total": 28, - "train_accuracy_recipient_token": 0.9, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_check": 1.0, - "train_accuracy_recipient_token_check_total": 1, - "train_accuracy_recipient_token_hare": 0.6666666666666666, - "train_accuracy_recipient_token_hare_total": 3, - "train_accuracy_recipient_token_invest": 1.0, - "train_accuracy_recipient_token_invest_total": 1, - "train_accuracy_recipient_token_plane": 1.0, - "train_accuracy_recipient_token_plane_total": 1, + "epoch": 0.6798418972332015, + "step": 344, + "train_accuracy_recipient_token": 0.8, + "train_accuracy_recipient_token_all": 0.8333333333333334, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.5, + "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 20 }, { - "epoch": 0.6373517786561265, - "grad_norm": 1.64437830735148, - "learning_rate": 1.5377296818668638e-06, - "loss": 0.1791, - "step": 215 + "epoch": 0.6818181818181818, + "grad_norm": 1.8184116146817706, + "learning_rate": 1.217751806485235e-06, + "loss": 0.184, + "step": 345 }, { - "epoch": 0.6373517786561265, - "step": 215, - "train_accuracy_first_token_arguments": 0.7407407407407407, - "train_accuracy_first_token_arguments_total": 27, - "train_accuracy_recipient_token": 0.8125, - "train_accuracy_recipient_token_all": 0.8181818181818182, - "train_accuracy_recipient_token_all_total": 11, + "epoch": 0.6818181818181818, + "step": 345, + "train_accuracy_recipient_token": 0.9285714285714286, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_update": 0.0, - "train_accuracy_recipient_token_update_total": 1, - "train_total_number_recipient_token": 16 + "train_total_number_recipient_token": 14 }, { - "epoch": 0.6373517786561265, - "step": 215, - "train_accuracy_first_token_arguments": 0.6764705882352942, - "train_accuracy_first_token_arguments_total": 34, - "train_accuracy_recipient_token": 0.8333333333333334, + "epoch": 0.6818181818181818, + "step": 345, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_evaluate": 0.0, - "train_accuracy_recipient_token_evaluate_total": 1, - "train_accuracy_recipient_token_find": 0.3333333333333333, - "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_pack": 1.0, - "train_accuracy_recipient_token_pack_total": 1, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.6373517786561265, - "step": 215, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 15, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.6837944664031621, + "grad_norm": 1.8206375079632138, + "learning_rate": 1.2040185792236874e-06, + "loss": 0.2078, + "step": 346 + }, + { + "epoch": 0.6837944664031621, + "step": 346, + "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 2, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_ind": 1.0, - "train_accuracy_recipient_token_ind_total": 4, - "train_accuracy_recipient_token_search": 1.0, - "train_accuracy_recipient_token_search_total": 1, - "train_total_number_recipient_token": 20 - }, - { - "epoch": 0.6403162055335968, - "grad_norm": 2.485352785980747, - "learning_rate": 1.5155389076625663e-06, - "loss": 0.293, - "step": 216 + "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_total_number_recipient_token": 16 }, { - "epoch": 0.6403162055335968, - "step": 216, - "train_accuracy_first_token_arguments": 0.9473684210526315, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.8421052631578947, + "epoch": 0.6837944664031621, + "step": 346, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_architecture": 1.0, - "train_accuracy_recipient_token_architecture_total": 1, - "train_accuracy_recipient_token_compare": 0.0, - "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_get": 0.7142857142857143, - "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_decode": 1.0, + "train_accuracy_recipient_token_decode_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.6403162055335968, - "step": 216, - "train_accuracy_first_token_arguments": 0.967741935483871, - "train_accuracy_first_token_arguments_total": 31, - "train_accuracy_recipient_token": 0.85, - "train_accuracy_recipient_token_all": 0.8461538461538461, - "train_accuracy_recipient_token_all_total": 13, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 0.8333333333333334, - "train_accuracy_recipient_token_get_total": 6, - "train_total_number_recipient_token": 20 + "epoch": 0.6857707509881423, + "grad_norm": 2.216136798863525, + "learning_rate": 1.1903386246823363e-06, + "loss": 0.2829, + "step": 347 }, { - "epoch": 0.6403162055335968, - "step": 216, - "train_accuracy_first_token_arguments": 0.8421052631578947, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.95, + "epoch": 0.6857707509881423, + "step": 347, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8666666666666667, + "train_accuracy_recipient_token_all_total": 15, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.6857707509881423, + "step": 347, + "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 8, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_total_number_recipient_token": 16 }, { - "epoch": 0.6432806324110671, - "grad_norm": 1.78457120796898, - "learning_rate": 1.4934395573577016e-06, - "loss": 0.3178, - "step": 217 + "epoch": 0.6877470355731226, + "grad_norm": 2.9404057405898345, + "learning_rate": 1.1767125051905314e-06, + "loss": 0.365, + "step": 348 }, { - "epoch": 0.6432806324110671, - "step": 217, - "train_accuracy_first_token_arguments": 0.9565217391304348, - "train_accuracy_first_token_arguments_total": 23, - "train_accuracy_recipient_token": 0.8421052631578947, + "epoch": 0.6877470355731226, + "step": 348, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_draw": 1.0, + "train_accuracy_recipient_token_draw_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_ins": 1.0, + "train_accuracy_recipient_token_ins_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.6877470355731226, + "step": 348, + "train_accuracy_recipient_token": 0.9333333333333333, "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 9, - "train_accuracy_recipient_token_compare": 0.0, - "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_dec": 0.0, + "train_accuracy_recipient_token_dec_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 0.75, - "train_accuracy_recipient_token_get_total": 8, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_verify": 1.0, + "train_accuracy_recipient_token_verify_total": 1, + "train_total_number_recipient_token": 15 }, { - "epoch": 0.6432806324110671, - "step": 217, - "train_accuracy_first_token_arguments": 0.7692307692307693, - "train_accuracy_first_token_arguments_total": 13, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 0.5, - "train_accuracy_recipient_token_find_total": 2, + "epoch": 0.6897233201581028, + "grad_norm": 1.6928281922145976, + "learning_rate": 1.1631407808646758e-06, + "loss": 0.2359, + "step": 349 + }, + { + "epoch": 0.6897233201581028, + "step": 349, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_convert": 1.0, + "train_accuracy_recipient_token_convert_total": 4, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_get_total": 1, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.6432806324110671, - "step": 217, - "train_accuracy_first_token_arguments": 0.9333333333333333, - "train_accuracy_first_token_arguments_total": 15, - "train_accuracy_recipient_token": 0.9411764705882353, + "epoch": 0.6897233201581028, + "step": 349, + "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_create": 1.0, + "train_accuracy_recipient_token_create_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 6, - "train_accuracy_recipient_token_inspect": 1.0, - "train_accuracy_recipient_token_inspect_total": 1, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_get_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.6462450592885376, - "grad_norm": 2.027667840038845, - "learning_rate": 1.4714336832515386e-06, - "loss": 0.167, - "step": 218 + "epoch": 0.691699604743083, + "grad_norm": 3.6558475010700313, + "learning_rate": 1.1496240095852002e-06, + "loss": 0.197, + "step": 350 }, { - "epoch": 0.6462450592885376, - "step": 218, - "train_accuracy_first_token_arguments": 0.9285714285714286, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.691699604743083, + "step": 350, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_find": 0.8571428571428571, + "train_accuracy_recipient_token_find_total": 7, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_language": 1.0, - "train_accuracy_recipient_token_language_total": 1, - "train_accuracy_recipient_token_translate": 1.0, - "train_accuracy_recipient_token_translate_total": 4, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_get_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.6462450592885376, - "step": 218, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 18, + "epoch": 0.691699604743083, + "step": 350, "train_accuracy_recipient_token": 0.8235294117647058, - "train_accuracy_recipient_token_all": 0.8181818181818182, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_find": 0.6666666666666666, - "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analy": 0.0, + "train_accuracy_recipient_token_analy_total": 1, + "train_accuracy_recipient_token_enh": 1.0, + "train_accuracy_recipient_token_enh_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_opt": 1.0, + "train_accuracy_recipient_token_opt_total": 2, "train_total_number_recipient_token": 17 }, { - "epoch": 0.6462450592885376, - "step": 218, - "train_accuracy_first_token_arguments": 0.85, - "train_accuracy_first_token_arguments_total": 20, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_art": 1.0, - "train_accuracy_recipient_token_art_total": 4, + "epoch": 0.6936758893280632, + "grad_norm": 1.4008973035356875, + "learning_rate": 1.1361627469736286e-06, + "loss": 0.1253, + "step": 351 + }, + { + "epoch": 0.6936758893280632, + "step": 351, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.7857142857142857, + "train_accuracy_recipient_token_all_total": 14, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_find_total": 4, "train_total_number_recipient_token": 18 }, - { - "epoch": 0.6492094861660079, - "grad_norm": 1.4967382429323417, - "learning_rate": 1.449523328962496e-06, - "loss": 0.2031, - "step": 219 + { + "epoch": 0.6936758893280632, + "step": 351, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.8333333333333334, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_get": 0.8571428571428571, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.6492094861660079, - "step": 219, - "train_accuracy_first_token_arguments": 0.7692307692307693, - "train_accuracy_first_token_arguments_total": 13, - "train_accuracy_recipient_token": 0.8235294117647058, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 0.3333333333333333, - "train_accuracy_recipient_token_find_total": 3, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_search": 1.0, - "train_accuracy_recipient_token_search_total": 1, - "train_total_number_recipient_token": 17 + "epoch": 0.6956521739130435, + "grad_norm": 1.8030546844266104, + "learning_rate": 1.122757546369744e-06, + "loss": 0.2569, + "step": 352 }, { - "epoch": 0.6492094861660079, - "step": 219, - "train_accuracy_first_token_arguments": 0.8461538461538461, - "train_accuracy_first_token_arguments_total": 13, - "train_accuracy_recipient_token": 0.95, + "epoch": 0.6956521739130435, + "step": 352, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_explain": 0.8, - "train_accuracy_recipient_token_explain_total": 5, + "train_accuracy_recipient_token_all_total": 9, + "train_accuracy_recipient_token_detect": 1.0, + "train_accuracy_recipient_token_detect_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_review": 1.0, + "train_accuracy_recipient_token_review_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.6492094861660079, - "step": 219, - "train_accuracy_first_token_arguments": 0.8125, - "train_accuracy_first_token_arguments_total": 16, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.6956521739130435, + "step": 352, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_b": 1.0, - "train_accuracy_recipient_token_b_total": 1, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_join": 1.0, + "train_accuracy_recipient_token_join_total": 1, + "train_accuracy_recipient_token_vol": 1.0, + "train_accuracy_recipient_token_vol_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.6521739130434783, - "grad_norm": 1.8670991248523237, - "learning_rate": 1.4277105292383594e-06, - "loss": 0.1995, - "step": 220 + "epoch": 0.6976284584980237, + "grad_norm": 1.1757496987091924, + "learning_rate": 1.1094089588088383e-06, + "loss": 0.144, + "step": 353 }, { - "epoch": 0.6521739130434783, - "step": 220, - "train_accuracy_first_token_arguments": 0.9090909090909091, - "train_accuracy_first_token_arguments_total": 11, + "epoch": 0.6976284584980237, + "step": 353, "train_accuracy_recipient_token": 0.9375, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 8, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_inspect": 1.0, - "train_accuracy_recipient_token_inspect_total": 1, - "train_total_number_recipient_token": 16 - }, - { - "epoch": 0.6521739130434783, - "step": 220, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 10, - "train_accuracy_recipient_token": 0.8125, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 0.6666666666666666, - "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, "train_total_number_recipient_token": 16 }, { - "epoch": 0.6521739130434783, - "step": 220, - "train_accuracy_first_token_arguments": 0.8333333333333334, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.8947368421052632, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 3, - "train_accuracy_recipient_token_get": 0.8333333333333334, - "train_accuracy_recipient_token_get_total": 6, - "train_total_number_recipient_token": 19 + "epoch": 0.6976284584980237, + "step": 353, + "train_accuracy_recipient_token": 0.85, + "train_accuracy_recipient_token_all": 0.7692307692307693, + "train_accuracy_recipient_token_all_total": 13, + "train_accuracy_recipient_token_detect": 1.0, + "train_accuracy_recipient_token_detect_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 1, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.6551383399209486, - "grad_norm": 2.550246191316389, - "learning_rate": 1.4059973097673187e-06, - "loss": 0.2752, - "step": 221 + "epoch": 0.6996047430830039, + "grad_norm": 1.7957489707100158, + "learning_rate": 1.0961175329990631e-06, + "loss": 0.2026, + "step": 354 }, { - "epoch": 0.6551383399209486, - "step": 221, - "train_accuracy_first_token_arguments": 0.9333333333333333, - "train_accuracy_first_token_arguments_total": 15, + "epoch": 0.6996047430830039, + "step": 354, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_invest": 1.0, - "train_accuracy_recipient_token_invest_total": 1, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_opt": 1.0, + "train_accuracy_recipient_token_opt_total": 1, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.6551383399209486, - "step": 221, - "train_accuracy_first_token_arguments": 0.625, - "train_accuracy_first_token_arguments_total": 24, - "train_accuracy_recipient_token": 0.8888888888888888, + "epoch": 0.6996047430830039, + "step": 354, + "train_accuracy_recipient_token": 0.8571428571428571, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_design": 1.0, + "train_accuracy_recipient_token_design_total": 1, + "train_accuracy_recipient_token_im": 1.0, + "train_accuracy_recipient_token_im_total": 1, + "train_accuracy_recipient_token_install": 0.0, + "train_accuracy_recipient_token_install_total": 1, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 14 + }, + { + "epoch": 0.7015810276679841, + "grad_norm": 1.7932806243824873, + "learning_rate": 1.082883815298876e-06, + "loss": 0.2821, + "step": 355 + }, + { + "epoch": 0.7015810276679841, + "step": 355, + "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_load": 1.0, - "train_accuracy_recipient_token_load_total": 1, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.6551383399209486, - "step": 221, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 9, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.7015810276679841, + "step": 355, + "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.7142857142857143, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { - "epoch": 0.658102766798419, - "grad_norm": 1.8305223962446249, - "learning_rate": 1.3843856869898486e-06, - "loss": 0.2692, - "step": 222 + "epoch": 0.7035573122529645, + "grad_norm": 2.0800060452488855, + "learning_rate": 1.0697083496945766e-06, + "loss": 0.2197, + "step": 356 }, { - "epoch": 0.658102766798419, - "step": 222, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 17, - "train_accuracy_recipient_token": 0.8823529411764706, - "train_accuracy_recipient_token_all": 0.8, + "epoch": 0.7035573122529645, + "step": 356, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_device": 1.0, - "train_accuracy_recipient_token_device_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_total_number_recipient_token": 17 - }, - { - "epoch": 0.658102766798419, - "step": 222, - "train_accuracy_first_token_arguments": 0.8823529411764706, - "train_accuracy_first_token_arguments_total": 17, - "train_accuracy_recipient_token": 0.95, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 0.9, - "train_accuracy_recipient_token_get_total": 10, + "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 20 }, { - "epoch": 0.658102766798419, - "step": 222, - "train_accuracy_first_token_arguments": 0.88, - "train_accuracy_first_token_arguments_total": 50, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.8, + "epoch": 0.7035573122529645, + "step": 356, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_generate": 1.0, + "train_accuracy_recipient_token_generate_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, - "train_accuracy_recipient_token_park": 1.0, - "train_accuracy_recipient_token_park_total": 2, - "train_accuracy_recipient_token_validate": 1.0, - "train_accuracy_recipient_token_validate_total": 2, - "train_total_number_recipient_token": 18 + "train_total_number_recipient_token": 19 }, { - "epoch": 0.6610671936758893, - "grad_norm": 1.7979076335597703, - "learning_rate": 1.3628776679114516e-06, - "loss": 0.1334, - "step": 223 + "epoch": 0.7055335968379447, + "grad_norm": 1.460137074708585, + "learning_rate": 1.0565916777779519e-06, + "loss": 0.1778, + "step": 357 }, { - "epoch": 0.6610671936758893, - "step": 223, - "train_accuracy_first_token_arguments": 0.7894736842105263, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.8666666666666667, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_check": 0.0, - "train_accuracy_recipient_token_check_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "epoch": 0.7055335968379447, + "step": 357, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.7055335968379447, + "step": 357, + "train_accuracy_recipient_token": 0.9333333333333333, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, + "train_accuracy_recipient_token_count": 1.0, + "train_accuracy_recipient_token_count_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_track": 1.0, + "train_accuracy_recipient_token_track_total": 1, "train_total_number_recipient_token": 15 }, { - "epoch": 0.6610671936758893, - "step": 223, - "train_accuracy_first_token_arguments": 0.5, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.95, + "epoch": 0.7075098814229249, + "grad_norm": 2.634085943263362, + "learning_rate": 1.04353433872401e-06, + "loss": 0.487, + "step": 358 + }, + { + "epoch": 0.7075098814229249, + "step": 358, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 12, - "train_accuracy_recipient_token_explain": 1.0, - "train_accuracy_recipient_token_explain_total": 1, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_translate": 0.0, - "train_accuracy_recipient_token_translate_total": 1, - "train_total_number_recipient_token": 20 + "train_total_number_recipient_token": 16 }, { - "epoch": 0.6610671936758893, - "step": 223, - "train_accuracy_first_token_arguments": 0.9210526315789473, - "train_accuracy_first_token_arguments_total": 38, - "train_accuracy_recipient_token": 0.9411764705882353, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, + "epoch": 0.7075098814229249, + "step": 358, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_local": 1.0, - "train_accuracy_recipient_token_local_total": 1, - "train_accuracy_recipient_token_plan": 1.0, - "train_accuracy_recipient_token_plan_total": 1, - "train_accuracy_recipient_token_restaurant": 1.0, - "train_accuracy_recipient_token_restaurant_total": 2, - "train_total_number_recipient_token": 17 - }, - { - "epoch": 0.6640316205533597, - "grad_norm": 2.7045237044612875, - "learning_rate": 1.3414752499162676e-06, - "loss": 0.3284, - "step": 224 + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.6640316205533597, - "step": 224, - "train_accuracy_first_token_arguments": 0.9310344827586207, - "train_accuracy_first_token_arguments_total": 29, - "train_accuracy_recipient_token": 0.8947368421052632, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, + "epoch": 0.7094861660079052, + "grad_norm": 1.699389503506117, + "learning_rate": 1.0305368692688175e-06, + "loss": 0.2116, + "step": 359 + }, + { + "epoch": 0.7094861660079052, + "step": 359, + "train_accuracy_recipient_token": 0.9333333333333333, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 2, - "train_accuracy_recipient_token_display": 1.0, - "train_accuracy_recipient_token_display_total": 1, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_suggest": 0.0, - "train_accuracy_recipient_token_suggest_total": 1, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_merge": 1.0, + "train_accuracy_recipient_token_merge_total": 1, + "train_total_number_recipient_token": 15 }, { - "epoch": 0.6640316205533597, - "step": 224, - "train_accuracy_first_token_arguments": 0.8333333333333334, - "train_accuracy_first_token_arguments_total": 18, + "epoch": 0.7094861660079052, + "step": 359, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 1, - "train_accuracy_recipient_token_classify": 0.0, - "train_accuracy_recipient_token_classify_total": 1, - "train_accuracy_recipient_token_decode": 1.0, - "train_accuracy_recipient_token_decode_total": 1, - "train_accuracy_recipient_token_g": 1.0, - "train_accuracy_recipient_token_g_total": 2, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 4, "train_total_number_recipient_token": 19 }, { - "epoch": 0.6640316205533597, - "step": 224, - "train_accuracy_first_token_arguments": 0.7692307692307693, - "train_accuracy_first_token_arguments_total": 13, + "epoch": 0.7114624505928854, + "grad_norm": 2.1791368035785204, + "learning_rate": 1.0175998036874357e-06, + "loss": 0.3231, + "step": 360 + }, + { + "epoch": 0.7114624505928854, + "step": 360, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_check": 1.0, - "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_order": 1.0, + "train_accuracy_recipient_token_order_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.7114624505928854, + "step": 360, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.6666666666666666, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_ins": 1.0, + "train_accuracy_recipient_token_ins_total": 1, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.6669960474308301, - "grad_norm": 1.8337644487638454, - "learning_rate": 1.3201804205815872e-06, - "loss": 0.2451, - "step": 225 + "epoch": 0.7134387351778656, + "grad_norm": 1.5860150416516434, + "learning_rate": 1.00472367377196e-06, + "loss": 0.1088, + "step": 361 }, { - "epoch": 0.6669960474308301, - "step": 225, - "train_accuracy_first_token_arguments": 0.9523809523809523, - "train_accuracy_first_token_arguments_total": 21, + "epoch": 0.7134387351778656, + "step": 361, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_count": 1.0, + "train_accuracy_recipient_token_count_total": 1, + "train_accuracy_recipient_token_d": 1.0, + "train_accuracy_recipient_token_d_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_vehicle": 0.6666666666666666, + "train_accuracy_recipient_token_vehicle_total": 3, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.7134387351778656, + "step": 361, "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_art": 1.0, - "train_accuracy_recipient_token_art_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.6669960474308301, - "step": 225, - "train_accuracy_first_token_arguments": 0.9166666666666666, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.8947368421052632, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_christ": 1.0, - "train_accuracy_recipient_token_christ_total": 1, - "train_accuracy_recipient_token_explain": 0.0, - "train_accuracy_recipient_token_explain_total": 1, + "epoch": 0.7154150197628458, + "grad_norm": 2.178306730601671, + "learning_rate": 9.91909008809659e-07, + "loss": 0.2397, + "step": 362 + }, + { + "epoch": 0.7154150197628458, + "step": 362, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.6669960474308301, - "step": 225, - "train_accuracy_first_token_arguments": 0.8333333333333334, - "train_accuracy_first_token_arguments_total": 12, + "epoch": 0.7154150197628458, + "step": 362, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compose": 1.0, - "train_accuracy_recipient_token_compose_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 3, - "train_accuracy_recipient_token_get": 0.8, - "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 2, + "train_accuracy_recipient_token_building": 1.0, + "train_accuracy_recipient_token_building_total": 3, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_object": 0.6666666666666666, + "train_accuracy_recipient_token_object_total": 3, "train_total_number_recipient_token": 19 }, { - "epoch": 0.6699604743083004, - "grad_norm": 1.601543108317484, - "learning_rate": 1.2989951574932693e-06, - "loss": 0.2814, - "step": 226 + "epoch": 0.717391304347826, + "grad_norm": 2.3506607158415993, + "learning_rate": 9.791563355612171e-07, + "loss": 0.3363, + "step": 363 }, { - "epoch": 0.6699604743083004, - "step": 226, - "train_accuracy_first_token_arguments": 0.967741935483871, - "train_accuracy_first_token_arguments_total": 31, - "train_accuracy_recipient_token": 0.8, - "train_accuracy_recipient_token_all": 0.8, + "epoch": 0.717391304347826, + "step": 363, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_configure": 0.0, - "train_accuracy_recipient_token_configure_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_total_number_recipient_token": 15 + "train_accuracy_recipient_token_get": 0.8888888888888888, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.6699604743083004, - "step": 226, - "train_accuracy_first_token_arguments": 0.8888888888888888, - "train_accuracy_first_token_arguments_total": 18, - "train_accuracy_recipient_token": 0.9047619047619048, + "epoch": 0.717391304347826, + "step": 363, + "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 0.6666666666666666, - "train_accuracy_recipient_token_find_total": 6, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 21 + "train_accuracy_recipient_token_join": 0.0, + "train_accuracy_recipient_token_join_total": 1, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.6699604743083004, - "step": 226, - "train_accuracy_first_token_arguments": 0.9333333333333333, - "train_accuracy_first_token_arguments_total": 30, - "train_accuracy_recipient_token": 0.9, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_calculate": 0.75, - "train_accuracy_recipient_token_calculate_total": 4, - "train_accuracy_recipient_token_get": 1.0, + "epoch": 0.7193675889328063, + "grad_norm": 2.4377826724234533, + "learning_rate": 9.66466178239084e-07, + "loss": 0.3347, + "step": 364 + }, + { + "epoch": 0.7193675889328063, + "step": 364, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compile": 1.0, + "train_accuracy_recipient_token_compile_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.6, "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 20 + "train_total_number_recipient_token": 17 }, { - "epoch": 0.6729249011857708, - "grad_norm": 1.9791094562714513, - "learning_rate": 1.277921428062091e-06, - "loss": 0.2229, - "step": 227 + "epoch": 0.7193675889328063, + "step": 364, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.7142857142857143, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.6729249011857708, - "step": 227, - "train_accuracy_first_token_arguments": 0.9629629629629629, - "train_accuracy_first_token_arguments_total": 54, - "train_accuracy_recipient_token": 0.7777777777777778, - "train_accuracy_recipient_token_all": 0.6666666666666666, - "train_accuracy_recipient_token_all_total": 12, + "epoch": 0.7213438735177866, + "grad_norm": 2.1195801120949644, + "learning_rate": 9.538390584859213e-07, + "loss": 0.2291, + "step": 365 + }, + { + "epoch": 0.7213438735177866, + "step": 365, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_order": 1.0, + "train_accuracy_recipient_token_order_total": 1, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.6729249011857708, - "step": 227, - "train_accuracy_first_token_arguments": 0.7777777777777778, - "train_accuracy_first_token_arguments_total": 9, + "epoch": 0.7213438735177866, + "step": 365, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_d": 1.0, - "train_accuracy_recipient_token_d_total": 1, + "train_accuracy_recipient_token_exp": 1.0, + "train_accuracy_recipient_token_exp_total": 4, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_total_number_recipient_token": 14 + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.6729249011857708, - "step": 227, - "train_accuracy_first_token_arguments": 0.6862745098039216, - "train_accuracy_first_token_arguments_total": 51, - "train_accuracy_recipient_token": 0.9375, + "epoch": 0.7233201581027668, + "grad_norm": 1.4023845894942437, + "learning_rate": 9.412754953531664e-07, + "loss": 0.1807, + "step": 366 + }, + { + "epoch": 0.7233201581027668, + "step": 366, + "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_customize": 1.0, - "train_accuracy_recipient_token_customize_total": 1, + "train_accuracy_recipient_token_create": 1.0, + "train_accuracy_recipient_token_create_total": 1, + "train_accuracy_recipient_token_design": 1.0, + "train_accuracy_recipient_token_design_total": 2, "train_accuracy_recipient_token_draw": 1.0, - "train_accuracy_recipient_token_draw_total": 2, + "train_accuracy_recipient_token_draw_total": 1, + "train_accuracy_recipient_token_evaluate": 0.0, + "train_accuracy_recipient_token_evaluate_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.7233201581027668, + "step": 366, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, - "train_total_number_recipient_token": 16 + "train_accuracy_recipient_token_list": 1.0, + "train_accuracy_recipient_token_list_total": 3, + "train_accuracy_recipient_token_use": 1.0, + "train_accuracy_recipient_token_use_total": 1, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.6758893280632411, - "grad_norm": 2.308708795701363, - "learning_rate": 1.2569611893410374e-06, - "loss": 0.3605, - "step": 228 + "epoch": 0.7252964426877471, + "grad_norm": 1.6925202838713784, + "learning_rate": 9.28776005279691e-07, + "loss": 0.2561, + "step": 367 }, { - "epoch": 0.6758893280632411, - "step": 228, - "train_accuracy_first_token_arguments": 0.6666666666666666, - "train_accuracy_first_token_arguments_total": 21, - "train_accuracy_recipient_token": 0.7894736842105263, - "train_accuracy_recipient_token_all": 0.7142857142857143, - "train_accuracy_recipient_token_all_total": 14, - "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_event": 1.0, - "train_accuracy_recipient_token_event_total": 1, + "epoch": 0.7252964426877471, + "step": 367, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_design": 1.0, + "train_accuracy_recipient_token_design_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 1, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.7252964426877471, + "step": 367, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_create": 0.5, + "train_accuracy_recipient_token_create_total": 2, + "train_accuracy_recipient_token_generate": 1.0, + "train_accuracy_recipient_token_generate_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.7272727272727273, + "grad_norm": 1.4692209621399164, + "learning_rate": 9.163411020705762e-07, + "loss": 0.2018, + "step": 368 }, { - "epoch": 0.6758893280632411, - "step": 228, - "train_accuracy_first_token_arguments": 0.7619047619047619, - "train_accuracy_first_token_arguments_total": 21, - "train_accuracy_recipient_token": 0.95, + "epoch": 0.7272727272727273, + "step": 368, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_find": 0.5, - "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_generate": 1.0, - "train_accuracy_recipient_token_generate_total": 5, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.7272727272727273, + "step": 368, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9166666666666666, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, - "train_total_number_recipient_token": 20 + "train_total_number_recipient_token": 17 }, { - "epoch": 0.6758893280632411, - "step": 228, - "train_accuracy_first_token_arguments": 0.7, - "train_accuracy_first_token_arguments_total": 10, - "train_accuracy_recipient_token": 0.8571428571428571, + "epoch": 0.7292490118577075, + "grad_norm": 1.8701554324040663, + "learning_rate": 9.039712968759864e-07, + "loss": 0.2345, + "step": 369 + }, + { + "epoch": 0.7292490118577075, + "step": 369, + "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_customize": 1.0, - "train_accuracy_recipient_token_customize_total": 1, - "train_accuracy_recipient_token_remove": 1.0, - "train_accuracy_recipient_token_remove_total": 1, - "train_accuracy_recipient_token_translate": 0.0, - "train_accuracy_recipient_token_translate_total": 1, - "train_total_number_recipient_token": 14 + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.6788537549407114, - "grad_norm": 2.4044180995153486, - "learning_rate": 1.2361163878435594e-06, - "loss": 0.4432, - "step": 229 + "epoch": 0.7292490118577075, + "step": 369, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 2, + "train_accuracy_recipient_token_inventory": 0.0, + "train_accuracy_recipient_token_inventory_total": 1, + "train_accuracy_recipient_token_sales": 1.0, + "train_accuracy_recipient_token_sales_total": 1, + "train_accuracy_recipient_token_shelf": 1.0, + "train_accuracy_recipient_token_shelf_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.6788537549407114, - "step": 229, - "train_accuracy_first_token_arguments": 0.9615384615384616, - "train_accuracy_first_token_arguments_total": 26, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_c": 1.0, - "train_accuracy_recipient_token_c_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_translate": 1.0, - "train_accuracy_recipient_token_translate_total": 1, - "train_total_number_recipient_token": 15 + "epoch": 0.7312252964426877, + "grad_norm": 1.9983260849142814, + "learning_rate": 8.916670981701656e-07, + "loss": 0.2475, + "step": 370 }, { - "epoch": 0.6788537549407114, - "step": 229, - "train_accuracy_first_token_arguments": 0.9230769230769231, - "train_accuracy_first_token_arguments_total": 13, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.7312252964426877, + "step": 370, + "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_buy": 0.0, + "train_accuracy_recipient_token_buy_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 7, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 10, + "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 20 }, { - "epoch": 0.6788537549407114, - "step": 229, - "train_accuracy_first_token_arguments": 0.7666666666666667, - "train_accuracy_first_token_arguments_total": 30, - "train_accuracy_recipient_token": 0.85, - "train_accuracy_recipient_token_all": 0.75, - "train_accuracy_recipient_token_all_total": 12, - "train_accuracy_recipient_token_book": 1.0, - "train_accuracy_recipient_token_book_total": 1, - "train_accuracy_recipient_token_c": 1.0, - "train_accuracy_recipient_token_c_total": 1, + "epoch": 0.7312252964426877, + "step": 370, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_locate": 1.0, + "train_accuracy_recipient_token_locate_total": 1, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 1, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.6818181818181818, - "grad_norm": 2.0846069978627417, - "learning_rate": 1.2153889593628032e-06, - "loss": 0.2884, - "step": 230 + "epoch": 0.733201581027668, + "grad_norm": 1.6154324345581008, + "learning_rate": 8.794290117305296e-07, + "loss": 0.1644, + "step": 371 }, { - "epoch": 0.6818181818181818, - "step": 230, - "train_accuracy_first_token_arguments": 0.5185185185185185, - "train_accuracy_first_token_arguments_total": 27, - "train_accuracy_recipient_token": 0.8571428571428571, - "train_accuracy_recipient_token_all": 0.8, + "epoch": 0.733201581027668, + "step": 371, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_total_number_recipient_token": 14 + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_m": 1.0, + "train_accuracy_recipient_token_m_total": 4, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.6818181818181818, - "step": 230, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 15, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, + "epoch": 0.733201581027668, + "step": 371, + "train_accuracy_recipient_token": 0.7777777777777778, + "train_accuracy_recipient_token_all": 0.7857142857142857, + "train_accuracy_recipient_token_all_total": 14, + "train_accuracy_recipient_token_calculate": 0.0, + "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 9, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.6818181818181818, - "step": 230, - "train_accuracy_first_token_arguments": 0.6060606060606061, - "train_accuracy_first_token_arguments_total": 33, + "epoch": 0.7351778656126482, + "grad_norm": 1.5991587060424923, + "learning_rate": 8.672575406168782e-07, + "loss": 0.1712, + "step": 372 + }, + { + "epoch": 0.7351778656126482, + "step": 372, "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_wine": 1.0, - "train_accuracy_recipient_token_wine_total": 2, "train_total_number_recipient_token": 18 }, { - "epoch": 0.6847826086956522, - "grad_norm": 1.903972306272471, - "learning_rate": 1.1947808287918406e-06, - "loss": 0.233, - "step": 231 + "epoch": 0.7351778656126482, + "step": 372, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 0.6, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.6847826086956522, - "step": 231, - "train_accuracy_first_token_arguments": 0.75, - "train_accuracy_first_token_arguments_total": 24, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.7371541501976284, + "grad_norm": 1.660305490888912, + "learning_rate": 8.551531851507186e-07, + "loss": 0.2554, + "step": 373 + }, + { + "epoch": 0.7371541501976284, + "step": 373, + "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_check": 1.0, - "train_accuracy_recipient_token_check_total": 1, - "train_accuracy_recipient_token_decode": 1.0, - "train_accuracy_recipient_token_decode_total": 1, + "train_accuracy_recipient_token_all_total": 14, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 6, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_translate": 0.0, + "train_accuracy_recipient_token_translate_total": 1, + "train_accuracy_recipient_token_vehicle": 1.0, + "train_accuracy_recipient_token_vehicle_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.6847826086956522, - "step": 231, - "train_accuracy_first_token_arguments": 0.9411764705882353, - "train_accuracy_first_token_arguments_total": 51, - "train_accuracy_recipient_token": 0.7647058823529411, - "train_accuracy_recipient_token_all": 0.7333333333333333, - "train_accuracy_recipient_token_all_total": 15, + "epoch": 0.7371541501976284, + "step": 373, + "train_accuracy_recipient_token": 0.9375, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_construction": 1.0, + "train_accuracy_recipient_token_construction_total": 2, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.7391304347826086, + "grad_norm": 2.4165564387939567, + "learning_rate": 8.431164428946928e-07, + "loss": 0.3098, + "step": 374 + }, + { + "epoch": 0.7391304347826086, + "step": 374, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.7142857142857143, + "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { - "epoch": 0.6847826086956522, - "step": 231, - "train_accuracy_first_token_arguments": 0.76, - "train_accuracy_first_token_arguments_total": 25, - "train_accuracy_recipient_token": 0.875, - "train_accuracy_recipient_token_all": 0.8181818181818182, - "train_accuracy_recipient_token_all_total": 11, + "epoch": 0.7391304347826086, + "step": 374, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_total_number_recipient_token": 16 + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.6877470355731226, - "grad_norm": 3.071022026928369, - "learning_rate": 1.1742939099449126e-06, - "loss": 0.3897, - "step": 232 + "epoch": 0.741106719367589, + "grad_norm": 1.801628808429806, + "learning_rate": 8.311478086321301e-07, + "loss": 0.1757, + "step": 375 }, { - "epoch": 0.6877470355731226, - "step": 232, - "train_accuracy_first_token_arguments": 0.6440677966101694, - "train_accuracy_first_token_arguments_total": 59, - "train_accuracy_recipient_token": 0.9411764705882353, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_draw": 1.0, - "train_accuracy_recipient_token_draw_total": 1, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_inspect": 1.0, - "train_accuracy_recipient_token_inspect_total": 1, - "train_total_number_recipient_token": 17 + "epoch": 0.741106719367589, + "step": 375, + "train_accuracy_recipient_token": 0.8, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 5, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 10 }, { - "epoch": 0.6877470355731226, - "step": 232, - "train_accuracy_first_token_arguments": 0.5454545454545454, - "train_accuracy_first_token_arguments_total": 33, + "epoch": 0.741106719367589, + "step": 375, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_dec": 1.0, - "train_accuracy_recipient_token_dec_total": 1, + "train_accuracy_recipient_token_animal": 1.0, + "train_accuracy_recipient_token_animal_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_verify": 1.0, - "train_accuracy_recipient_token_verify_total": 1, - "train_total_number_recipient_token": 15 + "train_accuracy_recipient_token_locate": 1.0, + "train_accuracy_recipient_token_locate_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.6877470355731226, - "step": 232, - "train_accuracy_first_token_arguments": 0.8571428571428571, - "train_accuracy_first_token_arguments_total": 21, - "train_accuracy_recipient_token": 0.9411764705882353, + "epoch": 0.7430830039525692, + "grad_norm": 12.540187334844513, + "learning_rate": 8.192477743467078e-07, + "loss": 0.7105, + "step": 376 + }, + { + "epoch": 0.7430830039525692, + "step": 376, + "train_accuracy_recipient_token": 0.8571428571428571, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_convert": 1.0, - "train_accuracy_recipient_token_convert_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.5, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 14 + }, + { + "epoch": 0.7430830039525692, + "step": 376, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, + "train_accuracy_recipient_token_compose": 1.0, + "train_accuracy_recipient_token_compose_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.6907114624505929, - "grad_norm": 2.858676916448621, - "learning_rate": 1.153930105379695e-06, - "loss": 0.2713, - "step": 233 + "epoch": 0.7450592885375494, + "grad_norm": 1.9972903002700673, + "learning_rate": 8.074168292022269e-07, + "loss": 0.25, + "step": 377 }, { - "epoch": 0.6907114624505929, - "step": 233, - "train_accuracy_first_token_arguments": 0.9411764705882353, - "train_accuracy_first_token_arguments_total": 17, + "epoch": 0.7450592885375494, + "step": 377, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_check": 1.0, - "train_accuracy_recipient_token_check_total": 1, - "train_accuracy_recipient_token_create": 1.0, - "train_accuracy_recipient_token_create_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { - "epoch": 0.6907114624505929, - "step": 233, - "train_accuracy_first_token_arguments": 0.9459459459459459, - "train_accuracy_first_token_arguments_total": 37, - "train_accuracy_recipient_token": 0.9444444444444444, + "epoch": 0.7450592885375494, + "step": 377, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 0.8571428571428571, - "train_accuracy_recipient_token_find_total": 7, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, - "train_total_number_recipient_token": 18 - }, - { - "epoch": 0.6907114624505929, - "step": 233, - "train_accuracy_first_token_arguments": 0.8260869565217391, - "train_accuracy_first_token_arguments_total": 23, - "train_accuracy_recipient_token": 0.9411764705882353, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 1, - "train_accuracy_recipient_token_enh": 1.0, - "train_accuracy_recipient_token_enh_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_optimize": 1.0, - "train_accuracy_recipient_token_optimize_total": 2, - "train_total_number_recipient_token": 17 - }, - { - "epoch": 0.6936758893280632, - "grad_norm": 1.4873714467096886, - "learning_rate": 1.1336913062206157e-06, - "loss": 0.1483, - "step": 234 + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.6936758893280632, - "step": 234, - "train_accuracy_first_token_arguments": 0.9, - "train_accuracy_first_token_arguments_total": 20, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.8571428571428571, - "train_accuracy_recipient_token_all_total": 14, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 4, - "train_total_number_recipient_token": 18 + "epoch": 0.7470355731225297, + "grad_norm": 2.3216872504414847, + "learning_rate": 7.956554595225017e-07, + "loss": 0.2433, + "step": 378 }, { - "epoch": 0.6936758893280632, - "step": 234, - "train_accuracy_first_token_arguments": 0.9090909090909091, - "train_accuracy_first_token_arguments_total": 11, - "train_accuracy_recipient_token": 0.8947368421052632, + "epoch": 0.7470355731225297, + "step": 378, + "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8333333333333334, "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 7, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_get_total": 1, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.6936758893280632, - "step": 234, - "train_accuracy_first_token_arguments": 0.8333333333333334, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.7470355731225297, + "step": 378, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_detect": 1.0, - "train_accuracy_recipient_token_detect_total": 1, + "train_accuracy_recipient_token_evaluate": 0.6666666666666666, + "train_accuracy_recipient_token_evaluate_total": 3, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_review": 1.0, - "train_accuracy_recipient_token_review_total": 1, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.6966403162055336, - "grad_norm": 1.5877712208219548, - "learning_rate": 1.1135793919832336e-06, - "loss": 0.2483, - "step": 235 + "epoch": 0.7490118577075099, + "grad_norm": 1.7733742585866188, + "learning_rate": 7.839641487713745e-07, + "loss": 0.2184, + "step": 379 }, { - "epoch": 0.6966403162055336, - "step": 235, - "train_accuracy_first_token_arguments": 0.9411764705882353, - "train_accuracy_first_token_arguments_total": 17, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.7490118577075099, + "step": 379, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_join": 1.0, - "train_accuracy_recipient_token_join_total": 1, - "train_accuracy_recipient_token_vol": 1.0, - "train_accuracy_recipient_token_vol_total": 1, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_te": 1.0, + "train_accuracy_recipient_token_te_total": 1, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.6966403162055336, - "step": 235, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 11, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 9, + "epoch": 0.7490118577075099, + "step": 379, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_evaluate": 0.0, + "train_accuracy_recipient_token_evaluate_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 3, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 1, - "train_total_number_recipient_token": 17 - }, - { - "epoch": 0.6966403162055336, - "step": 235, - "train_accuracy_first_token_arguments": 0.9047619047619048, - "train_accuracy_first_token_arguments_total": 21, - "train_accuracy_recipient_token": 0.85, - "train_accuracy_recipient_token_all": 0.7692307692307693, - "train_accuracy_recipient_token_all_total": 13, - "train_accuracy_recipient_token_detect": 1.0, - "train_accuracy_recipient_token_detect_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_validate": 1.0, - "train_accuracy_recipient_token_validate_total": 1, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_s": 1.0, + "train_accuracy_recipient_token_s_total": 2, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.6996047430830039, - "grad_norm": 1.392788536484163, - "learning_rate": 1.09359623039969e-06, - "loss": 0.1911, - "step": 236 + "epoch": 0.7509881422924901, + "grad_norm": 1.5040141373589382, + "learning_rate": 7.723433775328385e-07, + "loss": 0.1427, + "step": 380 }, { - "epoch": 0.6996047430830039, - "step": 236, - "train_accuracy_first_token_arguments": 0.9090909090909091, - "train_accuracy_first_token_arguments_total": 22, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.7509881422924901, + "step": 380, + "train_accuracy_recipient_token": 0.7777777777777778, + "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 0.0, + "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_review": 1.0, + "train_accuracy_recipient_token_review_total": 1, + "train_accuracy_recipient_token_stamp": 1.0, + "train_accuracy_recipient_token_stamp_total": 4, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.7509881422924901, + "step": 380, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_convert": 1.0, + "train_accuracy_recipient_token_convert_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_optimize": 0.0, - "train_accuracy_recipient_token_optimize_total": 1, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 1, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_sum": 1.0, + "train_accuracy_recipient_token_sum_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.6996047430830039, - "step": 236, - "train_accuracy_first_token_arguments": 0.8108108108108109, - "train_accuracy_first_token_arguments_total": 37, - "train_accuracy_recipient_token": 0.7857142857142857, - "train_accuracy_recipient_token_all": 0.8, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_design": 1.0, - "train_accuracy_recipient_token_design_total": 1, - "train_accuracy_recipient_token_im": 1.0, - "train_accuracy_recipient_token_im_total": 1, - "train_accuracy_recipient_token_install": 0.0, - "train_accuracy_recipient_token_install_total": 1, - "train_accuracy_recipient_token_schedule": 1.0, - "train_accuracy_recipient_token_schedule_total": 1, - "train_total_number_recipient_token": 14 + "epoch": 0.7529644268774703, + "grad_norm": 2.2982731810128127, + "learning_rate": 7.607936234912841e-07, + "loss": 0.3524, + "step": 381 }, { - "epoch": 0.6996047430830039, - "step": 236, - "train_accuracy_first_token_arguments": 0.8695652173913043, - "train_accuracy_first_token_arguments_total": 23, + "epoch": 0.7529644268774703, + "step": 381, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_calculate": 1.0, - "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 2, + "train_accuracy_recipient_token_decode": 1.0, + "train_accuracy_recipient_token_decode_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 18 }, { - "epoch": 0.7025691699604744, - "grad_norm": 1.9185387350295975, - "learning_rate": 1.0737436772452602e-06, - "loss": 0.2789, - "step": 237 + "epoch": 0.7529644268774703, + "step": 381, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_adjust": 1.0, + "train_accuracy_recipient_token_adjust_total": 1, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_dis": 1.0, + "train_accuracy_recipient_token_dis_total": 1, + "train_accuracy_recipient_token_enh": 1.0, + "train_accuracy_recipient_token_enh_total": 1, + "train_accuracy_recipient_token_event": 1.0, + "train_accuracy_recipient_token_event_total": 4, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.7025691699604744, - "step": 237, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 17, - "train_accuracy_recipient_token": 0.8235294117647058, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 0.7142857142857143, - "train_accuracy_recipient_token_get_total": 7, - "train_total_number_recipient_token": 17 + "epoch": 0.7549407114624506, + "grad_norm": 2.126256397171694, + "learning_rate": 7.493153614118634e-07, + "loss": 0.1726, + "step": 382 }, { - "epoch": 0.7025691699604744, - "step": 237, - "train_accuracy_first_token_arguments": 0.9285714285714286, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.7549407114624506, + "step": 382, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_compress": 1.0, + "train_accuracy_recipient_token_compress_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 9, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.7025691699604744, - "step": 237, - "train_accuracy_first_token_arguments": 0.7435897435897436, - "train_accuracy_first_token_arguments_total": 39, + "epoch": 0.7549407114624506, + "step": 382, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_check": 1.0, - "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_custom": 1.0, + "train_accuracy_recipient_token_custom_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_generate": 1.0, - "train_accuracy_recipient_token_generate_total": 5, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_style": 1.0, + "train_accuracy_recipient_token_style_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.7055335968379447, - "grad_norm": 1.6960089025056184, - "learning_rate": 1.054023576166014e-06, - "loss": 0.2385, - "step": 238 + "epoch": 0.7569169960474308, + "grad_norm": 1.8984072674752601, + "learning_rate": 7.379090631209712e-07, + "loss": 0.1575, + "step": 383 }, { - "epoch": 0.7055335968379447, - "step": 238, - "train_accuracy_first_token_arguments": 0.9, - "train_accuracy_first_token_arguments_total": 20, - "train_accuracy_recipient_token": 0.9411764705882353, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, + "epoch": 0.7569169960474308, + "step": 383, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_total_number_recipient_token": 17 - }, - { - "epoch": 0.7055335968379447, - "step": 238, - "train_accuracy_first_token_arguments": 0.8947368421052632, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.9333333333333333, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 1, - "train_accuracy_recipient_token_count": 1.0, - "train_accuracy_recipient_token_count_total": 1, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_track": 1.0, - "train_accuracy_recipient_token_track_total": 1, - "train_total_number_recipient_token": 15 + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.7055335968379447, - "step": 238, - "train_accuracy_first_token_arguments": 0.9090909090909091, - "train_accuracy_first_token_arguments_total": 44, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.7569169960474308, + "step": 383, + "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_find": 0.6666666666666666, + "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 16 + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_rec": 1.0, + "train_accuracy_recipient_token_rec_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.708498023715415, - "grad_norm": 2.366641831793585, - "learning_rate": 1.0344377585076e-06, - "loss": 0.4772, - "step": 239 + "epoch": 0.758893280632411, + "grad_norm": 1.8794511099778521, + "learning_rate": 7.265751974868554e-07, + "loss": 0.18, + "step": 384 }, { - "epoch": 0.708498023715415, - "step": 239, - "train_accuracy_first_token_arguments": 0.9444444444444444, - "train_accuracy_first_token_arguments_total": 18, + "epoch": 0.758893280632411, + "step": 384, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_order": 1.0, + "train_accuracy_recipient_token_order_total": 1, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.708498023715415, - "step": 239, - "train_accuracy_first_token_arguments": 0.75, - "train_accuracy_first_token_arguments_total": 28, - "train_accuracy_recipient_token": 0.9333333333333333, + "epoch": 0.758893280632411, + "step": 384, + "train_accuracy_recipient_token": 0.8125, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_book": 0.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_clean": 1.0, + "train_accuracy_recipient_token_clean_total": 1, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_rent": 1.0, + "train_accuracy_recipient_token_rent_total": 2, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.7608695652173914, + "grad_norm": 2.08074863120985, + "learning_rate": 7.153142304003419e-07, + "loss": 0.1936, + "step": 385 + }, + { + "epoch": 0.7608695652173914, + "step": 385, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_merge": 1.0, - "train_accuracy_recipient_token_merge_total": 1, - "train_total_number_recipient_token": 15 + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.708498023715415, - "step": 239, - "train_accuracy_first_token_arguments": 0.9411764705882353, - "train_accuracy_first_token_arguments_total": 17, + "epoch": 0.7608695652173914, + "step": 385, "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 4, + "train_accuracy_recipient_token_get": 0.8888888888888888, + "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { - "epoch": 0.7114624505928854, - "grad_norm": 2.0428633070085302, - "learning_rate": 1.0149880431451736e-06, - "loss": 0.2845, - "step": 240 + "epoch": 0.7628458498023716, + "grad_norm": 1.8910489893410165, + "learning_rate": 7.041266247556814e-07, + "loss": 0.1853, + "step": 386 }, { - "epoch": 0.7114624505928854, - "step": 240, - "train_accuracy_first_token_arguments": 0.9473684210526315, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.95, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.7628458498023716, + "step": 386, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_book": 0.0, - "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_order": 1.0, - "train_accuracy_recipient_token_order_total": 1, - "train_total_number_recipient_token": 20 + "train_total_number_recipient_token": 19 }, { - "epoch": 0.7114624505928854, - "step": 240, - "train_accuracy_first_token_arguments": 0.8421052631578947, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.8333333333333334, + "epoch": 0.7628458498023716, + "step": 386, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_calculate": 0.5, + "train_accuracy_recipient_token_calculate_total": 2, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.7648221343873518, + "grad_norm": 2.1682658052582444, + "learning_rate": 6.930128404315214e-07, + "loss": 0.1819, + "step": 387 + }, + { + "epoch": 0.7648221343873518, + "step": 387, + "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 0.6666666666666666, - "train_accuracy_recipient_token_find_total": 3, - "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_ap": 1.0, + "train_accuracy_recipient_token_ap_total": 1, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_inspect": 1.0, - "train_accuracy_recipient_token_inspect_total": 1, - "train_accuracy_recipient_token_schedule": 1.0, - "train_accuracy_recipient_token_schedule_total": 1, - "train_total_number_recipient_token": 18 + "train_total_number_recipient_token": 17 }, { - "epoch": 0.7114624505928854, - "step": 240, - "train_accuracy_first_token_arguments": 0.9, - "train_accuracy_first_token_arguments_total": 30, - "train_accuracy_recipient_token": 0.8888888888888888, + "epoch": 0.7648221343873518, + "step": 387, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_count": 1.0, - "train_accuracy_recipient_token_count_total": 1, - "train_accuracy_recipient_token_d": 1.0, - "train_accuracy_recipient_token_d_total": 1, - "train_accuracy_recipient_token_find": 0.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_vehicle": 0.6666666666666666, - "train_accuracy_recipient_token_vehicle_total": 3, - "train_total_number_recipient_token": 18 - }, - { - "epoch": 0.7144268774703557, - "grad_norm": 1.870434102867515, - "learning_rate": 9.956762363144892e-07, - "loss": 0.1578, - "step": 241 + "train_accuracy_recipient_token_all_total": 9, + "train_accuracy_recipient_token_create": 1.0, + "train_accuracy_recipient_token_create_total": 1, + "train_accuracy_recipient_token_dis": 1.0, + "train_accuracy_recipient_token_dis_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_interpret": 1.0, + "train_accuracy_recipient_token_interpret_total": 1, + "train_total_number_recipient_token": 16 }, { - "epoch": 0.7144268774703557, - "step": 241, - "train_accuracy_first_token_arguments": 0.7222222222222222, - "train_accuracy_first_token_arguments_total": 36, - "train_accuracy_recipient_token": 0.7894736842105263, - "train_accuracy_recipient_token_all": 0.8181818181818182, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 0.6666666666666666, - "train_accuracy_recipient_token_get_total": 6, - "train_accuracy_recipient_token_search": 1.0, - "train_accuracy_recipient_token_search_total": 1, - "train_total_number_recipient_token": 19 + "epoch": 0.766798418972332, + "grad_norm": 1.581806277615435, + "learning_rate": 6.819733342720067e-07, + "loss": 0.1822, + "step": 388 }, { - "epoch": 0.7144268774703557, - "step": 241, - "train_accuracy_first_token_arguments": 0.8235294117647058, - "train_accuracy_first_token_arguments_total": 17, + "epoch": 0.766798418972332, + "step": 388, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_event": 0.6666666666666666, + "train_accuracy_recipient_token_event_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_recommend": 0.0, - "train_accuracy_recipient_token_recommend_total": 1, - "train_accuracy_recipient_token_search": 1.0, - "train_accuracy_recipient_token_search_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_s": 1.0, + "train_accuracy_recipient_token_s_total": 2, "train_total_number_recipient_token": 18 }, { - "epoch": 0.7144268774703557, - "step": 241, - "train_accuracy_first_token_arguments": 0.7272727272727273, - "train_accuracy_first_token_arguments_total": 11, + "epoch": 0.766798418972332, + "step": 388, "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 0.5, - "train_accuracy_recipient_token_analyze_total": 2, - "train_accuracy_recipient_token_building": 1.0, - "train_accuracy_recipient_token_building_total": 3, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_object": 1.0, - "train_accuracy_recipient_token_object_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.717391304347826, - "grad_norm": 2.192061804607294, - "learning_rate": 9.765041314441529e-07, - "loss": 0.3503, - "step": 242 + "epoch": 0.7687747035573123, + "grad_norm": 1.2710521758130064, + "learning_rate": 6.710085600679967e-07, + "loss": 0.0792, + "step": 389 }, { - "epoch": 0.717391304347826, - "step": 242, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 0.9473684210526315, + "epoch": 0.7687747035573123, + "step": 389, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 0.8888888888888888, - "train_accuracy_recipient_token_get_total": 9, - "train_total_number_recipient_token": 19 - }, - { - "epoch": 0.717391304347826, - "step": 242, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 11, - "train_accuracy_recipient_token": 0.9411764705882353, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_join": 1.0, - "train_accuracy_recipient_token_join_total": 1, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_order": 1.0, + "train_accuracy_recipient_token_order_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.717391304347826, - "step": 242, - "train_accuracy_first_token_arguments": 0.9090909090909091, - "train_accuracy_first_token_arguments_total": 11, - "train_accuracy_recipient_token": 0.9411764705882353, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compile": 1.0, - "train_accuracy_recipient_token_compile_total": 1, + "epoch": 0.7687747035573123, + "step": 389, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 0.8, - "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.7203557312252964, - "grad_norm": 2.540677501201676, - "learning_rate": 9.574735089890765e-07, - "loss": 0.3029, - "step": 243 + "epoch": 0.7707509881422925, + "grad_norm": 1.618155327367614, + "learning_rate": 6.601189685384127e-07, + "loss": 0.1764, + "step": 390 }, { - "epoch": 0.7203557312252964, - "step": 243, - "train_accuracy_first_token_arguments": 0.8, - "train_accuracy_first_token_arguments_total": 10, - "train_accuracy_recipient_token": 0.9411764705882353, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, + "epoch": 0.7707509881422925, + "step": 390, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8888888888888888, + "train_accuracy_recipient_token_all_total": 9, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, + "train_accuracy_recipient_token_restore": 0.0, + "train_accuracy_recipient_token_restore_total": 1, "train_total_number_recipient_token": 17 }, { - "epoch": 0.7203557312252964, - "step": 243, - "train_accuracy_first_token_arguments": 0.8235294117647058, - "train_accuracy_first_token_arguments_total": 17, + "epoch": 0.7707509881422925, + "step": 390, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_evaluate": 1.0, + "train_accuracy_recipient_token_evaluate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_order": 1.0, - "train_accuracy_recipient_token_order_total": 1, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 1, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.7203557312252964, - "step": 243, - "train_accuracy_first_token_arguments": 0.6875, - "train_accuracy_first_token_arguments_total": 16, + "epoch": 0.7727272727272727, + "grad_norm": 1.5648952371083706, + "learning_rate": 6.493050073117115e-07, + "loss": 0.2366, + "step": 391 + }, + { + "epoch": 0.7727272727272727, + "step": 391, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.7727272727272727, + "step": 391, "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_exp": 1.0, - "train_accuracy_recipient_token_exp_total": 4, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_find": 0.8, + "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_schedule": 1.0, - "train_accuracy_recipient_token_schedule_total": 1, + "train_accuracy_recipient_token_locate": 1.0, + "train_accuracy_recipient_token_locate_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.7233201581027668, - "grad_norm": 1.681752257297182, - "learning_rate": 9.385861362651322e-07, - "loss": 0.2678, - "step": 244 + "epoch": 0.7747035573122529, + "grad_norm": 1.8885528316331135, + "learning_rate": 6.385671209074829e-07, + "loss": 0.2286, + "step": 392 }, { - "epoch": 0.7233201581027668, - "step": 244, - "train_accuracy_first_token_arguments": 0.717391304347826, - "train_accuracy_first_token_arguments_total": 46, + "epoch": 0.7747035573122529, + "step": 392, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_create": 1.0, - "train_accuracy_recipient_token_create_total": 1, - "train_accuracy_recipient_token_design": 1.0, - "train_accuracy_recipient_token_design_total": 2, - "train_accuracy_recipient_token_draw": 1.0, - "train_accuracy_recipient_token_draw_total": 1, - "train_accuracy_recipient_token_evaluate": 0.0, - "train_accuracy_recipient_token_evaluate_total": 1, + "train_accuracy_recipient_token_ass": 0.0, + "train_accuracy_recipient_token_ass_total": 1, + "train_accuracy_recipient_token_environment": 1.0, + "train_accuracy_recipient_token_environment_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_report": 1.0, + "train_accuracy_recipient_token_report_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.7233201581027668, - "step": 244, - "train_accuracy_first_token_arguments": 0.9130434782608695, - "train_accuracy_first_token_arguments_total": 23, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.7747035573122529, + "step": 392, + "train_accuracy_recipient_token": 0.9523809523809523, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_check": 1.0, - "train_accuracy_recipient_token_check_total": 3, + "train_accuracy_recipient_token_f": 0.75, + "train_accuracy_recipient_token_f_total": 4, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_list": 1.0, - "train_accuracy_recipient_token_list_total": 3, - "train_accuracy_recipient_token_use": 1.0, - "train_accuracy_recipient_token_use_total": 1, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_locate": 1.0, + "train_accuracy_recipient_token_locate_total": 1, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 2, + "train_total_number_recipient_token": 21 }, { - "epoch": 0.7233201581027668, - "step": 244, - "train_accuracy_first_token_arguments": 0.9423076923076923, - "train_accuracy_first_token_arguments_total": 52, - "train_accuracy_recipient_token": 0.8823529411764706, + "epoch": 0.7766798418972332, + "grad_norm": 1.841125886164984, + "learning_rate": 6.279057507181796e-07, + "loss": 0.2481, + "step": 393 + }, + { + "epoch": 0.7766798418972332, + "step": 393, + "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_design": 1.0, - "train_accuracy_recipient_token_design_total": 1, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 1, - "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 17 }, { - "epoch": 0.7262845849802372, - "grad_norm": 1.6081692582595575, - "learning_rate": 9.198437672850249e-07, - "loss": 0.2291, - "step": 245 - }, - { - "epoch": 0.7262845849802372, - "step": 245, - "train_accuracy_first_token_arguments": 0.8, - "train_accuracy_first_token_arguments_total": 10, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.7766798418972332, + "step": 393, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_create": 1.0, - "train_accuracy_recipient_token_create_total": 2, - "train_accuracy_recipient_token_generate": 1.0, - "train_accuracy_recipient_token_generate_total": 2, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, - "train_total_number_recipient_token": 18 + "train_total_number_recipient_token": 16 }, { - "epoch": 0.7262845849802372, - "step": 245, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 17, + "epoch": 0.7786561264822134, + "grad_norm": 1.3411412038428177, + "learning_rate": 6.17321334990973e-07, + "loss": 0.0975, + "step": 394 + }, + { + "epoch": 0.7786561264822134, + "step": 394, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 2, + "train_accuracy_recipient_token_enh": 1.0, + "train_accuracy_recipient_token_enh_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 7, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_get_total": 1, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.7262845849802372, - "step": 245, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 20, - "train_accuracy_recipient_token": 0.9411764705882353, - "train_accuracy_recipient_token_all": 0.9166666666666666, - "train_accuracy_recipient_token_all_total": 12, + "epoch": 0.7786561264822134, + "step": 394, + "train_accuracy_recipient_token": 0.85, + "train_accuracy_recipient_token_all": 0.7692307692307693, + "train_accuracy_recipient_token_all_total": 13, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.7292490118577075, - "grad_norm": 1.786148716474924, - "learning_rate": 9.012481425954053e-07, - "loss": 0.2712, - "step": 246 + "epoch": 0.7806324110671937, + "grad_norm": 2.0638574101441773, + "learning_rate": 6.068143088097372e-07, + "loss": 0.2713, + "step": 395 }, { - "epoch": 0.7292490118577075, - "step": 246, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 9, - "train_accuracy_recipient_token": 0.9411764705882353, + "epoch": 0.7806324110671937, + "step": 395, + "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 6, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_locate": 1.0, + "train_accuracy_recipient_token_locate_total": 2, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 2, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.7292490118577075, - "step": 246, - "train_accuracy_first_token_arguments": 0.9333333333333333, - "train_accuracy_first_token_arguments_total": 30, + "epoch": 0.7806324110671937, + "step": 395, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_calculate": 1.0, - "train_accuracy_recipient_token_calculate_total": 2, - "train_accuracy_recipient_token_inventory": 1.0, - "train_accuracy_recipient_token_inventory_total": 1, - "train_accuracy_recipient_token_sales": 1.0, - "train_accuracy_recipient_token_sales_total": 1, - "train_accuracy_recipient_token_shelf": 1.0, - "train_accuracy_recipient_token_shelf_total": 1, - "train_total_number_recipient_token": 18 - }, - { - "epoch": 0.7292490118577075, - "step": 246, - "train_accuracy_first_token_arguments": 0.875, - "train_accuracy_first_token_arguments_total": 24, - "train_accuracy_recipient_token": 0.9, - "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_buy": 0.0, - "train_accuracy_recipient_token_buy_total": 1, - "train_accuracy_recipient_token_find": 0.8571428571428571, - "train_accuracy_recipient_token_find_total": 7, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 20 }, { - "epoch": 0.7322134387351779, - "grad_norm": 1.7149609341130112, - "learning_rate": 8.828009891152301e-07, - "loss": 0.2038, - "step": 247 + "epoch": 0.782608695652174, + "grad_norm": 1.5396226003765137, + "learning_rate": 5.963851040771639e-07, + "loss": 0.2094, + "step": 396 }, { - "epoch": 0.7322134387351779, - "step": 247, - "train_accuracy_first_token_arguments": 0.7857142857142857, - "train_accuracy_first_token_arguments_total": 28, + "epoch": 0.782608695652174, + "step": 396, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_locate": 1.0, - "train_accuracy_recipient_token_locate_total": 1, - "train_accuracy_recipient_token_validate": 1.0, - "train_accuracy_recipient_token_validate_total": 1, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_stamp": 1.0, + "train_accuracy_recipient_token_stamp_total": 4, "train_total_number_recipient_token": 19 }, { - "epoch": 0.7322134387351779, - "step": 247, - "train_accuracy_first_token_arguments": 0.8571428571428571, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 0.95, + "epoch": 0.782608695652174, + "step": 396, + "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_m": 1.0, - "train_accuracy_recipient_token_m_total": 4, - "train_total_number_recipient_token": 20 - }, - { - "epoch": 0.7322134387351779, - "step": 247, - "train_accuracy_first_token_arguments": 0.9761904761904762, - "train_accuracy_first_token_arguments_total": 42, - "train_accuracy_recipient_token": 0.7777777777777778, - "train_accuracy_recipient_token_all": 0.7857142857142857, - "train_accuracy_recipient_token_all_total": 14, - "train_accuracy_recipient_token_calculate": 0.0, - "train_accuracy_recipient_token_calculate_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_total_number_recipient_token": 18 - }, - { - "epoch": 0.7351778656126482, - "grad_norm": 1.6828278608062666, - "learning_rate": 8.64504019975386e-07, - "loss": 0.218, - "step": 248 - }, - { - "epoch": 0.7351778656126482, - "step": 248, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 4, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.7351778656126482, - "step": 248, - "train_accuracy_first_token_arguments": 0.9166666666666666, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.8421052631578947, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 3, - "train_accuracy_recipient_token_get": 0.6, - "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 19 - }, - { - "epoch": 0.7351778656126482, - "step": 248, - "train_accuracy_first_token_arguments": 0.9523809523809523, - "train_accuracy_first_token_arguments_total": 21, - "train_accuracy_recipient_token": 0.8947368421052632, - "train_accuracy_recipient_token_all": 0.9285714285714286, - "train_accuracy_recipient_token_all_total": 14, - "train_accuracy_recipient_token_calculate": 1.0, - "train_accuracy_recipient_token_calculate_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_translate": 0.0, - "train_accuracy_recipient_token_translate_total": 1, - "train_accuracy_recipient_token_vehicle": 1.0, - "train_accuracy_recipient_token_vehicle_total": 1, - "train_total_number_recipient_token": 19 - }, - { - "epoch": 0.7381422924901185, - "grad_norm": 2.180566937973925, - "learning_rate": 8.463589343595976e-07, - "loss": 0.3458, - "step": 249 + "epoch": 0.7845849802371542, + "grad_norm": 2.0476241421143033, + "learning_rate": 5.860341494970132e-07, + "loss": 0.3632, + "step": 397 }, { - "epoch": 0.7381422924901185, - "step": 249, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 27, + "epoch": 0.7845849802371542, + "step": 397, "train_accuracy_recipient_token": 0.8947368421052632, - "train_accuracy_recipient_token_all": 0.9166666666666666, - "train_accuracy_recipient_token_all_total": 12, - "train_accuracy_recipient_token_check": 1.0, - "train_accuracy_recipient_token_check_total": 1, - "train_accuracy_recipient_token_construction": 1.0, - "train_accuracy_recipient_token_construction_total": 2, - "train_accuracy_recipient_token_get": 0.75, - "train_accuracy_recipient_token_get_total": 4, - "train_total_number_recipient_token": 19 - }, - { - "epoch": 0.7381422924901185, - "step": 249, - "train_accuracy_first_token_arguments": 0.9375, - "train_accuracy_first_token_arguments_total": 16, - "train_accuracy_recipient_token": 0.8823529411764706, - "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 0.7142857142857143, - "train_accuracy_recipient_token_get_total": 7, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_city": 1.0, + "train_accuracy_recipient_token_city_total": 1, + "train_accuracy_recipient_token_generate": 1.0, + "train_accuracy_recipient_token_generate_total": 1, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_urban": 1.0, + "train_accuracy_recipient_token_urban_total": 1, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.7381422924901185, - "step": 249, - "train_accuracy_first_token_arguments": 0.9285714285714286, - "train_accuracy_first_token_arguments_total": 28, + "epoch": 0.7845849802371542, + "step": 397, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_gift": 1.0, + "train_accuracy_recipient_token_gift_total": 1, + "train_accuracy_recipient_token_personal": 1.0, + "train_accuracy_recipient_token_personal_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.741106719367589, - "grad_norm": 1.7757645324819282, - "learning_rate": 8.28367417346633e-07, - "loss": 0.2223, - "step": 250 - }, - { - "epoch": 0.741106719367589, - "step": 250, - "train_accuracy_first_token_arguments": 0.75, - "train_accuracy_first_token_arguments_total": 8, - "train_accuracy_recipient_token": 0.8, - "train_accuracy_recipient_token_all": 0.8, - "train_accuracy_recipient_token_all_total": 5, - "train_accuracy_recipient_token_get": 0.8, - "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 10 + "epoch": 0.7865612648221344, + "grad_norm": 1.9696813260070691, + "learning_rate": 5.757618705564849e-07, + "loss": 0.2611, + "step": 398 }, { - "epoch": 0.741106719367589, - "step": 250, - "train_accuracy_first_token_arguments": 0.8571428571428571, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 0.9444444444444444, + "epoch": 0.7865612648221344, + "step": 398, + "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_animal": 1.0, - "train_accuracy_recipient_token_animal_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_locate": 0.0, - "train_accuracy_recipient_token_locate_total": 1, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_get_total": 8, + "train_accuracy_recipient_token_initialize": 0.0, + "train_accuracy_recipient_token_initialize_total": 1, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.741106719367589, - "step": 250, - "train_accuracy_first_token_arguments": 0.95, - "train_accuracy_first_token_arguments_total": 20, - "train_accuracy_recipient_token": 0.7857142857142857, - "train_accuracy_recipient_token_all": 0.8181818181818182, - "train_accuracy_recipient_token_all_total": 11, + "epoch": 0.7865612648221344, + "step": 398, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analy": 0.0, + "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 0.5, - "train_accuracy_recipient_token_get_total": 2, - "train_total_number_recipient_token": 14 - }, - { - "epoch": 0.7440711462450593, - "grad_norm": 11.280998921270033, - "learning_rate": 8.105311397538085e-07, - "loss": 0.5545, - "step": 251 + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_verify": 1.0, + "train_accuracy_recipient_token_verify_total": 1, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.7440711462450593, - "step": 251, - "train_accuracy_first_token_arguments": 0.7142857142857143, - "train_accuracy_first_token_arguments_total": 28, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 1, - "train_accuracy_recipient_token_compose": 1.0, - "train_accuracy_recipient_token_compose_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 1, - "train_total_number_recipient_token": 20 + "epoch": 0.7885375494071146, + "grad_norm": 2.175586246163446, + "learning_rate": 5.655686895087328e-07, + "loss": 0.1876, + "step": 399 }, { - "epoch": 0.7440711462450593, - "step": 251, - "train_accuracy_first_token_arguments": 0.75, - "train_accuracy_first_token_arguments_total": 20, + "epoch": 0.7885375494071146, + "step": 399, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_building": 1.0, + "train_accuracy_recipient_token_building_total": 1, + "train_accuracy_recipient_token_flag": 1.0, + "train_accuracy_recipient_token_flag_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 8, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 2, "train_total_number_recipient_token": 18 }, { - "epoch": 0.7440711462450593, - "step": 251, - "train_accuracy_first_token_arguments": 0.9090909090909091, - "train_accuracy_first_token_arguments_total": 11, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, + "epoch": 0.7885375494071146, + "step": 399, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_deb": 0.0, + "train_accuracy_recipient_token_deb_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 9, - "train_total_number_recipient_token": 19 - }, - { - "epoch": 0.7470355731225297, - "grad_norm": 2.177175537608394, - "learning_rate": 7.928517579818312e-07, - "loss": 0.2545, - "step": 252 + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.7470355731225297, - "step": 252, - "train_accuracy_first_token_arguments": 0.8, - "train_accuracy_first_token_arguments_total": 30, - "train_accuracy_recipient_token": 0.8235294117647058, - "train_accuracy_recipient_token_all": 0.75, - "train_accuracy_recipient_token_all_total": 12, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 4, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, - "train_total_number_recipient_token": 17 + "epoch": 0.7905138339920948, + "grad_norm": 2.494414189208174, + "learning_rate": 5.554550253555066e-07, + "loss": 0.2566, + "step": 400 }, { - "epoch": 0.7470355731225297, - "step": 252, - "train_accuracy_first_token_arguments": 0.6666666666666666, - "train_accuracy_first_token_arguments_total": 15, - "train_accuracy_recipient_token": 0.95, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.7905138339920948, + "step": 400, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_evaluate": 0.6666666666666666, - "train_accuracy_recipient_token_evaluate_total": 3, + "train_accuracy_recipient_token_building": 1.0, + "train_accuracy_recipient_token_building_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 6, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.7470355731225297, - "step": 252, - "train_accuracy_first_token_arguments": 0.9473684210526315, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.95, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.7905138339920948, + "step": 400, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_book": 1.0, - "train_accuracy_recipient_token_book_total": 1, - "train_accuracy_recipient_token_find": 0.5, - "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_configure": 1.0, + "train_accuracy_recipient_token_configure_total": 5, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 6, - "train_accuracy_recipient_token_te": 1.0, - "train_accuracy_recipient_token_te_total": 1, - "train_total_number_recipient_token": 20 - }, - { - "epoch": 0.75, - "grad_norm": 1.527663654975652, - "learning_rate": 7.753309138609705e-07, - "loss": 0.1884, - "step": 253 + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.75, - "step": 253, - "train_accuracy_first_token_arguments": 0.7884615384615384, - "train_accuracy_first_token_arguments_total": 52, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_c": 1.0, - "train_accuracy_recipient_token_c_total": 1, - "train_accuracy_recipient_token_evaluate": 0.0, - "train_accuracy_recipient_token_evaluate_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_suggest": 1.0, - "train_accuracy_recipient_token_suggest_total": 2, - "train_total_number_recipient_token": 18 + "epoch": 0.7924901185770751, + "grad_norm": 1.953703115430371, + "learning_rate": 5.454212938299256e-07, + "loss": 0.1746, + "step": 401 }, { - "epoch": 0.75, - "step": 253, - "train_accuracy_first_token_arguments": 0.6764705882352942, - "train_accuracy_first_token_arguments_total": 34, - "train_accuracy_recipient_token": 0.7777777777777778, + "epoch": 0.7924901185770751, + "step": 401, + "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_check": 0.0, - "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_review": 1.0, - "train_accuracy_recipient_token_review_total": 1, - "train_accuracy_recipient_token_stamp": 0.75, - "train_accuracy_recipient_token_stamp_total": 4, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.75, - "step": 253, - "train_accuracy_first_token_arguments": 0.8125, - "train_accuracy_first_token_arguments_total": 16, + "epoch": 0.7924901185770751, + "step": 401, "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_convert": 1.0, - "train_accuracy_recipient_token_convert_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_sum": 1.0, - "train_accuracy_recipient_token_sum_total": 1, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, + "train_accuracy_recipient_token_rel": 1.0, + "train_accuracy_recipient_token_rel_total": 2, + "train_accuracy_recipient_token_y": 1.0, + "train_accuracy_recipient_token_y_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.7529644268774703, - "grad_norm": 1.9700572495421322, - "learning_rate": 7.579702344985899e-07, - "loss": 0.3442, - "step": 254 + "epoch": 0.7944664031620553, + "grad_norm": 1.8516896605082545, + "learning_rate": 5.354679073793942e-07, + "loss": 0.1904, + "step": 402 }, { - "epoch": 0.7529644268774703, - "step": 254, - "train_accuracy_first_token_arguments": 0.84, - "train_accuracy_first_token_arguments_total": 25, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.8181818181818182, + "epoch": 0.7944664031620553, + "step": 402, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_check": 1.0, - "train_accuracy_recipient_token_check_total": 2, - "train_accuracy_recipient_token_decode": 1.0, - "train_accuracy_recipient_token_decode_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.7529644268774703, - "step": 254, - "train_accuracy_first_token_arguments": 0.8125, - "train_accuracy_first_token_arguments_total": 16, + "epoch": 0.7944664031620553, + "step": 402, "train_accuracy_recipient_token": 0.8947368421052632, - "train_accuracy_recipient_token_adjust": 1.0, - "train_accuracy_recipient_token_adjust_total": 1, - "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_discover": 1.0, - "train_accuracy_recipient_token_discover_total": 1, - "train_accuracy_recipient_token_enh": 1.0, - "train_accuracy_recipient_token_enh_total": 1, - "train_accuracy_recipient_token_event": 1.0, - "train_accuracy_recipient_token_event_total": 4, + "train_accuracy_recipient_token_get": 0.875, + "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { - "epoch": 0.7529644268774703, - "step": 254, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 15, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.7964426877470355, + "grad_norm": 1.8578121936874719, + "learning_rate": 5.255952751486442e-07, + "loss": 0.2239, + "step": 403 + }, + { + "epoch": 0.7964426877470355, + "step": 403, + "train_accuracy_recipient_token": 0.9523809523809523, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_book": 1.0, - "train_accuracy_recipient_token_book_total": 1, - "train_accuracy_recipient_token_compress": 1.0, - "train_accuracy_recipient_token_compress_total": 1, + "train_accuracy_recipient_token_count": 1.0, + "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_locate": 1.0, + "train_accuracy_recipient_token_locate_total": 1, + "train_total_number_recipient_token": 21 + }, + { + "epoch": 0.7964426877470355, + "step": 403, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_short": 1.0, + "train_accuracy_recipient_token_short_total": 1, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 1, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.7559288537549407, - "grad_norm": 2.171316152952409, - "learning_rate": 7.407713321280377e-07, - "loss": 0.1791, - "step": 255 + "epoch": 0.7984189723320159, + "grad_norm": 2.1579654219661144, + "learning_rate": 5.158038029629195e-07, + "loss": 0.2192, + "step": 404 }, { - "epoch": 0.7559288537549407, - "step": 255, - "train_accuracy_first_token_arguments": 0.8947368421052632, - "train_accuracy_first_token_arguments_total": 38, - "train_accuracy_recipient_token": 0.9473684210526315, + "epoch": 0.7984189723320159, + "step": 404, + "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_customize": 1.0, - "train_accuracy_recipient_token_customize_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 3, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_style": 1.0, - "train_accuracy_recipient_token_style_total": 1, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_recipe": 1.0, + "train_accuracy_recipient_token_recipe_total": 4, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.7559288537549407, - "step": 255, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 26, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.7984189723320159, + "step": 404, + "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_calculate": 1.0, - "train_accuracy_recipient_token_calculate_total": 1, - "train_accuracy_recipient_token_check": 1.0, - "train_accuracy_recipient_token_check_total": 1, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 18 + "train_total_number_recipient_token": 19 }, { - "epoch": 0.7559288537549407, - "step": 255, - "train_accuracy_first_token_arguments": 0.9285714285714286, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_calculate": 1.0, - "train_accuracy_recipient_token_calculate_total": 1, - "train_accuracy_recipient_token_find": 0.6666666666666666, - "train_accuracy_recipient_token_find_total": 3, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_rec": 1.0, - "train_accuracy_recipient_token_rec_total": 1, - "train_total_number_recipient_token": 18 + "epoch": 0.8003952569169961, + "grad_norm": 1.5522372172075602, + "learning_rate": 5.060938933112891e-07, + "loss": 0.14, + "step": 405 }, { - "epoch": 0.758893280632411, - "grad_norm": 1.7774995166747949, - "learning_rate": 7.237358039589271e-07, - "loss": 0.2064, - "step": 256 + "epoch": 0.8003952569169961, + "step": 405, + "train_accuracy_recipient_token": 0.75, + "train_accuracy_recipient_token_all": 0.75, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_crop": 1.0, + "train_accuracy_recipient_token_crop_total": 1, + "train_accuracy_recipient_token_f": 1.0, + "train_accuracy_recipient_token_f_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_har": 0.0, + "train_accuracy_recipient_token_har_total": 1, + "train_total_number_recipient_token": 16 }, { - "epoch": 0.758893280632411, - "step": 256, - "train_accuracy_first_token_arguments": 0.8529411764705882, - "train_accuracy_first_token_arguments_total": 34, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 1, + "epoch": 0.8003952569169961, + "step": 405, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.8888888888888888, + "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 7, - "train_accuracy_recipient_token_order": 1.0, - "train_accuracy_recipient_token_order_total": 1, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_land": 0.5, + "train_accuracy_recipient_token_land_total": 4, + "train_accuracy_recipient_token_mount": 1.0, + "train_accuracy_recipient_token_mount_total": 2, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.758893280632411, - "step": 256, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 17, - "train_accuracy_recipient_token": 0.875, + "epoch": 0.8023715415019763, + "grad_norm": 2.1036940569062943, + "learning_rate": 4.964659453301088e-07, + "loss": 0.2345, + "step": 406 + }, + { + "epoch": 0.8023715415019763, + "step": 406, + "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_book": 1.0, - "train_accuracy_recipient_token_book_total": 1, - "train_accuracy_recipient_token_clean": 1.0, - "train_accuracy_recipient_token_clean_total": 1, - "train_accuracy_recipient_token_find": 0.5, - "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_rent": 1.0, - "train_accuracy_recipient_token_rent_total": 2, - "train_total_number_recipient_token": 16 + "train_accuracy_recipient_token_compare": 0.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_go": 1.0, + "train_accuracy_recipient_token_go_total": 3, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.758893280632411, - "step": 256, - "train_accuracy_first_token_arguments": 0.9090909090909091, - "train_accuracy_first_token_arguments_total": 11, + "epoch": 0.8023715415019763, + "step": 406, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.7618577075098815, - "grad_norm": 1.7945920171198126, - "learning_rate": 7.068652320288081e-07, - "loss": 0.1652, - "step": 257 + "epoch": 0.8043478260869565, + "grad_norm": 1.6667772877286244, + "learning_rate": 4.869203547866097e-07, + "loss": 0.1708, + "step": 407 }, { - "epoch": 0.7618577075098815, - "step": 257, - "train_accuracy_first_token_arguments": 0.9166666666666666, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.9473684210526315, + "epoch": 0.8043478260869565, + "step": 407, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 0.8888888888888888, - "train_accuracy_recipient_token_get_total": 9, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.7618577075098815, - "step": 257, - "train_accuracy_first_token_arguments": 0.8105263157894737, - "train_accuracy_first_token_arguments_total": 95, - "train_accuracy_recipient_token": 0.9473684210526315, + "epoch": 0.8043478260869565, + "step": 407, + "train_accuracy_recipient_token": 0.7894736842105263, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 4, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_c": 0.5, + "train_accuracy_recipient_token_c_total": 2, + "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_ident": 0.5, + "train_accuracy_recipient_token_ident_total": 2, "train_total_number_recipient_token": 19 }, { - "epoch": 0.7618577075098815, - "step": 257, - "train_accuracy_first_token_arguments": 0.8666666666666667, - "train_accuracy_first_token_arguments_total": 15, - "train_accuracy_recipient_token": 0.8947368421052632, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_calculate": 0.5, - "train_accuracy_recipient_token_calculate_total": 2, - "train_accuracy_recipient_token_find": 0.5, - "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_total_number_recipient_token": 19 + "epoch": 0.8063241106719368, + "grad_norm": 2.2458743380912662, + "learning_rate": 4.774575140626317e-07, + "loss": 0.2932, + "step": 408 }, { - "epoch": 0.7648221343873518, - "grad_norm": 2.033550650864743, - "learning_rate": 6.901611830562469e-07, - "loss": 0.2407, - "step": 258 + "epoch": 0.8063241106719368, + "step": 408, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.7272727272727273, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.7648221343873518, - "step": 258, - "train_accuracy_first_token_arguments": 0.9047619047619048, - "train_accuracy_first_token_arguments_total": 21, - "train_accuracy_recipient_token": 0.9411764705882353, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.8063241106719368, + "step": 408, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_ap": 1.0, - "train_accuracy_recipient_token_ap_total": 1, - "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 1, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 17 }, { - "epoch": 0.7648221343873518, - "step": 258, - "train_accuracy_first_token_arguments": 0.72, - "train_accuracy_first_token_arguments_total": 25, - "train_accuracy_recipient_token": 0.8823529411764706, - "train_accuracy_recipient_token_all": 0.8, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_create": 1.0, - "train_accuracy_recipient_token_create_total": 1, - "train_accuracy_recipient_token_discover": 1.0, - "train_accuracy_recipient_token_discover_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_interpret": 1.0, - "train_accuracy_recipient_token_interpret_total": 1, - "train_total_number_recipient_token": 17 + "epoch": 0.808300395256917, + "grad_norm": 2.48794385983657, + "learning_rate": 4.6807781213849354e-07, + "loss": 0.2428, + "step": 409 }, { - "epoch": 0.7648221343873518, - "step": 258, - "train_accuracy_first_token_arguments": 0.9090909090909091, - "train_accuracy_first_token_arguments_total": 22, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.808300395256917, + "step": 409, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_event": 0.6666666666666666, - "train_accuracy_recipient_token_event_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_suggest": 1.0, - "train_accuracy_recipient_token_suggest_total": 2, - "train_total_number_recipient_token": 18 - }, - { - "epoch": 0.7677865612648221, - "grad_norm": 1.6980201197958136, - "learning_rate": 6.736252082953307e-07, - "loss": 0.2026, - "step": 259 - }, - { - "epoch": 0.7677865612648221, - "step": 259, - "train_accuracy_first_token_arguments": 0.9459459459459459, - "train_accuracy_first_token_arguments_total": 37, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 8, - "train_accuracy_recipient_token_search": 1.0, - "train_accuracy_recipient_token_search_total": 1, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 2, "train_total_number_recipient_token": 19 }, { - "epoch": 0.7677865612648221, - "step": 259, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 22, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.8, + "epoch": 0.808300395256917, + "step": 409, + "train_accuracy_recipient_token": 0.7647058823529411, + "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 3, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_order": 1.0, - "train_accuracy_recipient_token_order_total": 1, - "train_total_number_recipient_token": 18 - }, - { - "epoch": 0.7677865612648221, - "step": 259, - "train_accuracy_first_token_arguments": 0.92, - "train_accuracy_first_token_arguments_total": 50, - "train_accuracy_recipient_token": 0.95, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 7, - "train_total_number_recipient_token": 20 - }, - { - "epoch": 0.7707509881422925, - "grad_norm": 1.4870369650234745, - "learning_rate": 6.572588433916082e-07, - "loss": 0.1544, - "step": 260 + "train_accuracy_recipient_token_convert": 1.0, + "train_accuracy_recipient_token_convert_total": 3, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.7707509881422925, - "step": 260, - "train_accuracy_first_token_arguments": 0.8064516129032258, - "train_accuracy_first_token_arguments_total": 31, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 1, - "train_accuracy_recipient_token_restore": 0.0, - "train_accuracy_recipient_token_restore_total": 1, - "train_total_number_recipient_token": 18 + "epoch": 0.8102766798418972, + "grad_norm": 2.7307307205035567, + "learning_rate": 4.5878163457700327e-07, + "loss": 0.27, + "step": 410 }, { - "epoch": 0.7707509881422925, - "step": 260, - "train_accuracy_first_token_arguments": 0.7413793103448276, - "train_accuracy_first_token_arguments_total": 58, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_evaluate": 1.0, - "train_accuracy_recipient_token_evaluate_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 4, + "epoch": 0.8102766798418972, + "step": 410, + "train_accuracy_recipient_token": 0.9285714285714286, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_data": 1.0, + "train_accuracy_recipient_token_data_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_validate": 1.0, - "train_accuracy_recipient_token_validate_total": 1, - "train_total_number_recipient_token": 20 + "train_total_number_recipient_token": 14 }, { - "epoch": 0.7707509881422925, - "step": 260, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 29, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, + "epoch": 0.8102766798418972, + "step": 410, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 9, + "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 19 }, { - "epoch": 0.7737154150197628, - "grad_norm": 1.6094384600411795, - "learning_rate": 6.410636082394772e-07, - "loss": 0.2592, - "step": 261 + "epoch": 0.8122529644268774, + "grad_norm": 2.7985267824858187, + "learning_rate": 4.495693635076101e-07, + "loss": 0.2315, + "step": 411 }, { - "epoch": 0.7737154150197628, - "step": 261, - "train_accuracy_first_token_arguments": 0.7692307692307693, - "train_accuracy_first_token_arguments_total": 13, + "epoch": 0.8122529644268774, + "step": 411, "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 0.8, - "train_accuracy_recipient_token_find_total": 5, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_locate": 1.0, - "train_accuracy_recipient_token_locate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_h": 1.0, + "train_accuracy_recipient_token_h_total": 4, "train_total_number_recipient_token": 18 }, { - "epoch": 0.7737154150197628, - "step": 261, - "train_accuracy_first_token_arguments": 0.7666666666666667, - "train_accuracy_first_token_arguments_total": 30, - "train_accuracy_recipient_token": 0.9473684210526315, + "epoch": 0.8122529644268774, + "step": 411, + "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_ass": 1.0, - "train_accuracy_recipient_token_ass_total": 1, - "train_accuracy_recipient_token_environment": 1.0, - "train_accuracy_recipient_token_environment_total": 1, + "train_accuracy_recipient_token_compare": 0.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_w": 1.0, - "train_accuracy_recipient_token_w_total": 1, + "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { - "epoch": 0.7737154150197628, - "step": 261, - "train_accuracy_first_token_arguments": 0.975, - "train_accuracy_first_token_arguments_total": 40, - "train_accuracy_recipient_token": 0.9523809523809523, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_f": 1.0, - "train_accuracy_recipient_token_f_total": 4, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 3, - "train_accuracy_recipient_token_locate": 1.0, - "train_accuracy_recipient_token_locate_total": 1, - "train_accuracy_recipient_token_recommend": 0.5, - "train_accuracy_recipient_token_recommend_total": 2, - "train_total_number_recipient_token": 21 - }, - { - "epoch": 0.7766798418972332, - "grad_norm": 1.997171944251961, - "learning_rate": 6.250410068410367e-07, - "loss": 0.2894, - "step": 262 + "epoch": 0.8142292490118577, + "grad_norm": 1.2979415141987987, + "learning_rate": 4.4044137761069584e-07, + "loss": 0.0836, + "step": 412 }, { - "epoch": 0.7766798418972332, - "step": 262, - "train_accuracy_first_token_arguments": 0.9333333333333333, - "train_accuracy_first_token_arguments_total": 30, - "train_accuracy_recipient_token": 0.9411764705882353, + "epoch": 0.8142292490118577, + "step": 412, + "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 17 - }, - { - "epoch": 0.7766798418972332, - "step": 262, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 11, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_book": 1.0, - "train_accuracy_recipient_token_book_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_crop": 1.0, + "train_accuracy_recipient_token_crop_total": 1, + "train_accuracy_recipient_token_f": 1.0, + "train_accuracy_recipient_token_f_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, - "train_total_number_recipient_token": 16 + "train_accuracy_recipient_token_har": 0.0, + "train_accuracy_recipient_token_har_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.7766798418972332, - "step": 262, - "train_accuracy_first_token_arguments": 0.6956521739130435, - "train_accuracy_first_token_arguments_total": 69, + "epoch": 0.8142292490118577, + "step": 412, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 2, - "train_accuracy_recipient_token_enh": 1.0, - "train_accuracy_recipient_token_enh_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, - "train_total_number_recipient_token": 17 - }, - { - "epoch": 0.7796442687747036, - "grad_norm": 1.5690225372388775, - "learning_rate": 6.091925271664156e-07, - "loss": 0.1593, - "step": 263 + "train_accuracy_recipient_token_get_total": 10, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.7796442687747036, - "step": 263, - "train_accuracy_first_token_arguments": 0.75, - "train_accuracy_first_token_arguments_total": 24, - "train_accuracy_recipient_token": 0.85, - "train_accuracy_recipient_token_all": 0.7692307692307693, - "train_accuracy_recipient_token_all_total": 13, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_total_number_recipient_token": 20 + "epoch": 0.8162055335968379, + "grad_norm": 2.3682354960537606, + "learning_rate": 4.3139805210200917e-07, + "loss": 0.1285, + "step": 413 }, { - "epoch": 0.7796442687747036, - "step": 263, - "train_accuracy_first_token_arguments": 0.8333333333333334, - "train_accuracy_first_token_arguments_total": 18, + "epoch": 0.8162055335968379, + "step": 413, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.8162055335968379, + "step": 413, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_locate": 1.0, - "train_accuracy_recipient_token_locate_total": 2, + "train_accuracy_recipient_token_sum": 1.0, + "train_accuracy_recipient_token_sum_total": 1, "train_accuracy_recipient_token_translate": 1.0, - "train_accuracy_recipient_token_translate_total": 2, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_translate_total": 1, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.7796442687747036, - "step": 263, - "train_accuracy_first_token_arguments": 0.9285714285714286, - "train_accuracy_first_token_arguments_total": 14, + "epoch": 0.8181818181818182, + "grad_norm": 2.1831995030818256, + "learning_rate": 4.2243975871724017e-07, + "loss": 0.2866, + "step": 414 + }, + { + "epoch": 0.8181818181818182, + "step": 414, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.8181818181818182, + "step": 414, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 10, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_r": 1.0, + "train_accuracy_recipient_token_r_total": 4, "train_total_number_recipient_token": 20 }, { - "epoch": 0.782608695652174, - "grad_norm": 1.769966354853594, - "learning_rate": 5.93519641015591e-07, - "loss": 0.2617, - "step": 264 + "epoch": 0.8201581027667985, + "grad_norm": 2.3802602702771556, + "learning_rate": 4.1356686569674344e-07, + "loss": 0.2559, + "step": 415 }, { - "epoch": 0.782608695652174, - "step": 264, - "train_accuracy_first_token_arguments": 0.9, - "train_accuracy_first_token_arguments_total": 20, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, + "epoch": 0.8201581027667985, + "step": 415, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_stamp": 1.0, - "train_accuracy_recipient_token_stamp_total": 4, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.782608695652174, - "step": 264, - "train_accuracy_first_token_arguments": 0.6875, - "train_accuracy_first_token_arguments_total": 16, + "epoch": 0.8201581027667985, + "step": 415, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 0.75, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.8221343873517787, + "grad_norm": 2.0956524139072057, + "learning_rate": 4.0477973777039854e-07, + "loss": 0.177, + "step": 416 + }, + { + "epoch": 0.8221343873517787, + "step": 416, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_motor": 1.0, + "train_accuracy_recipient_token_motor_total": 3, "train_total_number_recipient_token": 18 }, { - "epoch": 0.782608695652174, - "step": 264, - "train_accuracy_first_token_arguments": 0.8529411764705882, - "train_accuracy_first_token_arguments_total": 34, - "train_accuracy_recipient_token": 0.9473684210526315, + "epoch": 0.8221343873517787, + "step": 416, + "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_city": 1.0, - "train_accuracy_recipient_token_city_total": 1, - "train_accuracy_recipient_token_generate": 1.0, - "train_accuracy_recipient_token_generate_total": 1, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_urban": 1.0, - "train_accuracy_recipient_token_urban_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 4, "train_total_number_recipient_token": 19 }, { - "epoch": 0.7855731225296443, - "grad_norm": 2.0500780698786483, - "learning_rate": 5.780238038817035e-07, - "loss": 0.4104, - "step": 265 + "epoch": 0.8241106719367589, + "grad_norm": 1.505715219382788, + "learning_rate": 3.9607873614261717e-07, + "loss": 0.2022, + "step": 417 }, { - "epoch": 0.7855731225296443, - "step": 265, - "train_accuracy_first_token_arguments": 0.9444444444444444, - "train_accuracy_first_token_arguments_total": 18, - "train_accuracy_recipient_token": 0.9473684210526315, + "epoch": 0.8241106719367589, + "step": 417, + "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_gift": 1.0, - "train_accuracy_recipient_token_gift_total": 1, - "train_accuracy_recipient_token_personal": 1.0, - "train_accuracy_recipient_token_personal_total": 1, - "train_total_number_recipient_token": 19 - }, - { - "epoch": 0.7855731225296443, - "step": 265, - "train_accuracy_first_token_arguments": 0.8333333333333334, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.95, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_compare": 0.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_get": 0.875, "train_accuracy_recipient_token_get_total": 8, - "train_accuracy_recipient_token_initialize": 0.0, - "train_accuracy_recipient_token_initialize_total": 1, - "train_total_number_recipient_token": 20 + "train_total_number_recipient_token": 19 }, { - "epoch": 0.7855731225296443, - "step": 265, - "train_accuracy_first_token_arguments": 0.782608695652174, - "train_accuracy_first_token_arguments_total": 23, - "train_accuracy_recipient_token": 0.9411764705882353, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.8241106719367589, + "step": 417, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 0.0, - "train_accuracy_recipient_token_analyze_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_verify": 1.0, - "train_accuracy_recipient_token_verify_total": 1, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 17 }, { - "epoch": 0.7885375494071146, - "grad_norm": 2.3088267661053083, - "learning_rate": 5.627064548158903e-07, - "loss": 0.215, - "step": 266 + "epoch": 0.8260869565217391, + "grad_norm": 1.7465367169322799, + "learning_rate": 3.874642184774977e-07, + "loss": 0.0952, + "step": 418 }, { - "epoch": 0.7885375494071146, - "step": 266, - "train_accuracy_first_token_arguments": 0.7692307692307693, - "train_accuracy_first_token_arguments_total": 13, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_building": 1.0, - "train_accuracy_recipient_token_building_total": 1, - "train_accuracy_recipient_token_flag": 1.0, - "train_accuracy_recipient_token_flag_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 2, - "train_total_number_recipient_token": 18 + "epoch": 0.8260869565217391, + "step": 418, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 9, + "train_accuracy_recipient_token_be": 0.75, + "train_accuracy_recipient_token_be_total": 4, + "train_accuracy_recipient_token_find": 0.8, + "train_accuracy_recipient_token_find_total": 5, + "train_accuracy_recipient_token_s": 1.0, + "train_accuracy_recipient_token_s_total": 1, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.7885375494071146, - "step": 266, - "train_accuracy_first_token_arguments": 0.8333333333333334, - "train_accuracy_first_token_arguments_total": 18, - "train_accuracy_recipient_token": 0.95, + "epoch": 0.8260869565217391, + "step": 418, + "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_deb": 1.0, - "train_accuracy_recipient_token_deb_total": 1, + "train_accuracy_recipient_token_create": 1.0, + "train_accuracy_recipient_token_create_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_find_total": 5, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.8280632411067194, + "grad_norm": 1.4978654089504846, + "learning_rate": 3.789365388841193e-07, + "loss": 0.1888, + "step": 419 + }, + { + "epoch": 0.8280632411067194, + "step": 419, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_gr": 1.0, + "train_accuracy_recipient_token_gr_total": 2, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_interpret": 1.0, + "train_accuracy_recipient_token_interpret_total": 1, + "train_accuracy_recipient_token_remove": 1.0, + "train_accuracy_recipient_token_remove_total": 1, "train_total_number_recipient_token": 20 }, { - "epoch": 0.7885375494071146, - "step": 266, - "train_accuracy_first_token_arguments": 0.7692307692307693, - "train_accuracy_first_token_arguments_total": 13, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.8, + "epoch": 0.8280632411067194, + "step": 419, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_building": 1.0, - "train_accuracy_recipient_token_building_total": 4, + "train_accuracy_recipient_token_check": 0.5, + "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.791501976284585, - "grad_norm": 2.3252209612398764, - "learning_rate": 5.47569016293649e-07, - "loss": 0.2737, - "step": 267 + "epoch": 0.8300395256916996, + "grad_norm": 1.5323629508097336, + "learning_rate": 3.7049604790198976e-07, + "loss": 0.1212, + "step": 420 }, { - "epoch": 0.791501976284585, - "step": 267, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 17, + "epoch": 0.8300395256916996, + "step": 420, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_configure": 1.0, - "train_accuracy_recipient_token_configure_total": 5, - "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_f": 1.0, + "train_accuracy_recipient_token_f_total": 2, + "train_accuracy_recipient_token_generate": 1.0, + "train_accuracy_recipient_token_generate_total": 1, + "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_port": 1.0, + "train_accuracy_recipient_token_port_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.791501976284585, - "step": 267, - "train_accuracy_first_token_arguments": 0.6363636363636364, - "train_accuracy_first_token_arguments_total": 22, - "train_accuracy_recipient_token": 0.9411764705882353, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, + "epoch": 0.8300395256916996, + "step": 420, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 1, - "train_accuracy_recipient_token_schedule": 1.0, - "train_accuracy_recipient_token_schedule_total": 1, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 2, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, "train_total_number_recipient_token": 17 }, { - "epoch": 0.791501976284585, - "step": 267, - "train_accuracy_first_token_arguments": 0.8666666666666667, - "train_accuracy_first_token_arguments_total": 15, + "epoch": 0.8320158102766798, + "grad_norm": 1.9086165088662692, + "learning_rate": 3.621430924866348e-07, + "loss": 0.2971, + "step": 421 + }, + { + "epoch": 0.8320158102766798, + "step": 421, "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 9, + "train_accuracy_recipient_token_calculate": 0.6666666666666666, + "train_accuracy_recipient_token_calculate_total": 3, + "train_accuracy_recipient_token_convert": 1.0, + "train_accuracy_recipient_token_convert_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 1, - "train_accuracy_recipient_token_rel": 1.0, - "train_accuracy_recipient_token_rel_total": 2, - "train_accuracy_recipient_token_y": 1.0, - "train_accuracy_recipient_token_y_total": 1, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_set": 1.0, + "train_accuracy_recipient_token_set_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.7944664031620553, - "grad_norm": 1.6931063369189647, - "learning_rate": 5.326128940827313e-07, - "loss": 0.172, - "step": 268 - }, - { - "epoch": 0.7944664031620553, - "step": 268, - "train_accuracy_first_token_arguments": 0.9166666666666666, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.95, - "train_accuracy_recipient_token_all": 0.9090909090909091, + "epoch": 0.8320158102766798, + "step": 421, + "train_accuracy_recipient_token": 0.8125, + "train_accuracy_recipient_token_a": 1.0, + "train_accuracy_recipient_token_a_total": 1, + "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 9, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_fish": 0.6666666666666666, + "train_accuracy_recipient_token_fish_total": 3, + "train_total_number_recipient_token": 16 }, { - "epoch": 0.7944664031620553, - "step": 268, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 17, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 8, - "train_total_number_recipient_token": 19 + "epoch": 0.83399209486166, + "grad_norm": 2.9759596008058633, + "learning_rate": 3.538780159953348e-07, + "loss": 0.212, + "step": 422 }, { - "epoch": 0.7944664031620553, - "step": 268, - "train_accuracy_first_token_arguments": 0.9375, - "train_accuracy_first_token_arguments_total": 16, - "train_accuracy_recipient_token": 0.9047619047619048, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_count": 1.0, - "train_accuracy_recipient_token_count_total": 1, - "train_accuracy_recipient_token_find": 0.75, - "train_accuracy_recipient_token_find_total": 4, - "train_accuracy_recipient_token_get": 0.6666666666666666, - "train_accuracy_recipient_token_get_total": 3, + "epoch": 0.83399209486166, + "step": 422, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9166666666666666, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_locate": 1.0, "train_accuracy_recipient_token_locate_total": 1, - "train_total_number_recipient_token": 21 + "train_total_number_recipient_token": 17 }, { - "epoch": 0.7974308300395256, - "grad_norm": 1.608131055033889, - "learning_rate": 5.178394771125969e-07, - "loss": 0.1891, - "step": 269 + "epoch": 0.83399209486166, + "step": 422, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.9, + "train_accuracy_recipient_token_get_total": 10, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.7974308300395256, - "step": 269, - "train_accuracy_first_token_arguments": 0.7058823529411765, - "train_accuracy_first_token_arguments_total": 17, - "train_accuracy_recipient_token": 0.8947368421052632, - "train_accuracy_recipient_token_all": 0.8333333333333334, - "train_accuracy_recipient_token_all_total": 12, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_short": 1.0, - "train_accuracy_recipient_token_short_total": 1, - "train_accuracy_recipient_token_validate": 1.0, - "train_accuracy_recipient_token_validate_total": 1, - "train_total_number_recipient_token": 19 + "epoch": 0.8359683794466403, + "grad_norm": 1.8348800188829963, + "learning_rate": 3.457011581730124e-07, + "loss": 0.2975, + "step": 423 }, { - "epoch": 0.7974308300395256, - "step": 269, - "train_accuracy_first_token_arguments": 0.9047619047619048, - "train_accuracy_first_token_arguments_total": 21, - "train_accuracy_recipient_token": 0.9, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_find": 1.0, + "epoch": 0.8359683794466403, + "step": 423, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_get": 0.6666666666666666, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_recipe": 1.0, - "train_accuracy_recipient_token_recipe_total": 4, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_get": 0.8333333333333334, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.7974308300395256, - "step": 269, - "train_accuracy_first_token_arguments": 0.9047619047619048, - "train_accuracy_first_token_arguments_total": 21, - "train_accuracy_recipient_token": 0.9473684210526315, + "epoch": 0.8359683794466403, + "step": 423, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_book": 1.0, - "train_accuracy_recipient_token_book_total": 1, - "train_accuracy_recipient_token_c": 1.0, - "train_accuracy_recipient_token_c_total": 1, - "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_pill": 1.0, + "train_accuracy_recipient_token_pill_total": 2, "train_total_number_recipient_token": 19 }, { - "epoch": 0.8003952569169961, - "grad_norm": 1.9940648204984408, - "learning_rate": 5.032501373454266e-07, - "loss": 0.2175, - "step": 270 + "epoch": 0.8379446640316206, + "grad_norm": 1.7013901333390404, + "learning_rate": 3.3761285513826627e-07, + "loss": 0.2019, + "step": 424 }, { - "epoch": 0.8003952569169961, - "step": 270, - "train_accuracy_first_token_arguments": 0.9354838709677419, - "train_accuracy_first_token_arguments_total": 31, - "train_accuracy_recipient_token": 0.8125, - "train_accuracy_recipient_token_all": 0.8333333333333334, - "train_accuracy_recipient_token_all_total": 12, - "train_accuracy_recipient_token_crop": 1.0, - "train_accuracy_recipient_token_crop_total": 1, - "train_accuracy_recipient_token_f": 1.0, - "train_accuracy_recipient_token_f_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_har": 0.0, - "train_accuracy_recipient_token_har_total": 1, - "train_total_number_recipient_token": 16 + "epoch": 0.8379446640316206, + "step": 424, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_evaluate": 1.0, + "train_accuracy_recipient_token_evaluate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_plan": 1.0, + "train_accuracy_recipient_token_plan_total": 1, + "train_accuracy_recipient_token_s": 1.0, + "train_accuracy_recipient_token_s_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.8003952569169961, - "step": 270, - "train_accuracy_first_token_arguments": 0.9259259259259259, - "train_accuracy_first_token_arguments_total": 27, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.8888888888888888, - "train_accuracy_recipient_token_all_total": 9, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, - "train_accuracy_recipient_token_land": 1.0, - "train_accuracy_recipient_token_land_total": 4, - "train_accuracy_recipient_token_mount": 1.0, - "train_accuracy_recipient_token_mount_total": 2, + "epoch": 0.8379446640316206, + "step": 424, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 0.5, + "train_accuracy_recipient_token_calculate_total": 2, + "train_accuracy_recipient_token_estimate": 0.5, + "train_accuracy_recipient_token_estimate_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_opt": 1.0, + "train_accuracy_recipient_token_opt_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.8003952569169961, - "step": 270, - "train_accuracy_first_token_arguments": 0.9090909090909091, - "train_accuracy_first_token_arguments_total": 11, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.8399209486166008, + "grad_norm": 2.3984171711273907, + "learning_rate": 3.2961343936955384e-07, + "loss": 0.229, + "step": 425 + }, + { + "epoch": 0.8399209486166008, + "step": 425, + "train_accuracy_recipient_token": 0.8125, + "train_accuracy_recipient_token_all": 0.7272727272727273, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_con": 1.0, + "train_accuracy_recipient_token_con_total": 1, + "train_accuracy_recipient_token_d": 1.0, + "train_accuracy_recipient_token_d_total": 1, + "train_accuracy_recipient_token_lands": 1.0, + "train_accuracy_recipient_token_lands_total": 1, + "train_accuracy_recipient_token_organ": 1.0, + "train_accuracy_recipient_token_organ_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.8399209486166008, + "step": 425, + "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_go": 1.0, - "train_accuracy_recipient_token_go_total": 3, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_get": 0.8888888888888888, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.8033596837944664, - "grad_norm": 1.941454727605692, - "learning_rate": 4.888462296487129e-07, - "loss": 0.2125, - "step": 271 + "epoch": 0.841897233201581, + "grad_norm": 2.1793174841282528, + "learning_rate": 3.217032396915265e-07, + "loss": 0.2292, + "step": 426 }, { - "epoch": 0.8033596837944664, - "step": 271, - "train_accuracy_first_token_arguments": 0.9411764705882353, - "train_accuracy_first_token_arguments_total": 34, + "epoch": 0.841897233201581, + "step": 426, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_dec": 1.0, + "train_accuracy_recipient_token_dec_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { - "epoch": 0.8033596837944664, - "step": 271, - "train_accuracy_first_token_arguments": 0.7407407407407407, - "train_accuracy_first_token_arguments_total": 27, + "epoch": 0.841897233201581, + "step": 426, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.8438735177865613, + "grad_norm": 2.181267379011323, + "learning_rate": 3.1388258126151093e-07, + "loss": 0.177, + "step": 427 + }, + { + "epoch": 0.8438735177865613, + "step": 427, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_validate": 1.0, - "train_accuracy_recipient_token_validate_total": 1, + "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { - "epoch": 0.8033596837944664, - "step": 271, - "train_accuracy_first_token_arguments": 0.7272727272727273, - "train_accuracy_first_token_arguments_total": 33, - "train_accuracy_recipient_token": 0.8947368421052632, + "epoch": 0.8438735177865613, + "step": 427, + "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_c": 1.0, - "train_accuracy_recipient_token_c_total": 2, - "train_accuracy_recipient_token_get": 0.8, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 2, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_l": 1.0, + "train_accuracy_recipient_token_l_total": 1, + "train_accuracy_recipient_token_remove": 1.0, + "train_accuracy_recipient_token_remove_total": 1, + "train_accuracy_recipient_token_write": 1.0, + "train_accuracy_recipient_token_write_total": 1, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.8063241106719368, - "grad_norm": 1.9052339985594693, - "learning_rate": 4.746290916694368e-07, - "loss": 0.2996, - "step": 272 + "epoch": 0.8458498023715415, + "grad_norm": 1.5941096991791754, + "learning_rate": 3.06151785556143e-07, + "loss": 0.0573, + "step": 428 }, { - "epoch": 0.8063241106719368, - "step": 272, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 10, - "train_accuracy_recipient_token": 0.8235294117647058, - "train_accuracy_recipient_token_all": 0.7272727272727273, - "train_accuracy_recipient_token_all_total": 11, + "epoch": 0.8458498023715415, + "step": 428, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_detect": 1.0, + "train_accuracy_recipient_token_detect_total": 1, + "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, - "train_total_number_recipient_token": 17 + "train_total_number_recipient_token": 19 }, { - "epoch": 0.8063241106719368, - "step": 272, - "train_accuracy_first_token_arguments": 0.8333333333333334, - "train_accuracy_first_token_arguments_total": 12, + "epoch": 0.8458498023715415, + "step": 428, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_write": 1.0, + "train_accuracy_recipient_token_write_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.8478260869565217, + "grad_norm": 1.5894279466737562, + "learning_rate": 2.98511170358155e-07, + "loss": 0.1393, + "step": 429 + }, + { + "epoch": 0.8478260869565217, + "step": 429, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 0.9166666666666666, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_paint": 1.0, + "train_accuracy_recipient_token_paint_total": 1, + "train_accuracy_recipient_token_rep": 1.0, + "train_accuracy_recipient_token_rep_total": 1, + "train_accuracy_recipient_token_surface": 1.0, + "train_accuracy_recipient_token_surface_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.8478260869565217, + "step": 429, + "train_accuracy_recipient_token": 0.875, + "train_accuracy_recipient_token_all": 0.8333333333333334, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, - "train_accuracy_recipient_token_translate": 1.0, - "train_accuracy_recipient_token_translate_total": 1, - "train_total_number_recipient_token": 17 + "train_total_number_recipient_token": 16 }, { - "epoch": 0.8063241106719368, - "step": 272, - "train_accuracy_first_token_arguments": 0.8181818181818182, - "train_accuracy_first_token_arguments_total": 11, + "epoch": 0.849802371541502, + "grad_norm": 1.872429124870318, + "learning_rate": 2.9096104974331186e-07, + "loss": 0.2275, + "step": 430 + }, + { + "epoch": 0.849802371541502, + "step": 430, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.849802371541502, + "step": 430, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 6, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 2, + "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { - "epoch": 0.8092885375494071, - "grad_norm": 2.933908602104762, - "learning_rate": 4.6060004370984763e-07, - "loss": 0.3296, - "step": 273 + "epoch": 0.8517786561264822, + "grad_norm": 1.9625847182922385, + "learning_rate": 2.8350173406749975e-07, + "loss": 0.1982, + "step": 431 }, { - "epoch": 0.8092885375494071, - "step": 273, - "train_accuracy_first_token_arguments": 0.9629629629629629, - "train_accuracy_first_token_arguments_total": 27, + "epoch": 0.8517786561264822, + "step": 431, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_convert": 1.0, - "train_accuracy_recipient_token_convert_total": 3, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_order": 1.0, + "train_accuracy_recipient_token_order_total": 1, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 2, "train_total_number_recipient_token": 17 }, { - "epoch": 0.8092885375494071, - "step": 273, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 9, - "train_accuracy_recipient_token": 0.7857142857142857, - "train_accuracy_recipient_token_all": 0.7, + "epoch": 0.8517786561264822, + "step": 431, + "train_accuracy_recipient_token": 0.875, + "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_data": 1.0, - "train_accuracy_recipient_token_data_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.8537549407114624, + "grad_norm": 2.051731787985664, + "learning_rate": 2.761335299539708e-07, + "loss": 0.249, + "step": 432 + }, + { + "epoch": 0.8537549407114624, + "step": 432, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_f": 1.0, + "train_accuracy_recipient_token_f_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, - "train_total_number_recipient_token": 14 + "train_accuracy_recipient_token_mul": 1.0, + "train_accuracy_recipient_token_mul_total": 1, + "train_accuracy_recipient_token_we": 1.0, + "train_accuracy_recipient_token_we_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.8092885375494071, - "step": 273, - "train_accuracy_first_token_arguments": 0.9, - "train_accuracy_first_token_arguments_total": 20, + "epoch": 0.8537549407114624, + "step": 432, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_histor": 1.0, + "train_accuracy_recipient_token_histor_total": 1, + "train_accuracy_recipient_token_land": 1.0, + "train_accuracy_recipient_token_land_total": 3, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.8557312252964426, + "grad_norm": 1.5632631855936685, + "learning_rate": 2.688567402807357e-07, + "loss": 0.1849, + "step": 433 + }, + { + "epoch": 0.8557312252964426, + "step": 433, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.8333333333333334, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.8557312252964426, + "step": 433, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_rep": 1.0, + "train_accuracy_recipient_token_rep_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.8122529644268774, - "grad_norm": 2.787651925675771, - "learning_rate": 4.467603886048452e-07, - "loss": 0.2215, - "step": 274 + "epoch": 0.857707509881423, + "grad_norm": 1.8054699895841433, + "learning_rate": 2.6167166416811745e-07, + "loss": 0.1723, + "step": 434 }, { - "epoch": 0.8122529644268774, - "step": 274, - "train_accuracy_first_token_arguments": 0.9583333333333334, - "train_accuracy_first_token_arguments_total": 24, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.857707509881423, + "step": 434, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_design": 1.0, + "train_accuracy_recipient_token_design_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 4, - "train_accuracy_recipient_token_h": 1.0, - "train_accuracy_recipient_token_h_total": 4, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_find_total": 5, + "train_accuracy_recipient_token_play": 1.0, + "train_accuracy_recipient_token_play_total": 1, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.8122529644268774, - "step": 274, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 16, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compare": 0.0, - "train_accuracy_recipient_token_compare_total": 1, + "epoch": 0.857707509881423, + "step": 434, + "train_accuracy_recipient_token": 0.7894736842105263, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_train": 0.25, + "train_accuracy_recipient_token_train_total": 4, "train_total_number_recipient_token": 19 }, { - "epoch": 0.8122529644268774, - "step": 274, - "train_accuracy_first_token_arguments": 0.9473684210526315, - "train_accuracy_first_token_arguments_total": 19, + "epoch": 0.8596837944664032, + "grad_norm": 1.723792192639983, + "learning_rate": 2.5457859696645245e-07, + "loss": 0.1502, + "step": 435 + }, + { + "epoch": 0.8596837944664032, + "step": 435, "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_crop": 1.0, - "train_accuracy_recipient_token_crop_total": 1, - "train_accuracy_recipient_token_f": 1.0, - "train_accuracy_recipient_token_f_total": 1, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 2, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_har": 0.0, - "train_accuracy_recipient_token_har_total": 1, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_mach": 1.0, + "train_accuracy_recipient_token_mach_total": 3, "train_total_number_recipient_token": 18 }, { - "epoch": 0.8152173913043478, - "grad_norm": 2.0435081993694, - "learning_rate": 4.331114116009938e-07, - "loss": 0.1421, - "step": 275 - }, - { - "epoch": 0.8152173913043478, - "step": 275, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 10, + "epoch": 0.8596837944664032, + "step": 435, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, @@ -15282,1858 +16744,1763 @@ "train_total_number_recipient_token": 20 }, { - "epoch": 0.8152173913043478, - "step": 275, - "train_accuracy_first_token_arguments": 0.9333333333333333, - "train_accuracy_first_token_arguments_total": 30, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 9, - "train_total_number_recipient_token": 20 + "epoch": 0.8616600790513834, + "grad_norm": 1.5466863752221973, + "learning_rate": 2.4757783024395244e-07, + "loss": 0.1792, + "step": 436 }, { - "epoch": 0.8152173913043478, - "step": 275, - "train_accuracy_first_token_arguments": 0.5925925925925926, - "train_accuracy_first_token_arguments_total": 27, - "train_accuracy_recipient_token": 0.8823529411764706, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.8616600790513834, + "step": 436, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_get": 0.5, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_sum": 1.0, - "train_accuracy_recipient_token_sum_total": 1, - "train_accuracy_recipient_token_translate": 1.0, - "train_accuracy_recipient_token_translate_total": 1, - "train_total_number_recipient_token": 17 - }, - { - "epoch": 0.8181818181818182, - "grad_norm": 2.30706005395235, - "learning_rate": 4.196543802371641e-07, - "loss": 0.2433, - "step": 276 - }, - { - "epoch": 0.8181818181818182, - "step": 276, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.9411764705882353, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 6, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_order": 1.0, + "train_accuracy_recipient_token_order_total": 1, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.8181818181818182, - "step": 276, - "train_accuracy_first_token_arguments": 0.7333333333333333, - "train_accuracy_first_token_arguments_total": 45, + "epoch": 0.8616600790513834, + "step": 436, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_c": 1.0, - "train_accuracy_recipient_token_c_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, - "train_accuracy_recipient_token_rail": 1.0, - "train_accuracy_recipient_token_rail_total": 4, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_p": 1.0, + "train_accuracy_recipient_token_p_total": 3, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.8181818181818182, - "step": 276, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 18, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.8181818181818182, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 7, - "train_total_number_recipient_token": 18 + "epoch": 0.8636363636363636, + "grad_norm": 2.6743224940140915, + "learning_rate": 2.4066965177471645e-07, + "loss": 0.3654, + "step": 437 }, { - "epoch": 0.8211462450592886, - "grad_norm": 1.9811813651011425, - "learning_rate": 4.063905442268201e-07, - "loss": 0.2254, - "step": 277 + "epoch": 0.8636363636363636, + "step": 437, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_accuracy_recipient_token_track": 1.0, + "train_accuracy_recipient_token_track_total": 1, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.8211462450592886, - "step": 277, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 28, - "train_accuracy_recipient_token": 0.8823529411764706, + "epoch": 0.8636363636363636, + "step": 437, + "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_find": 0.75, - "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, - "train_accuracy_recipient_token_schedule": 1.0, - "train_accuracy_recipient_token_schedule_total": 1, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.8211462450592886, - "step": 277, - "train_accuracy_first_token_arguments": 0.8095238095238095, - "train_accuracy_first_token_arguments_total": 21, + "epoch": 0.8656126482213439, + "grad_norm": 2.1181583086625415, + "learning_rate": 2.338543455269046e-07, + "loss": 0.3387, + "step": 438 + }, + { + "epoch": 0.8656126482213439, + "step": 438, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_motor": 1.0, - "train_accuracy_recipient_token_motor_total": 3, + "train_accuracy_recipient_token_ins": 1.0, + "train_accuracy_recipient_token_ins_total": 1, + "train_accuracy_recipient_token_remove": 1.0, + "train_accuracy_recipient_token_remove_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.8211462450592886, - "step": 277, - "train_accuracy_first_token_arguments": 0.9130434782608695, - "train_accuracy_first_token_arguments_total": 23, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.8656126482213439, + "step": 438, + "train_accuracy_recipient_token": 0.9130434782608695, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 4, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_calculate": 0.6666666666666666, + "train_accuracy_recipient_token_calculate_total": 3, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_m": 1.0, + "train_accuracy_recipient_token_m_total": 4, + "train_accuracy_recipient_token_s": 1.0, + "train_accuracy_recipient_token_s_total": 2, + "train_total_number_recipient_token": 23 }, { - "epoch": 0.8241106719367589, - "grad_norm": 1.7163941442053223, - "learning_rate": 3.9332113534196194e-07, - "loss": 0.2561, - "step": 278 + "epoch": 0.8675889328063241, + "grad_norm": 2.0841614154281136, + "learning_rate": 2.271321916510627e-07, + "loss": 0.1456, + "step": 439 }, { - "epoch": 0.8241106719367589, - "step": 278, - "train_accuracy_first_token_arguments": 0.9444444444444444, - "train_accuracy_first_token_arguments_total": 18, - "train_accuracy_recipient_token": 0.8947368421052632, + "epoch": 0.8675889328063241, + "step": 439, + "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compare": 0.0, - "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_get": 0.875, - "train_accuracy_recipient_token_get_total": 8, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_all_total": 8, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 2, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_join": 0.0, + "train_accuracy_recipient_token_join_total": 1, + "train_total_number_recipient_token": 16 }, { - "epoch": 0.8241106719367589, - "step": 278, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 25, - "train_accuracy_recipient_token": 0.8235294117647058, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 0.5, + "epoch": 0.8675889328063241, + "step": 439, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 17 - }, - { - "epoch": 0.8241106719367589, - "step": 278, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 15, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 9, - "train_accuracy_recipient_token_be": 0.6666666666666666, - "train_accuracy_recipient_token_be_total": 3, - "train_accuracy_recipient_token_find": 0.8, - "train_accuracy_recipient_token_find_total": 5, - "train_accuracy_recipient_token_suggest": 1.0, - "train_accuracy_recipient_token_suggest_total": 1, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_locate": 1.0, + "train_accuracy_recipient_token_locate_total": 1, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.8270750988142292, - "grad_norm": 1.6379930568149086, - "learning_rate": 3.80447367298738e-07, - "loss": 0.1022, - "step": 279 + "epoch": 0.8695652173913043, + "grad_norm": 1.2831293107819657, + "learning_rate": 2.205034664686076e-07, + "loss": 0.1333, + "step": 440 }, { - "epoch": 0.8270750988142292, - "step": 279, - "train_accuracy_first_token_arguments": 0.8823529411764706, - "train_accuracy_first_token_arguments_total": 17, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_create": 1.0, - "train_accuracy_recipient_token_create_total": 1, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 1, + "epoch": 0.8695652173913043, + "step": 440, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 5, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_order": 1.0, + "train_accuracy_recipient_token_order_total": 1, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 1, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.8270750988142292, - "step": 279, - "train_accuracy_first_token_arguments": 0.9047619047619048, - "train_accuracy_first_token_arguments_total": 21, + "epoch": 0.8695652173913043, + "step": 440, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_gr": 1.0, - "train_accuracy_recipient_token_gr_total": 2, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_interpret": 1.0, - "train_accuracy_recipient_token_interpret_total": 1, - "train_accuracy_recipient_token_remove": 1.0, - "train_accuracy_recipient_token_remove_total": 1, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_text": 1.0, + "train_accuracy_recipient_token_text_total": 6, + "train_total_number_recipient_token": 21 }, { - "epoch": 0.8270750988142292, - "step": 279, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 22, + "epoch": 0.8715415019762845, + "grad_norm": 2.007799864507481, + "learning_rate": 2.1396844246046904e-07, + "loss": 0.1843, + "step": 441 + }, + { + "epoch": 0.8715415019762845, + "step": 441, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_check": 1.0, - "train_accuracy_recipient_token_check_total": 2, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 1, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.8715415019762845, + "step": 441, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_draw": 1.0, + "train_accuracy_recipient_token_draw_total": 5, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 20 }, { - "epoch": 0.8300395256916996, - "grad_norm": 1.6541675532013322, - "learning_rate": 3.677704356447254e-07, - "loss": 0.1926, - "step": 280 + "epoch": 0.8735177865612648, + "grad_norm": 2.045124495201404, + "learning_rate": 2.075273882558873e-07, + "loss": 0.3035, + "step": 442 }, { - "epoch": 0.8300395256916996, - "step": 280, - "train_accuracy_first_token_arguments": 0.7903225806451613, - "train_accuracy_first_token_arguments_total": 62, + "epoch": 0.8735177865612648, + "step": 442, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_f": 1.0, - "train_accuracy_recipient_token_f_total": 2, - "train_accuracy_recipient_token_generate": 1.0, - "train_accuracy_recipient_token_generate_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_port": 1.0, - "train_accuracy_recipient_token_port_total": 1, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_get_total": 10, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.8300395256916996, - "step": 280, - "train_accuracy_first_token_arguments": 0.9833333333333333, - "train_accuracy_first_token_arguments_total": 120, - "train_accuracy_recipient_token": 0.8823529411764706, - "train_accuracy_recipient_token_all": 0.8181818181818182, + "epoch": 0.8735177865612648, + "step": 442, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 3, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 2, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 1, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_locate": 1.0, + "train_accuracy_recipient_token_locate_total": 1, + "train_accuracy_recipient_token_s": 1.0, + "train_accuracy_recipient_token_s_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.8300395256916996, - "step": 280, - "train_accuracy_first_token_arguments": 0.8846153846153846, - "train_accuracy_first_token_arguments_total": 26, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.8754940711462451, + "grad_norm": 2.0761359917977593, + "learning_rate": 2.0118056862137358e-07, + "loss": 0.1576, + "step": 443 + }, + { + "epoch": 0.8754940711462451, + "step": 443, + "train_accuracy_recipient_token": 0.8666666666666667, "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 9, - "train_accuracy_recipient_token_calculate": 1.0, - "train_accuracy_recipient_token_calculate_total": 3, - "train_accuracy_recipient_token_convert": 1.0, - "train_accuracy_recipient_token_convert_total": 1, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_design": 0.0, + "train_accuracy_recipient_token_design_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_extract": 0.5, + "train_accuracy_recipient_token_extract_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_set": 1.0, - "train_accuracy_recipient_token_set_total": 1, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_get_total": 1, + "train_total_number_recipient_token": 15 }, { - "epoch": 0.8330039525691699, - "grad_norm": 2.430973200540749, - "learning_rate": 3.5529151764790715e-07, - "loss": 0.282, - "step": 281 + "epoch": 0.8754940711462451, + "step": 443, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_att": 0.0, + "train_accuracy_recipient_token_att_total": 1, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_ex": 1.0, + "train_accuracy_recipient_token_ex_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.8330039525691699, - "step": 281, - "train_accuracy_first_token_arguments": 0.9166666666666666, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.9375, + "epoch": 0.8774703557312253, + "grad_norm": 2.2388845065518983, + "learning_rate": 1.9492824444982378e-07, + "loss": 0.3765, + "step": 444 + }, + { + "epoch": 0.8774703557312253, + "step": 444, + "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_aqu": 1.0, - "train_accuracy_recipient_token_aqu_total": 1, - "train_accuracy_recipient_token_check": 1.0, - "train_accuracy_recipient_token_check_total": 1, - "train_accuracy_recipient_token_fish": 1.0, - "train_accuracy_recipient_token_fish_total": 3, - "train_total_number_recipient_token": 16 + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_learn": 1.0, + "train_accuracy_recipient_token_learn_total": 1, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.8330039525691699, - "step": 281, - "train_accuracy_first_token_arguments": 0.9393939393939394, - "train_accuracy_first_token_arguments_total": 33, + "epoch": 0.8774703557312253, + "step": 444, "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9166666666666666, - "train_accuracy_recipient_token_all_total": 12, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_wild": 1.0, - "train_accuracy_recipient_token_wild_total": 2, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { - "epoch": 0.8330039525691699, - "step": 281, - "train_accuracy_first_token_arguments": 0.8181818181818182, - "train_accuracy_first_token_arguments_total": 11, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.8794466403162056, + "grad_norm": 1.1600613922342395, + "learning_rate": 1.887706727497965e-07, + "loss": 0.0662, + "step": 445 + }, + { + "epoch": 0.8794466403162056, + "step": 445, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.75, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.8794466403162056, + "step": 445, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 3, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 10, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 2, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.8359683794466403, - "grad_norm": 2.0245615896487577, - "learning_rate": 3.43011772187343e-07, - "loss": 0.3178, - "step": 282 + "epoch": 0.8814229249011858, + "grad_norm": 2.141770501735159, + "learning_rate": 1.8270810663494591e-07, + "loss": 0.2166, + "step": 446 }, { - "epoch": 0.8359683794466403, - "step": 282, - "train_accuracy_first_token_arguments": 0.8947368421052632, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.9444444444444444, + "epoch": 0.8814229249011858, + "step": 446, + "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_accuracy_recipient_token_sort": 0.0, + "train_accuracy_recipient_token_sort_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.8814229249011858, + "step": 446, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 0.9230769230769231, + "train_accuracy_recipient_token_all_total": 13, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_restaurant": 1.0, + "train_accuracy_recipient_token_restaurant_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.883399209486166, + "grad_norm": 1.668725614891495, + "learning_rate": 1.767407953136202e-07, + "loss": 0.2025, + "step": 447 + }, + { + "epoch": 0.883399209486166, + "step": 447, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 9, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.883399209486166, + "step": 447, + "train_accuracy_recipient_token": 0.7777777777777778, + "train_accuracy_recipient_token_all": 0.75, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_building": 0.5, + "train_accuracy_recipient_token_building_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_stat": 1.0, + "train_accuracy_recipient_token_stat_total": 2, "train_total_number_recipient_token": 18 }, { - "epoch": 0.8359683794466403, - "step": 282, - "train_accuracy_first_token_arguments": 0.8571428571428571, - "train_accuracy_first_token_arguments_total": 14, + "epoch": 0.8853754940711462, + "grad_norm": 2.0839098091259984, + "learning_rate": 1.7086898407861486e-07, + "loss": 0.3808, + "step": 448 + }, + { + "epoch": 0.8853754940711462, + "step": 448, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_compile": 1.0, + "train_accuracy_recipient_token_compile_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_pill": 1.0, - "train_accuracy_recipient_token_pill_total": 2, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, + "train_accuracy_recipient_token_restaurant": 1.0, + "train_accuracy_recipient_token_restaurant_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.8359683794466403, - "step": 282, - "train_accuracy_first_token_arguments": 0.8947368421052632, - "train_accuracy_first_token_arguments_total": 19, + "epoch": 0.8853754940711462, + "step": 448, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_evaluate": 1.0, - "train_accuracy_recipient_token_evaluate_total": 1, + "train_accuracy_recipient_token_draw": 1.0, + "train_accuracy_recipient_token_draw_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_plan": 1.0, - "train_accuracy_recipient_token_plan_total": 1, - "train_accuracy_recipient_token_suggest": 1.0, - "train_accuracy_recipient_token_suggest_total": 1, - "train_total_number_recipient_token": 18 + "train_total_number_recipient_token": 20 }, { - "epoch": 0.8389328063241107, - "grad_norm": 2.1514655537590763, - "learning_rate": 3.3093233964554464e-07, - "loss": 0.2761, - "step": 283 + "epoch": 0.8873517786561265, + "grad_norm": 2.6050856048557867, + "learning_rate": 1.6509291429709224e-07, + "loss": 0.2345, + "step": 449 }, { - "epoch": 0.8389328063241107, - "step": 283, - "train_accuracy_first_token_arguments": 0.9375, - "train_accuracy_first_token_arguments_total": 16, - "train_accuracy_recipient_token": 0.8888888888888888, + "epoch": 0.8873517786561265, + "step": 449, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_calculate": 0.5, - "train_accuracy_recipient_token_calculate_total": 2, - "train_accuracy_recipient_token_estimate": 0.5, - "train_accuracy_recipient_token_estimate_total": 2, + "train_accuracy_recipient_token_exp": 1.0, + "train_accuracy_recipient_token_exp_total": 3, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_optimize": 1.0, - "train_accuracy_recipient_token_optimize_total": 1, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.8389328063241107, - "step": 283, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 27, - "train_accuracy_recipient_token": 0.75, - "train_accuracy_recipient_token_all": 0.6363636363636364, + "epoch": 0.8873517786561265, + "step": 449, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_conduct": 1.0, - "train_accuracy_recipient_token_conduct_total": 1, - "train_accuracy_recipient_token_d": 1.0, - "train_accuracy_recipient_token_d_total": 1, - "train_accuracy_recipient_token_landscape": 1.0, - "train_accuracy_recipient_token_landscape_total": 1, - "train_accuracy_recipient_token_organ": 1.0, - "train_accuracy_recipient_token_organ_total": 1, - "train_total_number_recipient_token": 16 - }, - { - "epoch": 0.8389328063241107, - "step": 283, - "train_accuracy_first_token_arguments": 0.8421052631578947, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 0.8888888888888888, - "train_accuracy_recipient_token_get_total": 9, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_plant": 1.0, + "train_accuracy_recipient_token_plant_total": 3, "train_total_number_recipient_token": 19 }, { - "epoch": 0.841897233201581, - "grad_norm": 2.2087100686535135, - "learning_rate": 3.190543418025749e-07, - "loss": 0.2386, - "step": 284 + "epoch": 0.8893280632411067, + "grad_norm": 1.8361106243035428, + "learning_rate": 1.59412823400657e-07, + "loss": 0.2602, + "step": 450 }, { - "epoch": 0.841897233201581, - "step": 284, - "train_accuracy_first_token_arguments": 0.8666666666666667, - "train_accuracy_first_token_arguments_total": 15, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.8893280632411067, + "step": 450, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_dec": 1.0, - "train_accuracy_recipient_token_dec_total": 1, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.8893280632411067, + "step": 450, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_bird": 1.0, + "train_accuracy_recipient_token_bird_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_ident": 0.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.841897233201581, - "step": 284, - "train_accuracy_first_token_arguments": 0.8571428571428571, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 0.8947368421052632, + "epoch": 0.8913043478260869, + "grad_norm": 2.2065101723440863, + "learning_rate": 1.538289448755989e-07, + "loss": 0.2448, + "step": 451 + }, + { + "epoch": 0.8913043478260869, + "step": 451, + "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_get": 0.875, - "train_accuracy_recipient_token_get_total": 8, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { - "epoch": 0.841897233201581, - "step": 284, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 10, - "train_accuracy_recipient_token": 0.9473684210526315, + "epoch": 0.8913043478260869, + "step": 451, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 0.8888888888888888, - "train_accuracy_recipient_token_get_total": 9, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.8448616600790514, - "grad_norm": 2.1863171999841655, - "learning_rate": 3.0737888173187067e-07, - "loss": 0.1506, - "step": 285 + "epoch": 0.8932806324110671, + "grad_norm": 2.529268779693517, + "learning_rate": 1.483415082532938e-07, + "loss": 0.3118, + "step": 452 }, { - "epoch": 0.8448616600790514, - "step": 285, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 11, - "train_accuracy_recipient_token": 0.9411764705882353, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, + "epoch": 0.8932806324110671, + "step": 452, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8333333333333334, + "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_l": 1.0, - "train_accuracy_recipient_token_l_total": 1, - "train_accuracy_recipient_token_remove": 1.0, - "train_accuracy_recipient_token_remove_total": 1, - "train_accuracy_recipient_token_write": 1.0, - "train_accuracy_recipient_token_write_total": 1, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.8448616600790514, - "step": 285, - "train_accuracy_first_token_arguments": 0.6666666666666666, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.9473684210526315, + "epoch": 0.8932806324110671, + "step": 452, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_book": 1.0, - "train_accuracy_recipient_token_book_total": 1, - "train_accuracy_recipient_token_detect": 1.0, - "train_accuracy_recipient_token_detect_total": 1, - "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_list": 1.0, + "train_accuracy_recipient_token_list_total": 1, + "train_accuracy_recipient_token_stat": 1.0, + "train_accuracy_recipient_token_stat_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.8448616600790514, - "step": 285, - "train_accuracy_first_token_arguments": 0.76, - "train_accuracy_first_token_arguments_total": 25, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.8952569169960475, + "grad_norm": 1.9479548952801957, + "learning_rate": 1.4295073910076757e-07, + "loss": 0.3757, + "step": 453 + }, + { + "epoch": 0.8952569169960475, + "step": 453, + "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 7, - "train_accuracy_recipient_token_write": 1.0, - "train_accuracy_recipient_token_write_total": 1, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_object": 1.0, + "train_accuracy_recipient_token_object_total": 1, + "train_accuracy_recipient_token_people": 0.5, + "train_accuracy_recipient_token_people_total": 2, "train_total_number_recipient_token": 20 }, { - "epoch": 0.8478260869565217, - "grad_norm": 1.3956004649176736, - "learning_rate": 2.9590704369780313e-07, - "loss": 0.1174, - "step": 286 + "epoch": 0.8952569169960475, + "step": 453, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_verify": 1.0, + "train_accuracy_recipient_token_verify_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.8478260869565217, - "step": 286, - "train_accuracy_first_token_arguments": 0.8235294117647058, - "train_accuracy_first_token_arguments_total": 17, - "train_accuracy_recipient_token": 0.95, - "train_accuracy_recipient_token_all": 0.9166666666666666, - "train_accuracy_recipient_token_all_total": 12, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_paint": 1.0, - "train_accuracy_recipient_token_paint_total": 1, - "train_accuracy_recipient_token_rep": 1.0, - "train_accuracy_recipient_token_rep_total": 1, - "train_accuracy_recipient_token_surface": 1.0, - "train_accuracy_recipient_token_surface_total": 1, - "train_total_number_recipient_token": 20 + "epoch": 0.8972332015810277, + "grad_norm": 3.204511061904964, + "learning_rate": 1.3765685901142718e-07, + "loss": 0.2914, + "step": 454 }, { - "epoch": 0.8478260869565217, - "step": 286, - "train_accuracy_first_token_arguments": 0.9333333333333333, - "train_accuracy_first_token_arguments_total": 15, - "train_accuracy_recipient_token": 0.875, - "train_accuracy_recipient_token_all": 0.8333333333333334, - "train_accuracy_recipient_token_all_total": 12, + "epoch": 0.8972332015810277, + "step": 454, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_building": 1.0, + "train_accuracy_recipient_token_building_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, - "train_accuracy_recipient_token_search": 1.0, - "train_accuracy_recipient_token_search_total": 1, - "train_total_number_recipient_token": 16 + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.8478260869565217, - "step": 286, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.95, + "epoch": 0.8972332015810277, + "step": 454, + "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_book": 0.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 8, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.8507905138339921, - "grad_norm": 1.8154026629925726, - "learning_rate": 2.8463989305498596e-07, - "loss": 0.2343, - "step": 287 + "epoch": 0.8992094861660079, + "grad_norm": 1.3138994217003859, + "learning_rate": 1.3246008559594708e-07, + "loss": 0.1385, + "step": 455 }, { - "epoch": 0.8507905138339921, - "step": 287, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 14, + "epoch": 0.8992094861660079, + "step": 455, "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.8992094861660079, + "step": 455, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 9, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 2, "train_total_number_recipient_token": 19 }, { - "epoch": 0.8507905138339921, - "step": 287, - "train_accuracy_first_token_arguments": 0.8260869565217391, - "train_accuracy_first_token_arguments_total": 23, - "train_accuracy_recipient_token": 0.8823529411764706, - "train_accuracy_recipient_token_all": 0.8, + "epoch": 0.9011857707509882, + "grad_norm": 2.8924372432420418, + "learning_rate": 1.273606324733284e-07, + "loss": 0.3419, + "step": 456 + }, + { + "epoch": 0.9011857707509882, + "step": 456, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 0.6666666666666666, + "train_accuracy_recipient_token_check_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_order": 1.0, - "train_accuracy_recipient_token_order_total": 1, - "train_accuracy_recipient_token_schedule": 1.0, - "train_accuracy_recipient_token_schedule_total": 2, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.8507905138339921, - "step": 287, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 10, - "train_accuracy_recipient_token": 0.875, - "train_accuracy_recipient_token_all": 0.8, + "epoch": 0.9011857707509882, + "step": 456, + "train_accuracy_recipient_token": 0.9333333333333333, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 6, - "train_total_number_recipient_token": 16 + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_recogn": 1.0, + "train_accuracy_recipient_token_recogn_total": 1, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 1, + "train_total_number_recipient_token": 15 }, { - "epoch": 0.8537549407114624, - "grad_norm": 2.1946988449253766, - "learning_rate": 2.7357847614933876e-07, - "loss": 0.2904, - "step": 288 + "epoch": 0.9031620553359684, + "grad_norm": 2.118134582049429, + "learning_rate": 1.223587092621162e-07, + "loss": 0.3318, + "step": 457 }, { - "epoch": 0.8537549407114624, - "step": 288, - "train_accuracy_first_token_arguments": 0.8666666666666667, - "train_accuracy_first_token_arguments_total": 15, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.8, + "epoch": 0.9031620553359684, + "step": 457, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_f": 1.0, - "train_accuracy_recipient_token_f_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_mul": 1.0, - "train_accuracy_recipient_token_mul_total": 1, - "train_accuracy_recipient_token_weed": 1.0, - "train_accuracy_recipient_token_weed_total": 1, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.8537549407114624, - "step": 288, - "train_accuracy_first_token_arguments": 0.8846153846153846, - "train_accuracy_first_token_arguments_total": 26, - "train_accuracy_recipient_token": 0.95, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, + "epoch": 0.9031620553359684, + "step": 457, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_histor": 1.0, - "train_accuracy_recipient_token_histor_total": 1, - "train_accuracy_recipient_token_land": 1.0, - "train_accuracy_recipient_token_land_total": 3, + "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 20 }, { - "epoch": 0.8537549407114624, - "step": 288, - "train_accuracy_first_token_arguments": 0.85, - "train_accuracy_first_token_arguments_total": 40, - "train_accuracy_recipient_token": 0.8947368421052632, - "train_accuracy_recipient_token_all": 0.9166666666666666, - "train_accuracy_recipient_token_all_total": 12, - "train_accuracy_recipient_token_check": 0.0, - "train_accuracy_recipient_token_check_total": 1, - "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_schedule": 1.0, - "train_accuracy_recipient_token_schedule_total": 1, - "train_total_number_recipient_token": 19 + "epoch": 0.9051383399209486, + "grad_norm": 1.7617993832600736, + "learning_rate": 1.1745452157178206e-07, + "loss": 0.1621, + "step": 458 }, { - "epoch": 0.8567193675889329, - "grad_norm": 1.8692598511995904, - "learning_rate": 2.6272382022091704e-07, - "loss": 0.2141, - "step": 289 + "epoch": 0.9051383399209486, + "step": 458, + "train_accuracy_recipient_token": 0.875, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_event": 1.0, + "train_accuracy_recipient_token_event_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 16 }, { - "epoch": 0.8567193675889329, - "step": 289, - "train_accuracy_first_token_arguments": 0.9310344827586207, - "train_accuracy_first_token_arguments_total": 29, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, + "epoch": 0.9051383399209486, + "step": 458, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_car": 1.0, + "train_accuracy_recipient_token_car_total": 1, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_repair": 1.0, - "train_accuracy_recipient_token_repair_total": 1, + "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { - "epoch": 0.8567193675889329, - "step": 289, - "train_accuracy_first_token_arguments": 0.75, - "train_accuracy_first_token_arguments_total": 36, - "train_accuracy_recipient_token": 0.9411764705882353, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_design": 1.0, - "train_accuracy_recipient_token_design_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 5, - "train_accuracy_recipient_token_play": 1.0, - "train_accuracy_recipient_token_play_total": 1, - "train_total_number_recipient_token": 17 + "epoch": 0.9071146245059288, + "grad_norm": 1.5552007622687527, + "learning_rate": 1.1264827099427417e-07, + "loss": 0.1899, + "step": 459 }, { - "epoch": 0.8567193675889329, - "step": 289, - "train_accuracy_first_token_arguments": 0.92, - "train_accuracy_first_token_arguments_total": 25, - "train_accuracy_recipient_token": 0.8947368421052632, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_book": 1.0, - "train_accuracy_recipient_token_book_total": 1, + "epoch": 0.9071146245059288, + "step": 459, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 2, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_train": 0.75, - "train_accuracy_recipient_token_train_total": 4, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_stat": 1.0, + "train_accuracy_recipient_token_stat_total": 3, "train_total_number_recipient_token": 19 }, { - "epoch": 0.8596837944664032, - "grad_norm": 1.4764081056636353, - "learning_rate": 2.520769333085141e-07, - "loss": 0.1611, - "step": 290 + "epoch": 0.9071146245059288, + "step": 459, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.8596837944664032, - "step": 290, - "train_accuracy_first_token_arguments": 0.9626168224299065, - "train_accuracy_first_token_arguments_total": 107, - "train_accuracy_recipient_token": 0.8333333333333334, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_check": 0.5, - "train_accuracy_recipient_token_check_total": 2, - "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, - "train_accuracy_recipient_token_mach": 0.6666666666666666, - "train_accuracy_recipient_token_mach_total": 3, - "train_total_number_recipient_token": 18 + "epoch": 0.9090909090909091, + "grad_norm": 2.050523711336241, + "learning_rate": 1.0794015509572819e-07, + "loss": 0.171, + "step": 460 }, { - "epoch": 0.8596837944664032, - "step": 290, - "train_accuracy_first_token_arguments": 0.8666666666666667, - "train_accuracy_first_token_arguments_total": 15, - "train_accuracy_recipient_token": 0.95, + "epoch": 0.9090909090909091, + "step": 460, + "train_accuracy_recipient_token": 0.9047619047619048, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 0.9, - "train_accuracy_recipient_token_get_total": 10, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 2, + "train_accuracy_recipient_token_search": 0.0, + "train_accuracy_recipient_token_search_total": 1, + "train_total_number_recipient_token": 21 }, { - "epoch": 0.8596837944664032, - "step": 290, - "train_accuracy_first_token_arguments": 0.75, - "train_accuracy_first_token_arguments_total": 16, - "train_accuracy_recipient_token": 0.8947368421052632, + "epoch": 0.9090909090909091, + "step": 460, + "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_check": 1.0, - "train_accuracy_recipient_token_check_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_order": 1.0, - "train_accuracy_recipient_token_order_total": 1, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 2, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_count": 1.0, + "train_accuracy_recipient_token_count_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 2, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.8626482213438735, - "grad_norm": 1.920082186381157, - "learning_rate": 2.4163880415604913e-07, - "loss": 0.266, - "step": 291 + "epoch": 0.9110671936758893, + "grad_norm": 2.0269935707637354, + "learning_rate": 1.0333036740834857e-07, + "loss": 0.2898, + "step": 461 }, { - "epoch": 0.8626482213438735, - "step": 291, - "train_accuracy_first_token_arguments": 0.7333333333333333, - "train_accuracy_first_token_arguments_total": 15, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, + "epoch": 0.9110671936758893, + "step": 461, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_pig": 1.0, - "train_accuracy_recipient_token_pig_total": 3, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_gr": 1.0, + "train_accuracy_recipient_token_gr_total": 3, + "train_accuracy_recipient_token_ident": 0.5, + "train_accuracy_recipient_token_ident_total": 2, "train_total_number_recipient_token": 19 }, { - "epoch": 0.8626482213438735, - "step": 291, - "train_accuracy_first_token_arguments": 0.9680851063829787, - "train_accuracy_first_token_arguments_total": 94, - "train_accuracy_recipient_token": 0.9411764705882353, + "epoch": 0.9110671936758893, + "step": 461, + "train_accuracy_recipient_token": 0.9047619047619048, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_calculate": 1.0, - "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_get": 0.9090909090909091, + "train_accuracy_recipient_token_get_total": 11, + "train_total_number_recipient_token": 21 + }, + { + "epoch": 0.9130434782608695, + "grad_norm": 1.7100530929084616, + "learning_rate": 9.881909742245177e-08, + "loss": 0.2431, + "step": 462 + }, + { + "epoch": 0.9130434782608695, + "step": 462, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_schedule": 1.0, - "train_accuracy_recipient_token_schedule_total": 1, - "train_accuracy_recipient_token_track": 1.0, - "train_accuracy_recipient_token_track_total": 1, + "train_accuracy_recipient_token_verify": 1.0, + "train_accuracy_recipient_token_verify_total": 1, "train_total_number_recipient_token": 17 }, { - "epoch": 0.8626482213438735, - "step": 291, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 10, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_c": 1.0, - "train_accuracy_recipient_token_c_total": 1, - "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, + "epoch": 0.9130434782608695, + "step": 462, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_arr": 0.0, + "train_accuracy_recipient_token_arr_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_house": 1.0, + "train_accuracy_recipient_token_house_total": 3, + "train_accuracy_recipient_token_lands": 1.0, + "train_accuracy_recipient_token_lands_total": 1, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.8656126482213439, - "grad_norm": 2.4496922640460923, - "learning_rate": 2.3141040212074445e-07, - "loss": 0.3954, - "step": 292 + "epoch": 0.9150197628458498, + "grad_norm": 1.5464732166200588, + "learning_rate": 9.440653057867816e-08, + "loss": 0.1198, + "step": 463 }, { - "epoch": 0.8656126482213439, - "step": 292, - "train_accuracy_first_token_arguments": 0.8, - "train_accuracy_first_token_arguments_total": 20, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.9150197628458498, + "step": 463, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_inspect": 1.0, - "train_accuracy_recipient_token_inspect_total": 1, - "train_accuracy_recipient_token_remove": 1.0, - "train_accuracy_recipient_token_remove_total": 1, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_market": 1.0, + "train_accuracy_recipient_token_market_total": 3, + "train_accuracy_recipient_token_verify": 1.0, + "train_accuracy_recipient_token_verify_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.8656126482213439, - "step": 292, - "train_accuracy_first_token_arguments": 0.8666666666666667, - "train_accuracy_first_token_arguments_total": 30, - "train_accuracy_recipient_token": 0.8260869565217391, + "epoch": 0.9150197628458498, + "step": 463, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 0.8888888888888888, + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 20 + }, + { + "epoch": 0.9169960474308301, + "grad_norm": 1.6792924700521263, + "learning_rate": 9.00928482603669e-08, + "loss": 0.1242, + "step": 464 + }, + { + "epoch": 0.9169960474308301, + "step": 464, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_calculate": 1.0, - "train_accuracy_recipient_token_calculate_total": 3, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_multiply": 0.5, - "train_accuracy_recipient_token_multiply_total": 4, - "train_accuracy_recipient_token_suggest": 1.0, - "train_accuracy_recipient_token_suggest_total": 2, - "train_total_number_recipient_token": 23 + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 10, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.8656126482213439, - "step": 292, - "train_accuracy_first_token_arguments": 0.90625, - "train_accuracy_first_token_arguments_total": 32, - "train_accuracy_recipient_token": 0.9444444444444444, + "epoch": 0.9169960474308301, + "step": 464, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 9, - "train_accuracy_recipient_token_book": 1.0, - "train_accuracy_recipient_token_book_total": 2, - "train_accuracy_recipient_token_calculate": 1.0, - "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 5, - "train_accuracy_recipient_token_join": 0.0, - "train_accuracy_recipient_token_join_total": 1, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_research": 1.0, + "train_accuracy_recipient_token_research_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.8685770750988142, - "grad_norm": 2.4189295709433183, - "learning_rate": 2.2139267708310457e-07, - "loss": 0.1385, - "step": 293 + "epoch": 0.9189723320158103, + "grad_norm": 2.1453318249444084, + "learning_rate": 8.587822778610284e-08, + "loss": 0.2816, + "step": 465 }, { - "epoch": 0.8685770750988142, - "step": 293, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 10, - "train_accuracy_recipient_token": 0.9473684210526315, + "epoch": 0.9189723320158103, + "step": 465, + "train_accuracy_recipient_token": 0.9285714285714286, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_extract": 0.5, + "train_accuracy_recipient_token_extract_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 14 + }, + { + "epoch": 0.9189723320158103, + "step": 465, + "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_ex": 1.0, + "train_accuracy_recipient_token_ex_total": 3, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_locate": 1.0, - "train_accuracy_recipient_token_locate_total": 1, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_s": 1.0, + "train_accuracy_recipient_token_s_total": 1, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.9209486166007905, + "grad_norm": 2.6960675173414663, + "learning_rate": 8.176284240242638e-08, + "loss": 0.4503, + "step": 466 + }, + { + "epoch": 0.9209486166007905, + "step": 466, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 0.875, + "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { - "epoch": 0.8685770750988142, - "step": 293, - "train_accuracy_first_token_arguments": 0.9629629629629629, - "train_accuracy_first_token_arguments_total": 27, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.9209486166007905, + "step": 466, + "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 3, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_order": 1.0, - "train_accuracy_recipient_token_order_total": 1, - "train_accuracy_recipient_token_search": 1.0, - "train_accuracy_recipient_token_search_total": 1, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 16 }, { - "epoch": 0.8685770750988142, - "step": 293, - "train_accuracy_first_token_arguments": 0.8571428571428571, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_text": 1.0, - "train_accuracy_recipient_token_text_total": 6, - "train_total_number_recipient_token": 21 + "epoch": 0.9229249011857708, + "grad_norm": 1.7798090264319586, + "learning_rate": 7.774686127671183e-08, + "loss": 0.1686, + "step": 467 }, { - "epoch": 0.8715415019762845, - "grad_norm": 2.1088624243937426, - "learning_rate": 2.1158655935870325e-07, - "loss": 0.2126, - "step": 294 + "epoch": 0.9229249011857708, + "step": 467, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_property": 1.0, + "train_accuracy_recipient_token_property_total": 1, + "train_accuracy_recipient_token_schedule": 0.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.8715415019762845, - "step": 294, - "train_accuracy_first_token_arguments": 0.8947368421052632, - "train_accuracy_first_token_arguments_total": 19, + "epoch": 0.9229249011857708, + "step": 467, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 8, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.8715415019762845, - "step": 294, - "train_accuracy_first_token_arguments": 0.7230769230769231, - "train_accuracy_first_token_arguments_total": 65, - "train_accuracy_recipient_token": 0.95, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.924901185770751, + "grad_norm": 2.545370291066633, + "learning_rate": 7.383044949021339e-08, + "loss": 0.2601, + "step": 468 + }, + { + "epoch": 0.924901185770751, + "step": 468, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_draw": 1.0, - "train_accuracy_recipient_token_draw_total": 5, - "train_accuracy_recipient_token_get": 0.8, - "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_ap": 1.0, + "train_accuracy_recipient_token_ap_total": 1, + "train_accuracy_recipient_token_ger": 1.0, + "train_accuracy_recipient_token_ger_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 1, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.8715415019762845, - "step": 294, - "train_accuracy_first_token_arguments": 0.9444444444444444, - "train_accuracy_first_token_arguments_total": 18, + "epoch": 0.924901185770751, + "step": 468, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 10, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_s": 1.0, + "train_accuracy_recipient_token_s_total": 4, "train_total_number_recipient_token": 20 }, { - "epoch": 0.8745059288537549, - "grad_norm": 2.032905445433328, - "learning_rate": 2.0199295961178893e-07, - "loss": 0.2617, - "step": 295 + "epoch": 0.9268774703557312, + "grad_norm": 1.4300480959159698, + "learning_rate": 7.001376803128041e-08, + "loss": 0.162, + "step": 469 }, { - "epoch": 0.8745059288537549, - "step": 295, - "train_accuracy_first_token_arguments": 0.9166666666666666, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_locate": 1.0, - "train_accuracy_recipient_token_locate_total": 1, - "train_accuracy_recipient_token_suggest": 1.0, - "train_accuracy_recipient_token_suggest_total": 1, - "train_total_number_recipient_token": 18 + "epoch": 0.9268774703557312, + "step": 469, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 0.8888888888888888, + "train_accuracy_recipient_token_get_total": 9, + "train_accuracy_recipient_token_ident": 0.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.8745059288537549, - "step": 295, - "train_accuracy_first_token_arguments": 0.9210526315789473, - "train_accuracy_first_token_arguments_total": 38, - "train_accuracy_recipient_token": 0.9333333333333333, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.9268774703557312, + "step": 469, + "train_accuracy_recipient_token": 0.9375, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_design": 1.0, - "train_accuracy_recipient_token_design_total": 1, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 1, - "train_accuracy_recipient_token_extract": 0.5, - "train_accuracy_recipient_token_extract_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, - "train_total_number_recipient_token": 15 + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 16 }, { - "epoch": 0.8745059288537549, - "step": 295, - "train_accuracy_first_token_arguments": 0.7272727272727273, - "train_accuracy_first_token_arguments_total": 22, + "epoch": 0.9288537549407114, + "grad_norm": 2.7747664296518257, + "learning_rate": 6.629697378873839e-08, + "loss": 0.2523, + "step": 470 + }, + { + "epoch": 0.9288537549407114, + "step": 470, "train_accuracy_recipient_token": 0.8823529411764706, - "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_att": 1.0, - "train_accuracy_recipient_token_att_total": 1, - "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_explain": 1.0, - "train_accuracy_recipient_token_explain_total": 2, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_arr": 0.0, + "train_accuracy_recipient_token_arr_total": 1, + "train_accuracy_recipient_token_mar": 1.0, + "train_accuracy_recipient_token_mar_total": 1, + "train_accuracy_recipient_token_o": 1.0, + "train_accuracy_recipient_token_o_total": 1, + "train_accuracy_recipient_token_under": 1.0, + "train_accuracy_recipient_token_under_total": 1, "train_total_number_recipient_token": 17 }, { - "epoch": 0.8774703557312253, - "grad_norm": 2.284386546160239, - "learning_rate": 1.9261276877071354e-07, - "loss": 0.3743, - "step": 296 + "epoch": 0.9288537549407114, + "step": 470, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 0.5, + "train_accuracy_recipient_token_compare_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.8774703557312253, - "step": 296, - "train_accuracy_first_token_arguments": 0.8, - "train_accuracy_first_token_arguments_total": 15, - "train_accuracy_recipient_token": 0.9, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_get": 0.75, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_learn": 1.0, - "train_accuracy_recipient_token_learn_total": 1, - "train_accuracy_recipient_token_schedule": 1.0, - "train_accuracy_recipient_token_schedule_total": 1, - "train_total_number_recipient_token": 20 + "epoch": 0.9308300395256917, + "grad_norm": 1.8580136680680779, + "learning_rate": 6.268021954544095e-08, + "loss": 0.2213, + "step": 471 }, { - "epoch": 0.8774703557312253, - "step": 296, - "train_accuracy_first_token_arguments": 0.9285714285714286, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 0.9473684210526315, + "epoch": 0.9308300395256917, + "step": 471, + "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 9, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_plan": 0.0, + "train_accuracy_recipient_token_plan_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.8774703557312253, - "step": 296, - "train_accuracy_first_token_arguments": 0.9069767441860465, - "train_accuracy_first_token_arguments_total": 43, - "train_accuracy_recipient_token": 0.8823529411764706, - "train_accuracy_recipient_token_all": 0.8333333333333334, - "train_accuracy_recipient_token_all_total": 12, + "epoch": 0.9308300395256917, + "step": 471, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_stat": 1.0, + "train_accuracy_recipient_token_stat_total": 2, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.8804347826086957, - "grad_norm": 1.6291217601575605, - "learning_rate": 1.8344685794519507e-07, - "loss": 0.1641, - "step": 297 + "epoch": 0.932806324110672, + "grad_norm": 2.204067976920768, + "learning_rate": 5.9163653971989754e-08, + "loss": 0.3436, + "step": 472 }, { - "epoch": 0.8804347826086957, - "step": 297, - "train_accuracy_first_token_arguments": 0.9649122807017544, - "train_accuracy_first_token_arguments_total": 57, - "train_accuracy_recipient_token": 0.8947368421052632, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.932806324110672, + "step": 472, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_calculate": 1.0, - "train_accuracy_recipient_token_calculate_total": 3, - "train_accuracy_recipient_token_check": 1.0, - "train_accuracy_recipient_token_check_total": 1, - "train_accuracy_recipient_token_get": 0.5, - "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_search": 1.0, - "train_accuracy_recipient_token_search_total": 2, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, + "train_accuracy_recipient_token_arch": 1.0, + "train_accuracy_recipient_token_arch_total": 2, + "train_accuracy_recipient_token_building": 1.0, + "train_accuracy_recipient_token_building_total": 3, + "train_accuracy_recipient_token_gr": 1.0, + "train_accuracy_recipient_token_gr_total": 3, "train_total_number_recipient_token": 19 }, { - "epoch": 0.8804347826086957, - "step": 297, - "train_accuracy_first_token_arguments": 0.825, - "train_accuracy_first_token_arguments_total": 40, - "train_accuracy_recipient_token": 0.95, + "epoch": 0.932806324110672, + "step": 472, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 8, - "train_accuracy_recipient_token_sort": 0.0, - "train_accuracy_recipient_token_sort_total": 1, - "train_total_number_recipient_token": 20 - }, - { - "epoch": 0.8804347826086957, - "step": 297, - "train_accuracy_first_token_arguments": 0.8648648648648649, - "train_accuracy_first_token_arguments_total": 37, - "train_accuracy_recipient_token": 0.85, - "train_accuracy_recipient_token_all": 0.9230769230769231, - "train_accuracy_recipient_token_all_total": 13, - "train_accuracy_recipient_token_get": 0.6666666666666666, - "train_accuracy_recipient_token_get_total": 6, - "train_accuracy_recipient_token_restaurant": 1.0, - "train_accuracy_recipient_token_restaurant_total": 1, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_inter": 1.0, + "train_accuracy_recipient_token_inter_total": 1, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.883399209486166, - "grad_norm": 1.7794723846623823, - "learning_rate": 1.744960783454186e-07, - "loss": 0.2034, - "step": 298 + "epoch": 0.9347826086956522, + "grad_norm": 1.3085005755431722, + "learning_rate": 5.574742162062163e-08, + "loss": 0.193, + "step": 473 }, { - "epoch": 0.883399209486166, - "step": 298, - "train_accuracy_first_token_arguments": 0.8421052631578947, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.95, + "epoch": 0.9347826086956522, + "step": 473, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, - "train_accuracy_recipient_token_identify": 0.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_total_number_recipient_token": 20 - }, - { - "epoch": 0.883399209486166, - "step": 298, - "train_accuracy_first_token_arguments": 0.7872340425531915, - "train_accuracy_first_token_arguments_total": 47, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.8333333333333334, - "train_accuracy_recipient_token_all_total": 12, - "train_accuracy_recipient_token_building": 1.0, - "train_accuracy_recipient_token_building_total": 2, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_stat": 1.0, - "train_accuracy_recipient_token_stat_total": 2, - "train_total_number_recipient_token": 18 + "train_total_number_recipient_token": 19 }, { - "epoch": 0.883399209486166, - "step": 298, - "train_accuracy_first_token_arguments": 0.8837209302325582, - "train_accuracy_first_token_arguments_total": 43, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.9347826086956522, + "step": 473, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compile": 1.0, - "train_accuracy_recipient_token_compile_total": 1, - "train_accuracy_recipient_token_get": 0.75, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 1, - "train_accuracy_recipient_token_restaurant": 1.0, - "train_accuracy_recipient_token_restaurant_total": 1, + "train_accuracy_recipient_token_f": 1.0, + "train_accuracy_recipient_token_f_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_send": 1.0, + "train_accuracy_recipient_token_send_total": 1, + "train_accuracy_recipient_token_set": 1.0, + "train_accuracy_recipient_token_set_total": 2, "train_total_number_recipient_token": 18 }, { - "epoch": 0.8863636363636364, - "grad_norm": 1.9655423422483125, - "learning_rate": 1.6576126120299046e-07, - "loss": 0.3209, - "step": 299 + "epoch": 0.9367588932806324, + "grad_norm": 2.537954422029753, + "learning_rate": 5.2431662919267825e-08, + "loss": 0.2243, + "step": 474 }, { - "epoch": 0.8863636363636364, - "step": 299, - "train_accuracy_first_token_arguments": 0.8333333333333334, - "train_accuracy_first_token_arguments_total": 30, + "epoch": 0.9367588932806324, + "step": 474, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_draw": 1.0, - "train_accuracy_recipient_token_draw_total": 5, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.8863636363636364, - "step": 299, - "train_accuracy_first_token_arguments": 0.7777777777777778, - "train_accuracy_first_token_arguments_total": 9, - "train_accuracy_recipient_token": 0.9444444444444444, + "epoch": 0.9367588932806324, + "step": 474, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_exp": 1.0, - "train_accuracy_recipient_token_exp_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_translate": 0.0, - "train_accuracy_recipient_token_translate_total": 1, - "train_total_number_recipient_token": 18 - }, - { - "epoch": 0.8863636363636364, - "step": 299, - "train_accuracy_first_token_arguments": 0.9473684210526315, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.8947368421052632, - "train_accuracy_recipient_token_all": 0.8181818181818182, - "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_format": 1.0, + "train_accuracy_recipient_token_format_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_plant": 1.0, - "train_accuracy_recipient_token_plant_total": 3, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_sum": 1.0, + "train_accuracy_recipient_token_sum_total": 1, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.8893280632411067, - "grad_norm": 2.3395853773637407, - "learning_rate": 1.5724321769374023e-07, - "loss": 0.3217, - "step": 300 + "epoch": 0.9387351778656127, + "grad_norm": 1.805153682895635, + "learning_rate": 4.9216514165781885e-08, + "loss": 0.145, + "step": 475 }, { - "epoch": 0.8893280632411067, - "step": 300, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 9, - "train_accuracy_recipient_token": 0.8235294117647058, - "train_accuracy_recipient_token_all": 0.7, + "epoch": 0.9387351778656127, + "step": 475, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_arr": 1.0, + "train_accuracy_recipient_token_arr_total": 1, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_search": 1.0, - "train_accuracy_recipient_token_search_total": 1, - "train_total_number_recipient_token": 17 - }, - { - "epoch": 0.8893280632411067, - "step": 300, - "train_accuracy_first_token_arguments": 0.9, - "train_accuracy_first_token_arguments_total": 30, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_bird": 1.0, - "train_accuracy_recipient_token_bird_total": 2, + "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.8893280632411067, - "step": 300, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 10, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, + "epoch": 0.9387351778656127, + "step": 475, + "train_accuracy_recipient_token": 0.8235294117647058, + "train_accuracy_recipient_token_all": 0.8333333333333334, + "train_accuracy_recipient_token_all_total": 12, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_compare": 0.0, + "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 7, - "train_total_number_recipient_token": 19 - }, - { - "epoch": 0.892292490118577, - "grad_norm": 2.265776160882912, - "learning_rate": 1.4894273886239208e-07, - "loss": 0.3318, - "step": 301 + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.892292490118577, - "step": 301, - "train_accuracy_first_token_arguments": 0.9375, - "train_accuracy_first_token_arguments_total": 16, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_c": 1.0, - "train_accuracy_recipient_token_c_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 7, - "train_total_number_recipient_token": 18 + "epoch": 0.9407114624505929, + "grad_norm": 2.045602997144322, + "learning_rate": 4.61021075223364e-08, + "loss": 0.2222, + "step": 476 }, { - "epoch": 0.892292490118577, - "step": 301, - "train_accuracy_first_token_arguments": 0.8888888888888888, - "train_accuracy_first_token_arguments_total": 9, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.8333333333333334, - "train_accuracy_recipient_token_all_total": 12, + "epoch": 0.9407114624505929, + "step": 476, + "train_accuracy_recipient_token": 0.7894736842105263, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_ex": 1.0, + "train_accuracy_recipient_token_ex_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 6, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_plane": 1.0, + "train_accuracy_recipient_token_plane_total": 3, + "train_accuracy_recipient_token_zi": 0.0, + "train_accuracy_recipient_token_zi_total": 2, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.892292490118577, - "step": 301, - "train_accuracy_first_token_arguments": 0.9, - "train_accuracy_first_token_arguments_total": 20, - "train_accuracy_recipient_token": 0.9473684210526315, + "epoch": 0.9407114624505929, + "step": 476, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_arr": 1.0, + "train_accuracy_recipient_token_arr_total": 2, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 0.8333333333333334, - "train_accuracy_recipient_token_get_total": 6, - "train_accuracy_recipient_token_list": 1.0, - "train_accuracy_recipient_token_list_total": 1, - "train_accuracy_recipient_token_stat": 1.0, - "train_accuracy_recipient_token_stat_total": 1, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.8952569169960475, - "grad_norm": 2.3746950651302927, - "learning_rate": 1.4086059554910186e-07, - "loss": 0.3669, - "step": 302 + "epoch": 0.9426877470355731, + "grad_norm": 1.3694314670334982, + "learning_rate": 4.308857100999042e-08, + "loss": 0.1326, + "step": 477 }, { - "epoch": 0.8952569169960475, - "step": 302, - "train_accuracy_first_token_arguments": 0.6590909090909091, - "train_accuracy_first_token_arguments_total": 44, - "train_accuracy_recipient_token": 0.95, + "epoch": 0.9426877470355731, + "step": 477, + "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 1, - "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_object": 1.0, - "train_accuracy_recipient_token_object_total": 1, - "train_accuracy_recipient_token_people": 1.0, - "train_accuracy_recipient_token_people_total": 2, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_ident": 0.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.8952569169960475, - "step": 302, - "train_accuracy_first_token_arguments": 0.9285714285714286, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_verify": 1.0, - "train_accuracy_recipient_token_verify_total": 1, + "epoch": 0.9426877470355731, + "step": 477, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 0.8571428571428571, + "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { - "epoch": 0.8952569169960475, - "step": 302, - "train_accuracy_first_token_arguments": 0.9090909090909091, - "train_accuracy_first_token_arguments_total": 11, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_building": 1.0, - "train_accuracy_recipient_token_building_total": 4, + "epoch": 0.9446640316205533, + "grad_norm": 3.126805689186157, + "learning_rate": 4.017602850342584e-08, + "loss": 0.3008, + "step": 478 + }, + { + "epoch": 0.9446640316205533, + "step": 478, + "train_accuracy_recipient_token": 0.85, + "train_accuracy_recipient_token_all": 0.75, + "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 1, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 4, "train_total_number_recipient_token": 20 }, { - "epoch": 0.8982213438735178, - "grad_norm": 2.5356426721293994, - "learning_rate": 1.3299753831787193e-07, - "loss": 0.2913, - "step": 303 - }, - { - "epoch": 0.8982213438735178, - "step": 303, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 16, + "epoch": 0.9446640316205533, + "step": 478, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_book": 1.0, - "train_accuracy_recipient_token_book_total": 1, - "train_accuracy_recipient_token_c": 1.0, - "train_accuracy_recipient_token_c_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { - "epoch": 0.8982213438735178, - "step": 303, - "train_accuracy_first_token_arguments": 0.8947368421052632, - "train_accuracy_first_token_arguments_total": 19, + "epoch": 0.9466403162055336, + "grad_norm": 2.434210637252682, + "learning_rate": 3.736459972585815e-08, + "loss": 0.2991, + "step": 479 + }, + { + "epoch": 0.9466403162055336, + "step": 479, "train_accuracy_recipient_token": 0.8947368421052632, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_get": 0.875, - "train_accuracy_recipient_token_get_total": 8, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.75, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_search": 1.0, + "train_accuracy_recipient_token_search_total": 2, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.8982213438735178, - "step": 303, - "train_accuracy_first_token_arguments": 0.9230769230769231, - "train_accuracy_first_token_arguments_total": 13, + "epoch": 0.9466403162055336, + "step": 479, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_check": 1.0, - "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 6, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 2, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_plant": 1.0, + "train_accuracy_recipient_token_plant_total": 3, "train_total_number_recipient_token": 19 }, { - "epoch": 0.9011857707509882, - "grad_norm": 2.3231357183648993, - "learning_rate": 1.2535429738684822e-07, - "loss": 0.2729, - "step": 304 + "epoch": 0.9486166007905138, + "grad_norm": 1.5988207864411084, + "learning_rate": 3.4654400244112654e-08, + "loss": 0.1611, + "step": 480 }, { - "epoch": 0.9011857707509882, - "step": 304, - "train_accuracy_first_token_arguments": 0.8571428571428571, - "train_accuracy_first_token_arguments_total": 21, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_check": 1.0, - "train_accuracy_recipient_token_check_total": 3, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, + "epoch": 0.9486166007905138, + "step": 480, + "train_accuracy_recipient_token": 0.85, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 15, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_get_total": 5, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.9011857707509882, - "step": 304, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 6, + "epoch": 0.9486166007905138, + "step": 480, "train_accuracy_recipient_token": 0.8, - "train_accuracy_recipient_token_all": 0.8, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_recogn": 1.0, - "train_accuracy_recipient_token_recogn_total": 1, - "train_accuracy_recipient_token_translate": 1.0, - "train_accuracy_recipient_token_translate_total": 1, - "train_total_number_recipient_token": 15 - }, - { - "epoch": 0.9011857707509882, - "step": 304, - "train_accuracy_first_token_arguments": 0.9444444444444444, - "train_accuracy_first_token_arguments_total": 18, - "train_accuracy_recipient_token": 0.9411764705882353, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_predict": 0.0, + "train_accuracy_recipient_token_predict_total": 1, + "train_accuracy_recipient_token_stamp": 1.0, + "train_accuracy_recipient_token_stamp_total": 2, + "train_total_number_recipient_token": 15 }, { - "epoch": 0.9041501976284585, - "grad_norm": 2.0889543788782716, - "learning_rate": 1.1793158256050708e-07, - "loss": 0.3167, - "step": 305 + "epoch": 0.950592885375494, + "grad_norm": 1.9240578271638804, + "learning_rate": 3.204554146387456e-08, + "loss": 0.3299, + "step": 481 }, { - "epoch": 0.9041501976284585, - "step": 305, - "train_accuracy_first_token_arguments": 0.9285714285714286, - "train_accuracy_first_token_arguments_total": 14, + "epoch": 0.950592885375494, + "step": 481, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 10, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_get_total": 9, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.9041501976284585, - "step": 305, - "train_accuracy_first_token_arguments": 0.8059701492537313, - "train_accuracy_first_token_arguments_total": 67, - "train_accuracy_recipient_token": 0.9375, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_event": 1.0, - "train_accuracy_recipient_token_event_total": 2, + "epoch": 0.950592885375494, + "step": 481, + "train_accuracy_recipient_token": 0.8333333333333334, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_total_number_recipient_token": 16 + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_opt": 0.0, + "train_accuracy_recipient_token_opt_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.9041501976284585, - "step": 305, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.9525691699604744, + "grad_norm": 2.5790139743271983, + "learning_rate": 2.9538130625110796e-08, + "loss": 0.2576, + "step": 482 + }, + { + "epoch": 0.9525691699604744, + "step": 482, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_car": 1.0, - "train_accuracy_recipient_token_car_total": 1, - "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { - "epoch": 0.9071146245059288, - "grad_norm": 1.546141851099797, - "learning_rate": 1.1073008316373812e-07, - "loss": 0.1956, - "step": 306 - }, - { - "epoch": 0.9071146245059288, - "step": 306, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 30, + "epoch": 0.9525691699604744, + "step": 482, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 2, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, - "train_accuracy_recipient_token_stat": 1.0, - "train_accuracy_recipient_token_stat_total": 3, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 16 }, { - "epoch": 0.9071146245059288, - "step": 306, - "train_accuracy_first_token_arguments": 0.9166666666666666, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.9545454545454546, + "grad_norm": 2.8948073860814616, + "learning_rate": 2.7132270797659567e-08, + "loss": 0.3431, + "step": 483 + }, + { + "epoch": 0.9545454545454546, + "step": 483, + "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_create": 1.0, + "train_accuracy_recipient_token_create_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 8, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_ident": 0.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_ne": 1.0, + "train_accuracy_recipient_token_ne_total": 3, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.9071146245059288, - "step": 306, - "train_accuracy_first_token_arguments": 0.7857142857142857, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 0.9523809523809523, + "epoch": 0.9545454545454546, + "step": 483, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_convert": 1.0, + "train_accuracy_recipient_token_convert_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 2, - "train_accuracy_recipient_token_search": 0.0, - "train_accuracy_recipient_token_search_total": 1, - "train_total_number_recipient_token": 21 + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 15 }, { - "epoch": 0.9100790513833992, - "grad_norm": 2.2381733984414756, - "learning_rate": 1.0375046797782868e-07, - "loss": 0.2507, - "step": 307 + "epoch": 0.9565217391304348, + "grad_norm": 3.1119629331028023, + "learning_rate": 2.4828060876995462e-08, + "loss": 0.2377, + "step": 484 }, { - "epoch": 0.9100790513833992, - "step": 307, - "train_accuracy_first_token_arguments": 0.8947368421052632, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.9444444444444444, + "epoch": 0.9565217391304348, + "step": 484, + "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_calculate": 1.0, - "train_accuracy_recipient_token_calculate_total": 2, - "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_count": 1.0, - "train_accuracy_recipient_token_count_total": 1, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 2, - "train_total_number_recipient_token": 18 - }, - { - "epoch": 0.9100790513833992, - "step": 307, - "train_accuracy_first_token_arguments": 0.8, - "train_accuracy_first_token_arguments_total": 15, - "train_accuracy_recipient_token": 0.8947368421052632, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_custom": 1.0, + "train_accuracy_recipient_token_custom_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_gr": 1.0, - "train_accuracy_recipient_token_gr_total": 3, - "train_accuracy_recipient_token_identify": 0.5, - "train_accuracy_recipient_token_identify_total": 2, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_total_number_recipient_token": 16 }, { - "epoch": 0.9100790513833992, - "step": 307, - "train_accuracy_first_token_arguments": 0.9565217391304348, - "train_accuracy_first_token_arguments_total": 23, - "train_accuracy_recipient_token": 0.9523809523809523, + "epoch": 0.9565217391304348, + "step": 484, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 0.9090909090909091, - "train_accuracy_recipient_token_get_total": 11, - "train_total_number_recipient_token": 21 + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_pool": 1.0, + "train_accuracy_recipient_token_pool_total": 4, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.9130434782608695, - "grad_norm": 1.6564222632639842, - "learning_rate": 9.699338517835611e-08, - "loss": 0.2674, - "step": 308 + "epoch": 0.958498023715415, + "grad_norm": 1.7117541555790323, + "learning_rate": 2.262559558016325e-08, + "loss": 0.2084, + "step": 485 }, { - "epoch": 0.9130434782608695, - "step": 308, - "train_accuracy_first_token_arguments": 0.9166666666666666, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.9411764705882353, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.958498023715415, + "step": 485, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_verify": 1.0, - "train_accuracy_recipient_token_verify_total": 1, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_robot": 1.0, + "train_accuracy_recipient_token_robot_total": 2, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.9130434782608695, - "step": 308, - "train_accuracy_first_token_arguments": 0.8571428571428571, - "train_accuracy_first_token_arguments_total": 28, - "train_accuracy_recipient_token": 0.95, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_arr": 0.0, - "train_accuracy_recipient_token_arr_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_house": 1.0, - "train_accuracy_recipient_token_house_total": 3, - "train_accuracy_recipient_token_landscape": 1.0, - "train_accuracy_recipient_token_landscape_total": 1, - "train_total_number_recipient_token": 20 + "epoch": 0.958498023715415, + "step": 485, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 0.875, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.9130434782608695, - "step": 308, - "train_accuracy_first_token_arguments": 0.8125, - "train_accuracy_first_token_arguments_total": 16, - "train_accuracy_recipient_token": 0.9444444444444444, + "epoch": 0.9604743083003953, + "grad_norm": 1.620668314091392, + "learning_rate": 2.052496544188487e-08, + "loss": 0.1256, + "step": 486 + }, + { + "epoch": 0.9604743083003953, + "step": 486, + "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_market": 1.0, - "train_accuracy_recipient_token_market_total": 3, - "train_accuracy_recipient_token_verify": 1.0, - "train_accuracy_recipient_token_verify_total": 1, - "train_total_number_recipient_token": 18 - }, - { - "epoch": 0.91600790513834, - "grad_norm": 1.503775394527991, - "learning_rate": 9.045946227499298e-08, - "loss": 0.1276, - "step": 309 + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_learn": 1.0, + "train_accuracy_recipient_token_learn_total": 1, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.91600790513834, - "step": 309, - "train_accuracy_first_token_arguments": 0.8947368421052632, - "train_accuracy_first_token_arguments_total": 19, + "epoch": 0.9604743083003953, + "step": 486, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, @@ -17142,1418 +18509,1256 @@ "train_total_number_recipient_token": 20 }, { - "epoch": 0.91600790513834, - "step": 309, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 15, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.9624505928853755, + "grad_norm": 1.5982169963407395, + "learning_rate": 1.852625681083742e-08, + "loss": 0.112, + "step": 487 + }, + { + "epoch": 0.9624505928853755, + "step": 487, + "train_accuracy_recipient_token": 0.9523809523809523, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 10, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_count": 1.0, + "train_accuracy_recipient_token_count_total": 1, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_subtract": 1.0, + "train_accuracy_recipient_token_subtract_total": 1, + "train_accuracy_recipient_token_track": 1.0, + "train_accuracy_recipient_token_track_total": 1, + "train_total_number_recipient_token": 21 }, { - "epoch": 0.91600790513834, - "step": 309, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 11, + "epoch": 0.9624505928853755, + "step": 487, "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 0.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_compile": 1.0, + "train_accuracy_recipient_token_compile_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_research": 1.0, - "train_accuracy_recipient_token_research_total": 1, + "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 18 }, { - "epoch": 0.9189723320158103, - "grad_norm": 2.2000639611250423, - "learning_rate": 8.414930605323445e-08, - "loss": 0.2596, - "step": 310 + "epoch": 0.9644268774703557, + "grad_norm": 1.8281750558489605, + "learning_rate": 1.6629551846104874e-08, + "loss": 0.1926, + "step": 488 }, { - "epoch": 0.9189723320158103, - "step": 310, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 4, - "train_accuracy_recipient_token": 0.9285714285714286, + "epoch": 0.9644268774703557, + "step": 488, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_extract": 0.5, - "train_accuracy_recipient_token_extract_total": 2, + "train_accuracy_recipient_token_extract": 1.0, + "train_accuracy_recipient_token_extract_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_recogn": 1.0, + "train_accuracy_recipient_token_recogn_total": 1, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 1, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.9644268774703557, + "step": 488, + "train_accuracy_recipient_token": 0.9333333333333333, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_total_number_recipient_token": 14 + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_verify": 1.0, + "train_accuracy_recipient_token_verify_total": 1, + "train_total_number_recipient_token": 15 }, { - "epoch": 0.9189723320158103, - "step": 310, - "train_accuracy_first_token_arguments": 0.7272727272727273, - "train_accuracy_first_token_arguments_total": 11, - "train_accuracy_recipient_token": 0.8125, + "epoch": 0.9664031620553359, + "grad_norm": 2.6298463108507355, + "learning_rate": 1.483492851379914e-08, + "loss": 0.4112, + "step": 489 + }, + { + "epoch": 0.9664031620553359, + "step": 489, + "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_explain": 0.6666666666666666, - "train_accuracy_recipient_token_explain_total": 3, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 3, + "train_accuracy_recipient_token_check": 0.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, - "train_accuracy_recipient_token_suggest": 1.0, - "train_accuracy_recipient_token_suggest_total": 1, - "train_total_number_recipient_token": 16 + "train_total_number_recipient_token": 18 }, { - "epoch": 0.9189723320158103, - "step": 310, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 23, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, + "epoch": 0.9664031620553359, + "step": 489, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compile": 1.0, + "train_accuracy_recipient_token_compile_total": 1, + "train_accuracy_recipient_token_find": 0.75, + "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 8, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.9219367588932806, - "grad_norm": 2.4294485936947194, - "learning_rate": 7.806350251804484e-08, - "loss": 0.3957, - "step": 311 + "epoch": 0.9683794466403162, + "grad_norm": 2.433575683997582, + "learning_rate": 1.3142460583856487e-08, + "loss": 0.2541, + "step": 490 }, { - "epoch": 0.9219367588932806, - "step": 311, - "train_accuracy_first_token_arguments": 0.7619047619047619, - "train_accuracy_first_token_arguments_total": 42, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.9683794466403162, + "step": 490, + "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 16 - }, - { - "epoch": 0.9219367588932806, - "step": 311, - "train_accuracy_first_token_arguments": 0.8235294117647058, - "train_accuracy_first_token_arguments_total": 17, - "train_accuracy_recipient_token": 0.8823529411764706, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 1, - "train_accuracy_recipient_token_property": 1.0, - "train_accuracy_recipient_token_property_total": 1, - "train_accuracy_recipient_token_schedule": 0.0, - "train_accuracy_recipient_token_schedule_total": 1, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_download": 0.0, + "train_accuracy_recipient_token_download_total": 1, + "train_accuracy_recipient_token_get": 0.8333333333333334, + "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_search": 0.0, + "train_accuracy_recipient_token_search_total": 1, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.9219367588932806, - "step": 311, - "train_accuracy_first_token_arguments": 0.9444444444444444, - "train_accuracy_first_token_arguments_total": 18, + "epoch": 0.9683794466403162, + "step": 490, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 6, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.924901185770751, - "grad_norm": 2.1844882128572, - "learning_rate": 7.220261683943935e-08, - "loss": 0.26, - "step": 312 + "epoch": 0.9703557312252964, + "grad_norm": 1.7477968227475646, + "learning_rate": 1.1552217627004426e-08, + "loss": 0.1936, + "step": 491 }, { - "epoch": 0.924901185770751, - "step": 312, - "train_accuracy_first_token_arguments": 0.6666666666666666, - "train_accuracy_first_token_arguments_total": 15, - "train_accuracy_recipient_token": 0.9473684210526315, + "epoch": 0.9703557312252964, + "step": 491, + "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_ap": 1.0, - "train_accuracy_recipient_token_ap_total": 1, - "train_accuracy_recipient_token_ger": 1.0, - "train_accuracy_recipient_token_ger_total": 1, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, + "train_accuracy_recipient_token_deb": 0.0, + "train_accuracy_recipient_token_deb_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_validate": 1.0, - "train_accuracy_recipient_token_validate_total": 1, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_plan": 1.0, + "train_accuracy_recipient_token_plan_total": 1, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.924901185770751, - "step": 312, - "train_accuracy_first_token_arguments": 0.875, - "train_accuracy_first_token_arguments_total": 16, - "train_accuracy_recipient_token": 0.95, + "epoch": 0.9703557312252964, + "step": 491, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 6, - "train_accuracy_recipient_token_sand": 0.75, - "train_accuracy_recipient_token_sand_total": 4, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.924901185770751, - "step": 312, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 13, - "train_accuracy_recipient_token": 0.9, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 0.8888888888888888, - "train_accuracy_recipient_token_get_total": 9, - "train_accuracy_recipient_token_identify": 0.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_total_number_recipient_token": 20 + "epoch": 0.9723320158102767, + "grad_norm": 1.8404798729311513, + "learning_rate": 1.006426501190233e-08, + "loss": 0.2403, + "step": 492 }, { - "epoch": 0.9278656126482213, - "grad_norm": 1.842757768341881, - "learning_rate": 6.6567193299997e-08, - "loss": 0.1724, - "step": 313 + "epoch": 0.9723320158102767, + "step": 492, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.9278656126482213, - "step": 313, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 8, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.9723320158102767, + "step": 492, + "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 6, - "train_total_number_recipient_token": 16 + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.9278656126482213, - "step": 313, - "train_accuracy_first_token_arguments": 0.5833333333333334, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.8823529411764706, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_arr": 0.0, - "train_accuracy_recipient_token_arr_total": 1, - "train_accuracy_recipient_token_marine": 1.0, - "train_accuracy_recipient_token_marine_total": 1, - "train_accuracy_recipient_token_o": 1.0, - "train_accuracy_recipient_token_o_total": 1, - "train_accuracy_recipient_token_under": 1.0, - "train_accuracy_recipient_token_under_total": 1, - "train_total_number_recipient_token": 17 + "epoch": 0.974308300395257, + "grad_norm": 2.0079365969770246, + "learning_rate": 8.67866390245442e-09, + "loss": 0.216, + "step": 493 }, { - "epoch": 0.9278656126482213, - "step": 313, - "train_accuracy_first_token_arguments": 0.8571428571428571, - "train_accuracy_first_token_arguments_total": 35, - "train_accuracy_recipient_token": 0.8333333333333334, - "train_accuracy_recipient_token_all": 0.8, + "epoch": 0.974308300395257, + "step": 493, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compare": 0.5, - "train_accuracy_recipient_token_compare_total": 2, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, - "train_accuracy_recipient_token_search": 1.0, - "train_accuracy_recipient_token_search_total": 1, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_park": 1.0, + "train_accuracy_recipient_token_park_total": 4, "train_total_number_recipient_token": 18 }, { - "epoch": 0.9308300395256917, - "grad_norm": 2.101888459848309, - "learning_rate": 6.115775524431711e-08, - "loss": 0.3051, - "step": 314 - }, - { - "epoch": 0.9308300395256917, - "step": 314, - "train_accuracy_first_token_arguments": 0.9090909090909091, - "train_accuracy_first_token_arguments_total": 22, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.974308300395257, + "step": 493, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_ex": 1.0, + "train_accuracy_recipient_token_ex_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_plan": 1.0, - "train_accuracy_recipient_token_plan_total": 1, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.9308300395256917, - "step": 314, - "train_accuracy_first_token_arguments": 0.7857142857142857, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 0.8421052631578947, - "train_accuracy_recipient_token_all": 0.8181818181818182, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, + "epoch": 0.9762845849802372, + "grad_norm": 2.1557879594347273, + "learning_rate": 7.3954712552953835e-09, + "loss": 0.234, + "step": 494 + }, + { + "epoch": 0.9762845849802372, + "step": 494, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_stat": 0.5, - "train_accuracy_recipient_token_stat_total": 2, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_get_total": 7, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.9308300395256917, - "step": 314, - "train_accuracy_first_token_arguments": 0.9411764705882353, - "train_accuracy_first_token_arguments_total": 17, + "epoch": 0.9762845849802372, + "step": 494, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 1, - "train_accuracy_recipient_token_arch": 1.0, - "train_accuracy_recipient_token_arch_total": 2, - "train_accuracy_recipient_token_building": 1.0, - "train_accuracy_recipient_token_building_total": 3, - "train_accuracy_recipient_token_gr": 1.0, - "train_accuracy_recipient_token_gr_total": 3, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_art": 1.0, + "train_accuracy_recipient_token_art_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 1, + "train_total_number_recipient_token": 16 }, { - "epoch": 0.9337944664031621, - "grad_norm": 1.8025117128414663, - "learning_rate": 5.597480503041486e-08, - "loss": 0.3276, - "step": 315 + "epoch": 0.9782608695652174, + "grad_norm": 2.088360252773201, + "learning_rate": 6.214739817448634e-09, + "loss": 0.2639, + "step": 495 }, { - "epoch": 0.9337944664031621, - "step": 315, - "train_accuracy_first_token_arguments": 0.7647058823529411, - "train_accuracy_first_token_arguments_total": 102, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.9782608695652174, + "step": 495, + "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 1, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_inter": 1.0, - "train_accuracy_recipient_token_inter_total": 1, + "train_accuracy_recipient_token_get": 0.8888888888888888, + "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { - "epoch": 0.9337944664031621, - "step": 315, - "train_accuracy_first_token_arguments": 0.8461538461538461, - "train_accuracy_first_token_arguments_total": 13, + "epoch": 0.9782608695652174, + "step": 495, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_filter": 1.0, + "train_accuracy_recipient_token_filter_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 9, + "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { - "epoch": 0.9337944664031621, - "step": 315, - "train_accuracy_first_token_arguments": 0.9473684210526315, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.9444444444444444, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_fabric": 1.0, - "train_accuracy_recipient_token_fabric_total": 4, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, - "train_accuracy_recipient_token_send": 1.0, - "train_accuracy_recipient_token_send_total": 1, - "train_accuracy_recipient_token_set": 1.0, - "train_accuracy_recipient_token_set_total": 2, - "train_total_number_recipient_token": 18 + "epoch": 0.9802371541501976, + "grad_norm": 1.6208106966240963, + "learning_rate": 5.136518124159162e-09, + "loss": 0.1421, + "step": 496 }, { - "epoch": 0.9367588932806324, - "grad_norm": 2.303122934120546, - "learning_rate": 5.1018823983070285e-08, - "loss": 0.2539, - "step": 316 + "epoch": 0.9802371541501976, + "step": 496, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_motor": 1.0, + "train_accuracy_recipient_token_motor_total": 3, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.9367588932806324, - "step": 316, - "train_accuracy_first_token_arguments": 0.7, - "train_accuracy_first_token_arguments_total": 20, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 11, + "epoch": 0.9802371541501976, + "step": 496, + "train_accuracy_recipient_token": 0.6842105263157895, + "train_accuracy_recipient_token_all": 0.6666666666666666, + "train_accuracy_recipient_token_all_total": 15, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_get": 0.6666666666666666, + "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 19 }, { - "epoch": 0.9367588932806324, - "step": 316, - "train_accuracy_first_token_arguments": 0.9166666666666666, - "train_accuracy_first_token_arguments_total": 12, + "epoch": 0.9822134387351779, + "grad_norm": 2.040838086529817, + "learning_rate": 4.160850496897906e-09, + "loss": 0.164, + "step": 497 + }, + { + "epoch": 0.9822134387351779, + "step": 497, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_building": 1.0, + "train_accuracy_recipient_token_building_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_format": 1.0, - "train_accuracy_recipient_token_format_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_sum": 1.0, - "train_accuracy_recipient_token_sum_total": 1, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.9367588932806324, - "step": 316, - "train_accuracy_first_token_arguments": 0.7, - "train_accuracy_first_token_arguments_total": 10, - "train_accuracy_recipient_token": 0.7777777777777778, + "epoch": 0.9822134387351779, + "step": 497, + "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_arr": 0.0, - "train_accuracy_recipient_token_arr_total": 1, - "train_accuracy_recipient_token_check": 1.0, - "train_accuracy_recipient_token_check_total": 1, - "train_accuracy_recipient_token_find": 0.5, + "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_get": 0.75, - "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 2, + "train_accuracy_recipient_token_table": 1.0, + "train_accuracy_recipient_token_table_total": 2, "train_total_number_recipient_token": 18 }, { - "epoch": 0.9397233201581028, - "grad_norm": 1.7793933316663082, - "learning_rate": 4.629027234912986e-08, - "loss": 0.1763, - "step": 317 - }, - { - "epoch": 0.9397233201581028, - "step": 317, - "train_accuracy_first_token_arguments": 0.95, - "train_accuracy_first_token_arguments_total": 20, - "train_accuracy_recipient_token": 0.8823529411764706, - "train_accuracy_recipient_token_all": 0.9166666666666666, - "train_accuracy_recipient_token_all_total": 12, - "train_accuracy_recipient_token_book": 1.0, - "train_accuracy_recipient_token_book_total": 1, - "train_accuracy_recipient_token_compare": 0.0, - "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_total_number_recipient_token": 17 + "epoch": 0.9841897233201581, + "grad_norm": 1.8868735474227656, + "learning_rate": 3.287777041539042e-09, + "loss": 0.1909, + "step": 498 }, { - "epoch": 0.9397233201581028, - "step": 317, - "train_accuracy_first_token_arguments": 0.7428571428571429, - "train_accuracy_first_token_arguments_total": 35, + "epoch": 0.9841897233201581, + "step": 498, "train_accuracy_recipient_token": 0.8421052631578947, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_explain": 1.0, - "train_accuracy_recipient_token_explain_total": 1, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_generate": 0.0, + "train_accuracy_recipient_token_generate_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, - "train_accuracy_recipient_token_plane": 1.0, - "train_accuracy_recipient_token_plane_total": 3, - "train_accuracy_recipient_token_zi": 0.0, - "train_accuracy_recipient_token_zi_total": 2, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_sum": 1.0, + "train_accuracy_recipient_token_sum_total": 1, + "train_accuracy_recipient_token_translate": 0.5, + "train_accuracy_recipient_token_translate_total": 2, "train_total_number_recipient_token": 19 }, { - "epoch": 0.9397233201581028, - "step": 317, - "train_accuracy_first_token_arguments": 0.8888888888888888, - "train_accuracy_first_token_arguments_total": 18, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.9841897233201581, + "step": 498, + "train_accuracy_recipient_token": 0.9285714285714286, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_arr": 1.0, - "train_accuracy_recipient_token_arr_total": 2, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_evaluate": 0.0, + "train_accuracy_recipient_token_evaluate_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_total_number_recipient_token": 17 - }, - { - "epoch": 0.9426877470355731, - "grad_norm": 1.5774943505040582, - "learning_rate": 4.178958925476401e-08, - "loss": 0.1834, - "step": 318 - }, - { - "epoch": 0.9426877470355731, - "eval_accuracy_first_token_arguments": 0.9248395967002749, - "eval_accuracy_first_token_arguments_total": 1091, - "eval_accuracy_recipient_token": 0.9401330376940134, - "eval_accuracy_recipient_token_all": 0.9220272904483431, - "eval_accuracy_recipient_token_all_total": 513, - "eval_accuracy_recipient_token_calculate": 1.0, - "eval_accuracy_recipient_token_calculate_total": 13, - "eval_accuracy_recipient_token_find": 1.0, - "eval_accuracy_recipient_token_find_total": 57, - "eval_accuracy_recipient_token_get": 0.9795918367346939, - "eval_accuracy_recipient_token_get_total": 196, - "eval_accuracy_recipient_token_identify": 0.8333333333333334, - "eval_accuracy_recipient_token_identify_total": 6, - "eval_loss": 0.21341286599636078, - "eval_perplexity": 1.0333647517086502, - "eval_runtime": 112.5857, - "eval_samples_per_second": 0.897, - "eval_steps_per_second": 0.453, - "eval_total_number_recipient_token": 902, - "step": 318 + "train_accuracy_recipient_token_get_total": 1, + "train_total_number_recipient_token": 14 }, { - "epoch": 0.9426877470355731, - "step": 318, - "train_accuracy_first_token_arguments": 0.7777777777777778, - "train_accuracy_first_token_arguments_total": 18, - "train_accuracy_recipient_token": 0.8947368421052632, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 3, - "train_accuracy_recipient_token_get": 0.8, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_identify": 0.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_total_number_recipient_token": 19 + "epoch": 0.9861660079051383, + "grad_norm": 1.9933889585193483, + "learning_rate": 2.5173336467135266e-09, + "loss": 0.2492, + "step": 499 }, { - "epoch": 0.9426877470355731, - "step": 318, - "train_accuracy_first_token_arguments": 0.9375, - "train_accuracy_first_token_arguments_total": 16, - "train_accuracy_recipient_token": 0.8888888888888888, + "epoch": 0.9861660079051383, + "step": 499, + "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_get": 0.8571428571428571, - "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_property": 1.0, + "train_accuracy_recipient_token_property_total": 4, "train_total_number_recipient_token": 18 }, { - "epoch": 0.9426877470355731, - "step": 318, - "train_accuracy_first_token_arguments": 0.8461538461538461, - "train_accuracy_first_token_arguments_total": 26, - "train_accuracy_recipient_token": 0.85, - "train_accuracy_recipient_token_all": 0.75, - "train_accuracy_recipient_token_all_total": 12, + "epoch": 0.9861660079051383, + "step": 499, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_ex": 1.0, + "train_accuracy_recipient_token_ex_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_translate": 1.0, - "train_accuracy_recipient_token_translate_total": 4, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.9456521739130435, - "grad_norm": 2.4406440061460377, - "learning_rate": 3.7517192664685844e-08, - "loss": 0.3195, - "step": 319 + "epoch": 0.9881422924901185, + "grad_norm": 1.780716845359166, + "learning_rate": 1.8495519823308329e-09, + "loss": 0.1323, + "step": 500 }, { - "epoch": 0.9456521739130435, - "step": 319, - "train_accuracy_first_token_arguments": 0.9375, - "train_accuracy_first_token_arguments_total": 16, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.9881422924901185, + "step": 500, + "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 8, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 2, + "train_accuracy_recipient_token_calculate": 0.5, + "train_accuracy_recipient_token_calculate_total": 2, + "train_accuracy_recipient_token_find": 0.6666666666666666, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.9456521739130435, - "step": 319, - "train_accuracy_first_token_arguments": 0.7142857142857143, - "train_accuracy_first_token_arguments_total": 21, + "epoch": 0.9881422924901185, + "step": 500, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_art": 1.0, + "train_accuracy_recipient_token_art_total": 4, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_accuracy_recipient_token_search": 1.0, - "train_accuracy_recipient_token_search_total": 2, - "train_accuracy_recipient_token_validate": 1.0, - "train_accuracy_recipient_token_validate_total": 1, + "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { - "epoch": 0.9456521739130435, - "step": 319, - "train_accuracy_first_token_arguments": 0.9375, - "train_accuracy_first_token_arguments_total": 16, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.9901185770750988, + "grad_norm": 2.296947318897557, + "learning_rate": 1.284459498280266e-09, + "loss": 0.134, + "step": 501 + }, + { + "epoch": 0.9901185770750988, + "step": 501, + "train_accuracy_recipient_token": 0.7777777777777778, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_schedule": 0.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_accuracy_recipient_token_track": 0.5, + "train_accuracy_recipient_token_track_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.9901185770750988, + "step": 501, + "train_accuracy_recipient_token": 0.9473684210526315, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_plant": 1.0, - "train_accuracy_recipient_token_plant_total": 3, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.9486166007905138, - "grad_norm": 2.041210168988887, - "learning_rate": 3.347347934333778e-08, - "loss": 0.2436, - "step": 320 + "epoch": 0.9920948616600791, + "grad_norm": 1.8997002986472211, + "learning_rate": 8.220794233004814e-10, + "loss": 0.2113, + "step": 502 }, { - "epoch": 0.9486166007905138, - "step": 320, - "train_accuracy_first_token_arguments": 0.9375, - "train_accuracy_first_token_arguments_total": 16, - "train_accuracy_recipient_token": 0.85, - "train_accuracy_recipient_token_all": 0.8, - "train_accuracy_recipient_token_all_total": 15, + "epoch": 0.9920948616600791, + "step": 502, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.9486166007905138, - "step": 320, - "train_accuracy_first_token_arguments": 0.5925925925925926, - "train_accuracy_first_token_arguments_total": 54, - "train_accuracy_recipient_token": 0.8666666666666667, - "train_accuracy_recipient_token_all": 0.9090909090909091, + "epoch": 0.9920948616600791, + "step": 502, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, - "train_accuracy_recipient_token_predict": 1.0, - "train_accuracy_recipient_token_predict_total": 1, - "train_accuracy_recipient_token_stamp": 0.5, - "train_accuracy_recipient_token_stamp_total": 2, - "train_total_number_recipient_token": 15 + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_write": 1.0, + "train_accuracy_recipient_token_write_total": 3, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.9486166007905138, - "step": 320, - "train_accuracy_first_token_arguments": 0.9565217391304348, - "train_accuracy_first_token_arguments_total": 23, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.9940711462450593, + "grad_norm": 1.812032510476914, + "learning_rate": 4.624307640249681e-10, + "loss": 0.2103, + "step": 503 + }, + { + "epoch": 0.9940711462450593, + "step": 503, + "train_accuracy_recipient_token": 0.6111111111111112, + "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 0.0, + "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 9, + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_opt": 0.0, + "train_accuracy_recipient_token_opt_total": 1, + "train_accuracy_recipient_token_plan": 0.0, + "train_accuracy_recipient_token_plan_total": 1, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.9940711462450593, + "step": 503, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 15, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, "train_total_number_recipient_token": 19 }, { - "epoch": 0.9515810276679841, - "grad_norm": 2.151431626558894, - "learning_rate": 2.9658824818044328e-08, - "loss": 0.309, - "step": 321 + "epoch": 0.9960474308300395, + "grad_norm": 1.9968965605808304, + "learning_rate": 2.0552830420184077e-10, + "loss": 0.2541, + "step": 504 }, { - "epoch": 0.9515810276679841, - "step": 321, - "train_accuracy_first_token_arguments": 0.9811320754716981, - "train_accuracy_first_token_arguments_total": 53, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_calculate": 1.0, - "train_accuracy_recipient_token_calculate_total": 2, + "epoch": 0.9960474308300395, + "step": 504, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, - "train_accuracy_recipient_token_optimize": 0.0, - "train_accuracy_recipient_token_optimize_total": 1, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_order": 1.0, + "train_accuracy_recipient_token_order_total": 1, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.9515810276679841, - "step": 321, - "train_accuracy_first_token_arguments": 0.90625, - "train_accuracy_first_token_arguments_total": 32, - "train_accuracy_recipient_token": 1.0, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.9960474308300395, + "step": 504, + "train_accuracy_recipient_token": 0.9285714285714286, + "train_accuracy_recipient_token_all": 0.875, + "train_accuracy_recipient_token_all_total": 8, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 14 + }, + { + "epoch": 0.9960474308300395, + "step": 504, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 9, + "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 19 }, { - "epoch": 0.9515810276679841, - "step": 321, - "train_accuracy_first_token_arguments": 0.9545454545454546, - "train_accuracy_first_token_arguments_total": 22, + "epoch": 0.9960474308300395, + "step": 504, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 2, - "train_total_number_recipient_token": 16 - }, - { - "epoch": 0.9545454545454546, - "grad_norm": 2.5639718605032216, - "learning_rate": 2.607358334413779e-08, - "loss": 0.3681, - "step": 322 + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.9545454545454546, - "step": 322, - "train_accuracy_first_token_arguments": 0.8148148148148148, - "train_accuracy_first_token_arguments_total": 27, + "epoch": 0.9960474308300395, + "step": 504, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_create": 1.0, - "train_accuracy_recipient_token_create_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_ne": 1.0, - "train_accuracy_recipient_token_ne_total": 3, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 20 }, { - "epoch": 0.9545454545454546, - "step": 322, - "train_accuracy_first_token_arguments": 0.8888888888888888, - "train_accuracy_first_token_arguments_total": 18, - "train_accuracy_recipient_token": 0.9333333333333333, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.9960474308300395, + "step": 504, + "train_accuracy_recipient_token": 0.9375, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_convert": 1.0, - "train_accuracy_recipient_token_convert_total": 1, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, - "train_total_number_recipient_token": 15 + "train_accuracy_recipient_token_ident": 0.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_translate": 1.0, + "train_accuracy_recipient_token_translate_total": 2, + "train_total_number_recipient_token": 16 }, { - "epoch": 0.9545454545454546, - "step": 322, - "train_accuracy_first_token_arguments": 0.9, - "train_accuracy_first_token_arguments_total": 30, - "train_accuracy_recipient_token": 0.9375, - "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_customize": 1.0, - "train_accuracy_recipient_token_customize_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, + "epoch": 0.9960474308300395, + "step": 504, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_total_number_recipient_token": 16 + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_histor": 1.0, + "train_accuracy_recipient_token_histor_total": 3, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.9575098814229249, - "grad_norm": 2.8167178374609607, - "learning_rate": 2.2718087872060925e-08, - "loss": 0.2343, - "step": 323 + "epoch": 0.9960474308300395, + "step": 504, + "train_accuracy_recipient_token": 0.7647058823529411, + "train_accuracy_recipient_token_all": 0.7272727272727273, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 0.8, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_validate": 1.0, + "train_accuracy_recipient_token_validate_total": 1, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.9575098814229249, - "step": 323, - "train_accuracy_first_token_arguments": 0.6744186046511628, - "train_accuracy_first_token_arguments_total": 172, + "epoch": 0.9960474308300395, + "step": 504, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_c": 1.0, - "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 4, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.9960474308300395, + "step": 504, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_pool": 1.0, - "train_accuracy_recipient_token_pool_total": 4, - "train_total_number_recipient_token": 20 + "train_total_number_recipient_token": 17 }, { - "epoch": 0.9575098814229249, - "step": 323, - "train_accuracy_first_token_arguments": 0.9285714285714286, - "train_accuracy_first_token_arguments_total": 14, + "epoch": 0.9960474308300395, + "step": 504, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_c": 1.0, - "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 3, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, - "train_accuracy_recipient_token_robot": 1.0, - "train_accuracy_recipient_token_robot_total": 2, + "train_accuracy_recipient_token_find_total": 5, + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_interpret": 1.0, + "train_accuracy_recipient_token_interpret_total": 2, "train_total_number_recipient_token": 18 }, { - "epoch": 0.9575098814229249, - "step": 323, - "train_accuracy_first_token_arguments": 0.9583333333333334, - "train_accuracy_first_token_arguments_total": 24, + "epoch": 0.9960474308300395, + "step": 504, "train_accuracy_recipient_token": 0.8421052631578947, - "train_accuracy_recipient_token_all": 0.8181818181818182, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_get": 0.875, - "train_accuracy_recipient_token_get_total": 8, + "train_accuracy_recipient_token_all": 0.8666666666666667, + "train_accuracy_recipient_token_all_total": 15, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_find": 0.6666666666666666, + "train_accuracy_recipient_token_find_total": 3, "train_total_number_recipient_token": 19 }, { - "epoch": 0.9604743083003953, - "grad_norm": 1.6492281178442556, - "learning_rate": 1.9592650016444503e-08, - "loss": 0.1848, - "step": 324 - }, - { - "epoch": 0.9604743083003953, - "step": 324, - "train_accuracy_first_token_arguments": 0.896551724137931, - "train_accuracy_first_token_arguments_total": 29, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.9960474308300395, + "step": 504, + "train_accuracy_recipient_token": 0.95, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 4, + "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_learn": 1.0, - "train_accuracy_recipient_token_learn_total": 1, - "train_total_number_recipient_token": 19 - }, - { - "epoch": 0.9604743083003953, - "step": 324, - "train_accuracy_first_token_arguments": 0.9166666666666666, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.95, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_get": 0.8888888888888888, - "train_accuracy_recipient_token_get_total": 9, + "train_accuracy_recipient_token_tour": 0.75, + "train_accuracy_recipient_token_tour_total": 4, "train_total_number_recipient_token": 20 }, { - "epoch": 0.9604743083003953, - "step": 324, - "train_accuracy_first_token_arguments": 0.5384615384615384, - "train_accuracy_first_token_arguments_total": 13, - "train_accuracy_recipient_token": 0.9047619047619048, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_count": 0.0, - "train_accuracy_recipient_token_count_total": 1, - "train_accuracy_recipient_token_get": 0.8, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_subtract": 1.0, - "train_accuracy_recipient_token_subtract_total": 1, - "train_accuracy_recipient_token_track": 1.0, - "train_accuracy_recipient_token_track_total": 1, - "train_total_number_recipient_token": 21 - }, - { - "epoch": 0.9634387351778656, - "grad_norm": 1.754485036239026, - "learning_rate": 1.6697560027171543e-08, - "loss": 0.1782, - "step": 325 - }, - { - "epoch": 0.9634387351778656, - "step": 325, - "train_accuracy_first_token_arguments": 0.9642857142857143, - "train_accuracy_first_token_arguments_total": 28, - "train_accuracy_recipient_token": 0.8888888888888888, + "epoch": 0.9960474308300395, + "step": 504, + "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compare": 0.0, - "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_compile": 1.0, - "train_accuracy_recipient_token_compile_total": 1, + "train_accuracy_recipient_token_book": 0.0, + "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.9634387351778656, - "step": 325, - "train_accuracy_first_token_arguments": 0.7857142857142857, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 0.9473684210526315, + "epoch": 0.9960474308300395, + "step": 504, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_extract": 1.0, - "train_accuracy_recipient_token_extract_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_recogn": 0.0, - "train_accuracy_recipient_token_recogn_total": 1, - "train_accuracy_recipient_token_translate": 1.0, - "train_accuracy_recipient_token_translate_total": 1, - "train_total_number_recipient_token": 19 - }, - { - "epoch": 0.9634387351778656, - "step": 325, - "train_accuracy_first_token_arguments": 0.5333333333333333, - "train_accuracy_first_token_arguments_total": 15, - "train_accuracy_recipient_token": 0.9333333333333333, - "train_accuracy_recipient_token_all": 0.9090909090909091, - "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, - "train_accuracy_recipient_token_verify": 1.0, - "train_accuracy_recipient_token_verify_total": 1, - "train_total_number_recipient_token": 15 - }, - { - "epoch": 0.9664031620553359, - "grad_norm": 2.1745450667504445, - "learning_rate": 1.4033086762419989e-08, - "loss": 0.3463, - "step": 326 - }, - { - "epoch": 0.9664031620553359, - "step": 326, - "train_accuracy_first_token_arguments": 0.9322033898305084, - "train_accuracy_first_token_arguments_total": 59, - "train_accuracy_recipient_token": 0.7777777777777778, - "train_accuracy_recipient_token_all": 0.7272727272727273, - "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 3, - "train_accuracy_recipient_token_check": 0.0, - "train_accuracy_recipient_token_check_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.9664031620553359, - "step": 326, - "train_accuracy_first_token_arguments": 0.9375, - "train_accuracy_first_token_arguments_total": 16, + "epoch": 0.9960474308300395, + "step": 504, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compile": 1.0, - "train_accuracy_recipient_token_compile_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 4, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_hang": 1.0, + "train_accuracy_recipient_token_hang_total": 1, + "train_accuracy_recipient_token_re": 1.0, + "train_accuracy_recipient_token_re_total": 1, + "train_accuracy_recipient_token_remove": 1.0, + "train_accuracy_recipient_token_remove_total": 1, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.9664031620553359, - "step": 326, - "train_accuracy_first_token_arguments": 0.6875, - "train_accuracy_first_token_arguments_total": 16, - "train_accuracy_recipient_token": 0.8421052631578947, + "epoch": 0.9960474308300395, + "step": 504, + "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_compare": 1.0, - "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_download": 1.0, - "train_accuracy_recipient_token_download_total": 1, - "train_accuracy_recipient_token_get": 0.8333333333333334, - "train_accuracy_recipient_token_get_total": 6, - "train_accuracy_recipient_token_search": 0.0, - "train_accuracy_recipient_token_search_total": 1, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_invest": 1.0, + "train_accuracy_recipient_token_invest_total": 1, "train_total_number_recipient_token": 19 }, { - "epoch": 0.9693675889328063, - "grad_norm": 2.1355214484085425, - "learning_rate": 1.1599477663696845e-08, - "loss": 0.246, - "step": 327 - }, - { - "epoch": 0.9693675889328063, - "step": 327, - "train_accuracy_first_token_arguments": 0.9230769230769231, - "train_accuracy_first_token_arguments_total": 13, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.9960474308300395, + "step": 504, + "train_accuracy_recipient_token": 0.9375, + "train_accuracy_recipient_token_add": 1.0, + "train_accuracy_recipient_token_add_total": 1, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_construct": 1.0, + "train_accuracy_recipient_token_construct_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_total_number_recipient_token": 17 + "train_accuracy_recipient_token_select": 1.0, + "train_accuracy_recipient_token_select_total": 2, + "train_total_number_recipient_token": 16 }, { - "epoch": 0.9693675889328063, - "step": 327, - "train_accuracy_first_token_arguments": 0.8125, - "train_accuracy_first_token_arguments_total": 16, - "train_accuracy_recipient_token": 0.9411764705882353, - "train_accuracy_recipient_token_all": 0.9, + "epoch": 0.9960474308300395, + "step": 504, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 1, - "train_accuracy_recipient_token_deb": 1.0, - "train_accuracy_recipient_token_deb_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_plan": 1.0, - "train_accuracy_recipient_token_plan_total": 1, + "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { - "epoch": 0.9693675889328063, - "step": 327, - "train_accuracy_first_token_arguments": 0.7727272727272727, - "train_accuracy_first_token_arguments_total": 22, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_c": 1.0, - "train_accuracy_recipient_token_c_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 4, - "train_accuracy_recipient_token_get": 0.75, - "train_accuracy_recipient_token_get_total": 4, - "train_total_number_recipient_token": 19 - }, - { - "epoch": 0.9723320158102767, - "grad_norm": 1.6782990768747863, - "learning_rate": 9.396958732856843e-09, - "loss": 0.2385, - "step": 328 + "epoch": 0.9960474308300395, + "step": 504, + "train_accuracy_recipient_token": 0.9, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_art": 1.0, + "train_accuracy_recipient_token_art_total": 2, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_histor": 0.0, + "train_accuracy_recipient_token_histor_total": 1, + "train_accuracy_recipient_token_m": 1.0, + "train_accuracy_recipient_token_m_total": 1, + "train_total_number_recipient_token": 20 }, { - "epoch": 0.9723320158102767, - "step": 328, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 29, + "epoch": 0.9960474308300395, + "step": 504, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, - "train_accuracy_recipient_token_book": 1.0, - "train_accuracy_recipient_token_book_total": 1, - "train_accuracy_recipient_token_find": 0.75, - "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_gr": 1.0, + "train_accuracy_recipient_token_gr_total": 3, + "train_accuracy_recipient_token_ident": 0.5, + "train_accuracy_recipient_token_ident_total": 2, "train_total_number_recipient_token": 20 }, { - "epoch": 0.9723320158102767, - "step": 328, - "train_accuracy_first_token_arguments": 0.85, - "train_accuracy_first_token_arguments_total": 20, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 0.0, - "train_accuracy_recipient_token_find_total": 1, + "epoch": 0.9960474308300395, + "step": 504, + "train_accuracy_recipient_token": 0.9285714285714286, + "train_accuracy_recipient_token_all": 0.8888888888888888, + "train_accuracy_recipient_token_all_total": 9, + "train_accuracy_recipient_token_draw": 1.0, + "train_accuracy_recipient_token_draw_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 8, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 14 }, { - "epoch": 0.9723320158102767, - "step": 328, - "train_accuracy_first_token_arguments": 0.631578947368421, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.9444444444444444, + "epoch": 0.9960474308300395, + "step": 504, + "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_review": 0.5, + "train_accuracy_recipient_token_review_total": 2, + "train_total_number_recipient_token": 18 + }, + { + "epoch": 0.9960474308300395, + "step": 504, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_compare": 1.0, + "train_accuracy_recipient_token_compare_total": 1, + "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_accuracy_recipient_token_park": 1.0, - "train_accuracy_recipient_token_park_total": 4, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.9752964426877471, - "grad_norm": 2.45983910491564, - "learning_rate": 7.425734511117e-09, - "loss": 0.3068, - "step": 329 + "epoch": 0.9960474308300395, + "step": 504, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_create": 0.5, + "train_accuracy_recipient_token_create_total": 2, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.9752964426877471, - "step": 329, - "train_accuracy_first_token_arguments": 0.95, - "train_accuracy_first_token_arguments_total": 40, + "epoch": 0.9960474308300395, + "step": 504, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_explain": 1.0, - "train_accuracy_recipient_token_explain_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_translate": 1.0, - "train_accuracy_recipient_token_translate_total": 1, + "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { - "epoch": 0.9752964426877471, - "step": 329, - "train_accuracy_first_token_arguments": 0.95, - "train_accuracy_first_token_arguments_total": 20, + "epoch": 0.9960474308300395, + "step": 504, + "train_accuracy_recipient_token": 0.8421052631578947, + "train_accuracy_recipient_token_all": 0.8181818181818182, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_get": 0.875, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.9960474308300395, + "step": 504, + "train_accuracy_recipient_token": 0.8947368421052632, + "train_accuracy_recipient_token_all": 0.8888888888888888, + "train_accuracy_recipient_token_all_total": 9, + "train_accuracy_recipient_token_custom": 1.0, + "train_accuracy_recipient_token_custom_total": 4, + "train_accuracy_recipient_token_get": 0.8333333333333334, + "train_accuracy_recipient_token_get_total": 6, + "train_total_number_recipient_token": 19 + }, + { + "epoch": 0.9960474308300395, + "step": 504, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 7, + "train_accuracy_recipient_token_eng": 1.0, + "train_accuracy_recipient_token_eng_total": 2, + "train_accuracy_recipient_token_list": 1.0, + "train_accuracy_recipient_token_list_total": 1, + "train_accuracy_recipient_token_lock": 1.0, + "train_accuracy_recipient_token_lock_total": 1, + "train_accuracy_recipient_token_tem": 1.0, + "train_accuracy_recipient_token_tem_total": 3, "train_total_number_recipient_token": 17 }, { - "epoch": 0.9752964426877471, - "step": 329, - "train_accuracy_first_token_arguments": 0.8666666666666667, - "train_accuracy_first_token_arguments_total": 15, + "epoch": 0.9960474308300395, + "step": 504, + "train_accuracy_recipient_token": 0.9375, + "train_accuracy_recipient_token_all": 0.9090909090909091, + "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_total_number_recipient_token": 16 + }, + { + "epoch": 0.9960474308300395, + "step": 504, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_art": 1.0, - "train_accuracy_recipient_token_art_total": 1, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_car": 1.0, + "train_accuracy_recipient_token_car_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_validate": 1.0, - "train_accuracy_recipient_token_validate_total": 1, - "train_total_number_recipient_token": 16 - }, - { - "epoch": 0.9782608695652174, - "grad_norm": 1.9407711881422764, - "learning_rate": 5.685988060059045e-09, - "loss": 0.2393, - "step": 330 + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 19 }, { - "epoch": 0.9782608695652174, - "step": 330, - "train_accuracy_first_token_arguments": 0.8571428571428571, - "train_accuracy_first_token_arguments_total": 28, - "train_accuracy_recipient_token": 0.9473684210526315, - "train_accuracy_recipient_token_all": 1.0, + "epoch": 0.9960474308300395, + "step": 504, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_get": 0.8888888888888888, - "train_accuracy_recipient_token_get_total": 9, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 5, + "train_accuracy_recipient_token_invest": 1.0, + "train_accuracy_recipient_token_invest_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.9782608695652174, - "step": 330, - "train_accuracy_first_token_arguments": 0.7142857142857143, - "train_accuracy_first_token_arguments_total": 21, + "epoch": 0.9960474308300395, + "step": 504, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_filter": 1.0, - "train_accuracy_recipient_token_filter_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 8, + "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { - "epoch": 0.9782608695652174, - "step": 330, - "train_accuracy_first_token_arguments": 0.9166666666666666, - "train_accuracy_first_token_arguments_total": 12, - "train_accuracy_recipient_token": 0.8823529411764706, - "train_accuracy_recipient_token_all": 0.8, + "epoch": 0.9960474308300395, + "step": 504, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_motor": 1.0, - "train_accuracy_recipient_token_motor_total": 3, + "train_accuracy_recipient_token_bird": 1.0, + "train_accuracy_recipient_token_bird_total": 2, + "train_accuracy_recipient_token_c": 1.0, + "train_accuracy_recipient_token_c_total": 1, + "train_accuracy_recipient_token_histor": 1.0, + "train_accuracy_recipient_token_histor_total": 1, + "train_accuracy_recipient_token_symbol": 1.0, + "train_accuracy_recipient_token_symbol_total": 1, "train_total_number_recipient_token": 17 }, { - "epoch": 0.9812252964426877, - "grad_norm": 1.9607276731011434, - "learning_rate": 4.17788094463023e-09, - "loss": 0.1726, - "step": 331 - }, - { - "epoch": 0.9812252964426877, - "step": 331, - "train_accuracy_first_token_arguments": 0.8666666666666667, - "train_accuracy_first_token_arguments_total": 15, - "train_accuracy_recipient_token": 0.6842105263157895, - "train_accuracy_recipient_token_all": 0.6666666666666666, - "train_accuracy_recipient_token_all_total": 15, + "epoch": 0.9960474308300395, + "step": 504, + "train_accuracy_recipient_token": 0.875, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 0.6666666666666666, - "train_accuracy_recipient_token_get_total": 3, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_organ": 1.0, + "train_accuracy_recipient_token_organ_total": 1, + "train_accuracy_recipient_token_rent": 1.0, + "train_accuracy_recipient_token_rent_total": 1, + "train_total_number_recipient_token": 16 }, { - "epoch": 0.9812252964426877, - "step": 331, - "train_accuracy_first_token_arguments": 0.8571428571428571, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 0.9473684210526315, + "epoch": 0.9960474308300395, + "step": 504, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_building": 1.0, - "train_accuracy_recipient_token_building_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, - "train_accuracy_recipient_token_get": 0.75, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_ident": 1.0, + "train_accuracy_recipient_token_ident_total": 1, + "train_accuracy_recipient_token_predict": 1.0, + "train_accuracy_recipient_token_predict_total": 1, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.9812252964426877, - "step": 331, - "train_accuracy_first_token_arguments": 0.8, - "train_accuracy_first_token_arguments_total": 30, + "epoch": 0.9960474308300395, + "step": 504, "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_evaluate": 0.0, + "train_accuracy_recipient_token_evaluate_total": 1, "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 2, + "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, - "train_accuracy_recipient_token_schedule": 0.5, - "train_accuracy_recipient_token_schedule_total": 2, - "train_accuracy_recipient_token_table": 0.5, - "train_accuracy_recipient_token_table_total": 2, + "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { - "epoch": 0.9841897233201581, - "grad_norm": 1.9808670794420047, - "learning_rate": 2.9015532181397854e-09, - "loss": 0.2188, - "step": 332 - }, - { - "epoch": 0.9841897233201581, - "step": 332, - "train_accuracy_first_token_arguments": 0.8421052631578947, - "train_accuracy_first_token_arguments_total": 19, - "train_accuracy_recipient_token": 0.9473684210526315, + "epoch": 0.9960474308300395, + "step": 504, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_generate": 0.5, - "train_accuracy_recipient_token_generate_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_sum": 1.0, - "train_accuracy_recipient_token_sum_total": 1, - "train_accuracy_recipient_token_translate": 1.0, - "train_accuracy_recipient_token_translate_total": 2, + "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { - "epoch": 0.9841897233201581, - "step": 332, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 14, - "train_accuracy_recipient_token": 0.7857142857142857, - "train_accuracy_recipient_token_all": 0.8, + "epoch": 0.9960474308300395, + "step": 504, + "train_accuracy_recipient_token": 0.9444444444444444, + "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_calculate": 1.0, - "train_accuracy_recipient_token_calculate_total": 1, - "train_accuracy_recipient_token_estimate": 1.0, - "train_accuracy_recipient_token_estimate_total": 1, - "train_accuracy_recipient_token_evaluate": 0.0, - "train_accuracy_recipient_token_evaluate_total": 1, - "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, - "train_total_number_recipient_token": 14 - }, - { - "epoch": 0.9841897233201581, - "step": 332, - "train_accuracy_first_token_arguments": 0.96875, - "train_accuracy_first_token_arguments_total": 64, - "train_accuracy_recipient_token": 0.8888888888888888, - "train_accuracy_recipient_token_all": 0.8181818181818182, - "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 2, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 3, - "train_accuracy_recipient_token_property": 1.0, - "train_accuracy_recipient_token_property_total": 4, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_locate": 0.0, + "train_accuracy_recipient_token_locate_total": 1, + "train_accuracy_recipient_token_schedule": 1.0, + "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 18 }, { - "epoch": 0.9871541501976284, - "grad_norm": 2.0556535601550325, - "learning_rate": 1.857123409250705e-09, - "loss": 0.2487, - "step": 333 + "epoch": 0.9960474308300395, + "step": 504, + "train_accuracy_recipient_token": 0.9411764705882353, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 4, + "train_accuracy_recipient_token_park": 1.0, + "train_accuracy_recipient_token_park_total": 2, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.9871541501976284, - "step": 333, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 16, - "train_accuracy_recipient_token": 1.0, + "epoch": 0.9960474308300395, + "step": 504, + "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_explain": 1.0, - "train_accuracy_recipient_token_explain_total": 1, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 1, + "train_accuracy_recipient_token_art": 1.0, + "train_accuracy_recipient_token_art_total": 3, + "train_accuracy_recipient_token_artist": 0.0, + "train_accuracy_recipient_token_artist_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 6, + "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { - "epoch": 0.9871541501976284, - "step": 333, - "train_accuracy_first_token_arguments": 0.8571428571428571, - "train_accuracy_first_token_arguments_total": 21, - "train_accuracy_recipient_token": 0.9, + "epoch": 0.9960474308300395, + "step": 504, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 2, - "train_accuracy_recipient_token_calculate": 0.5, - "train_accuracy_recipient_token_calculate_total": 2, - "train_accuracy_recipient_token_find": 0.6666666666666666, - "train_accuracy_recipient_token_find_total": 3, - "train_accuracy_recipient_token_recommend": 1.0, - "train_accuracy_recipient_token_recommend_total": 1, - "train_total_number_recipient_token": 20 + "train_accuracy_recipient_token_analy": 1.0, + "train_accuracy_recipient_token_analy_total": 2, + "train_accuracy_recipient_token_check": 1.0, + "train_accuracy_recipient_token_check_total": 2, + "train_accuracy_recipient_token_object": 1.0, + "train_accuracy_recipient_token_object_total": 3, + "train_accuracy_recipient_token_transfer": 1.0, + "train_accuracy_recipient_token_transfer_total": 1, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.9871541501976284, - "step": 333, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 17, - "train_accuracy_recipient_token": 0.9473684210526315, + "epoch": 0.9960474308300395, + "step": 504, + "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_artifact": 1.0, - "train_accuracy_recipient_token_artifact_total": 4, + "train_accuracy_recipient_token_fruit": 1.0, + "train_accuracy_recipient_token_fruit_total": 4, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_get_total": 3, + "train_total_number_recipient_token": 17 }, { - "epoch": 0.9901185770750988, - "grad_norm": 2.092295914240381, - "learning_rate": 1.0446885109746673e-09, - "loss": 0.1504, - "step": 334 + "epoch": 0.9960474308300395, + "step": 504, + "train_accuracy_recipient_token": 1.0, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 8, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.9901185770750988, - "step": 334, - "train_accuracy_first_token_arguments": 0.8823529411764706, - "train_accuracy_first_token_arguments_total": 34, - "train_accuracy_recipient_token": 0.7222222222222222, - "train_accuracy_recipient_token_all": 0.8181818181818182, - "train_accuracy_recipient_token_all_total": 11, + "epoch": 0.9960474308300395, + "step": 504, + "train_accuracy_recipient_token": 0.875, + "train_accuracy_recipient_token_all": 0.8, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_book": 1.0, + "train_accuracy_recipient_token_book_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_schedule": 0.0, - "train_accuracy_recipient_token_schedule_total": 1, - "train_accuracy_recipient_token_track": 0.0, - "train_accuracy_recipient_token_track_total": 2, - "train_total_number_recipient_token": 18 + "train_accuracy_recipient_token_get_total": 3, + "train_accuracy_recipient_token_organ": 1.0, + "train_accuracy_recipient_token_organ_total": 1, + "train_total_number_recipient_token": 16 }, { - "epoch": 0.9901185770750988, - "step": 334, - "train_accuracy_first_token_arguments": 0.9230769230769231, - "train_accuracy_first_token_arguments_total": 13, + "epoch": 0.9960474308300395, + "step": 504, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_analyze": 1.0, - "train_accuracy_recipient_token_analyze_total": 1, - "train_accuracy_recipient_token_check": 1.0, - "train_accuracy_recipient_token_check_total": 1, + "train_accuracy_recipient_token_decode": 1.0, + "train_accuracy_recipient_token_decode_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 5, - "train_accuracy_recipient_token_identify": 1.0, - "train_accuracy_recipient_token_identify_total": 1, - "train_total_number_recipient_token": 19 + "train_accuracy_recipient_token_get_total": 1, + "train_accuracy_recipient_token_interpret": 1.0, + "train_accuracy_recipient_token_interpret_total": 1, + "train_accuracy_recipient_token_stage": 1.0, + "train_accuracy_recipient_token_stage_total": 4, + "train_total_number_recipient_token": 18 }, { - "epoch": 0.9901185770750988, - "step": 334, - "train_accuracy_first_token_arguments": 0.9166666666666666, - "train_accuracy_first_token_arguments_total": 12, + "epoch": 0.9960474308300395, + "step": 504, + "train_accuracy_recipient_token": 0.8823529411764706, + "train_accuracy_recipient_token_all": 0.9, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_estimate": 1.0, + "train_accuracy_recipient_token_estimate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_get": 0.5, + "train_accuracy_recipient_token_get_total": 2, + "train_accuracy_recipient_token_rent": 1.0, + "train_accuracy_recipient_token_rent_total": 1, + "train_total_number_recipient_token": 17 + }, + { + "epoch": 0.9960474308300395, + "step": 504, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, @@ -18562,68 +19767,58 @@ "train_total_number_recipient_token": 18 }, { - "epoch": 0.9930830039525692, - "grad_norm": 2.0272702589635965, - "learning_rate": 4.6432397166285e-10, - "loss": 0.1775, - "step": 335 - }, - { - "epoch": 0.9930830039525692, - "step": 335, - "train_accuracy_first_token_arguments": 0.4375, - "train_accuracy_first_token_arguments_total": 32, + "epoch": 0.9960474308300395, + "step": 504, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, - "train_accuracy_recipient_token_all_total": 11, + "train_accuracy_recipient_token_all_total": 10, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, - "train_accuracy_recipient_token_write": 1.0, - "train_accuracy_recipient_token_write_total": 3, "train_total_number_recipient_token": 18 }, { - "epoch": 0.9930830039525692, - "step": 335, - "train_accuracy_first_token_arguments": 0.8769230769230769, - "train_accuracy_first_token_arguments_total": 65, - "train_accuracy_recipient_token": 0.7222222222222222, + "epoch": 0.9960474308300395, + "step": 504, + "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 0.3333333333333333, - "train_accuracy_recipient_token_find_total": 3, + "train_accuracy_recipient_token_calculate": 1.0, + "train_accuracy_recipient_token_calculate_total": 1, + "train_accuracy_recipient_token_find": 1.0, + "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, - "train_accuracy_recipient_token_get_total": 1, - "train_accuracy_recipient_token_optimize": 0.0, - "train_accuracy_recipient_token_optimize_total": 1, - "train_accuracy_recipient_token_plan": 0.0, - "train_accuracy_recipient_token_plan_total": 1, + "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 18 }, { - "epoch": 0.9930830039525692, - "step": 335, - "train_accuracy_first_token_arguments": 0.85, - "train_accuracy_first_token_arguments_total": 20, - "train_accuracy_recipient_token": 0.8421052631578947, - "train_accuracy_recipient_token_all": 0.8, - "train_accuracy_recipient_token_all_total": 15, - "train_accuracy_recipient_token_find": 1.0, - "train_accuracy_recipient_token_find_total": 4, - "train_total_number_recipient_token": 19 + "epoch": 0.9960474308300395, + "step": 504, + "train_accuracy_recipient_token": 0.8888888888888888, + "train_accuracy_recipient_token_all": 1.0, + "train_accuracy_recipient_token_all_total": 5, + "train_accuracy_recipient_token_f": 0.0, + "train_accuracy_recipient_token_f_total": 1, + "train_accuracy_recipient_token_pest": 1.0, + "train_accuracy_recipient_token_pest_total": 1, + "train_accuracy_recipient_token_plant": 1.0, + "train_accuracy_recipient_token_plant_total": 1, + "train_accuracy_recipient_token_sun": 1.0, + "train_accuracy_recipient_token_sun_total": 1, + "train_total_number_recipient_token": 9 }, { "epoch": 0.9960474308300395, - "grad_norm": 1.9403721941845617, - "learning_rate": 1.160836880001459e-10, - "loss": 0.3228, - "step": 336 + "eval_loss": 0.18871143460273743, + "eval_runtime": 412.3849, + "eval_samples_per_second": 0.245, + "eval_steps_per_second": 0.124, + "step": 504 }, { "epoch": 0.9960474308300395, - "step": 336, - "train_accuracy_first_token_arguments": 0.868421052631579, - "train_accuracy_first_token_arguments_total": 38, + "step": 504, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.8333333333333334, "train_accuracy_recipient_token_all_total": 12, @@ -18637,16 +19832,14 @@ }, { "epoch": 0.9960474308300395, - "step": 336, - "train_accuracy_first_token_arguments": 0.9375, - "train_accuracy_first_token_arguments_total": 16, + "step": 504, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, - "train_accuracy_recipient_token_explain": 1.0, - "train_accuracy_recipient_token_explain_total": 1, + "train_accuracy_recipient_token_ex": 1.0, + "train_accuracy_recipient_token_ex_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, @@ -18654,38 +19847,53 @@ "train_total_number_recipient_token": 19 }, { - "epoch": 0.9960474308300395, - "step": 336, - "train_accuracy_first_token_arguments": 1.0, - "train_accuracy_first_token_arguments_total": 18, - "train_accuracy_recipient_token": 0.95, + "epoch": 0.9980237154150198, + "grad_norm": 1.6014381828385609, + "learning_rate": 5.138260408488194e-11, + "loss": 0.1502, + "step": 505 + }, + { + "epoch": 0.9980237154150198, + "step": 505, + "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, - "train_accuracy_recipient_token_find": 0.8, + "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 20 }, { - "epoch": 0.9990118577075099, - "grad_norm": 1.5288849501967092, + "epoch": 0.9980237154150198, + "step": 505, + "train_accuracy_recipient_token": 0.7142857142857143, + "train_accuracy_recipient_token_all": 0.6, + "train_accuracy_recipient_token_all_total": 5, + "train_accuracy_recipient_token_get": 1.0, + "train_accuracy_recipient_token_get_total": 2, + "train_total_number_recipient_token": 7 + }, + { + "epoch": 1.0, + "grad_norm": 1.3267207098037004, "learning_rate": 0.0, - "loss": 0.1852, - "step": 337 + "loss": 0.1335, + "step": 506 }, { - "epoch": 0.9990118577075099, - "step": 337, - "total_flos": 337785957384192.0, - "train_loss": 0.2736482885984002, - "train_runtime": 6297.3, - "train_samples_per_second": 0.321, - "train_steps_per_second": 0.054 + "epoch": 1.0, + "step": 506, + "total_flos": 0.0, + "train_loss": 0.24046664490647937, + "train_runtime": 13405.0682, + "train_samples_per_second": 0.151, + "train_steps_per_second": 0.038 } ], "logging_steps": 1.0, - "max_steps": 337, + "max_steps": 506, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100.0, @@ -18701,7 +19909,7 @@ "attributes": {} } }, - "total_flos": 337785957384192.0, + "total_flos": 0.0, "train_batch_size": 1, "trial_name": null, "trial_params": null