{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 168, "global_step": 506, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0, "step": 0, "train_accuracy_recipient_token": 0.8, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_accuracy_recipient_token_set": 0.0, "train_accuracy_recipient_token_set_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0, "step": 0, "train_accuracy_recipient_token": 0.6111111111111112, "train_accuracy_recipient_token_all": 0.6, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_ident": 0.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.001976284584980237, "grad_norm": 2.82898798080985, "learning_rate": 3.125e-07, "loss": 0.2927, "step": 1 }, { "epoch": 0.001976284584980237, "step": 1, "train_accuracy_recipient_token": 0.5789473684210527, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_histor": 0.0, "train_accuracy_recipient_token_histor_total": 2, "train_accuracy_recipient_token_med": 0.5, "train_accuracy_recipient_token_med_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.001976284584980237, "step": 1, "train_accuracy_recipient_token": 0.65, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 0.5, "train_accuracy_recipient_token_estimate_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_p": 0.0, "train_accuracy_recipient_token_p_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.003952569169960474, "grad_norm": 2.796922585421834, "learning_rate": 6.25e-07, "loss": 0.3515, "step": 2 }, { "epoch": 0.003952569169960474, "step": 2, "train_accuracy_recipient_token": 0.5555555555555556, "train_accuracy_recipient_token_all": 0.6, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.3333333333333333, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_search": 0.5, "train_accuracy_recipient_token_search_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.003952569169960474, "step": 2, "train_accuracy_recipient_token": 0.6842105263157895, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_search": 0.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.005928853754940711, "grad_norm": 2.1200152439803803, "learning_rate": 9.375000000000001e-07, "loss": 0.1468, "step": 3 }, { "epoch": 0.005928853754940711, "step": 3, "train_accuracy_recipient_token": 0.7368421052631579, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_c": 0.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_get": 0.875, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.005928853754940711, "step": 3, "train_accuracy_recipient_token": 0.8, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 2, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_interpret": 0.6666666666666666, "train_accuracy_recipient_token_interpret_total": 3, "train_total_number_recipient_token": 20 }, { "epoch": 0.007905138339920948, "grad_norm": 3.190242133778613, "learning_rate": 1.25e-06, "loss": 0.2953, "step": 4 }, { "epoch": 0.007905138339920948, "step": 4, "train_accuracy_recipient_token": 0.7368421052631579, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_check": 0.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 19 }, { "epoch": 0.007905138339920948, "step": 4, "train_accuracy_recipient_token": 0.6111111111111112, "train_accuracy_recipient_token_all": 0.6, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 0.3333333333333333, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.009881422924901186, "grad_norm": 3.13829102322951, "learning_rate": 1.5625e-06, "loss": 0.2338, "step": 5 }, { "epoch": 0.009881422924901186, "step": 5, "train_accuracy_recipient_token": 0.631578947368421, "train_accuracy_recipient_token_all": 0.6, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_list": 0.0, "train_accuracy_recipient_token_list_total": 1, "train_accuracy_recipient_token_vendor": 1.0, "train_accuracy_recipient_token_vendor_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.009881422924901186, "step": 5, "train_accuracy_recipient_token": 0.7, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_get": 0.7142857142857143, "train_accuracy_recipient_token_get_total": 7, "train_accuracy_recipient_token_ident": 0.0, "train_accuracy_recipient_token_ident_total": 2, "train_total_number_recipient_token": 20 }, { "epoch": 0.011857707509881422, "grad_norm": 3.8032606174393555, "learning_rate": 1.8750000000000003e-06, "loss": 0.2693, "step": 6 }, { "epoch": 0.011857707509881422, "step": 6, "train_accuracy_recipient_token": 0.6, "train_accuracy_recipient_token_all": 0.5, "train_accuracy_recipient_token_all_total": 8, "train_accuracy_recipient_token_d": 0.0, "train_accuracy_recipient_token_d_total": 1, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 1, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_total_number_recipient_token": 15 }, { "epoch": 0.011857707509881422, "step": 6, "train_accuracy_recipient_token": 0.6666666666666666, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 2, "train_accuracy_recipient_token_get": 0.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_locate": 0.0, "train_accuracy_recipient_token_locate_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 15 }, { "epoch": 0.01383399209486166, "grad_norm": 4.597413898848223, "learning_rate": 2.1875000000000002e-06, "loss": 0.3544, "step": 7 }, { "epoch": 0.01383399209486166, "step": 7, "train_accuracy_recipient_token": 0.5, "train_accuracy_recipient_token_all": 0.4, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 0.5, "train_accuracy_recipient_token_analy_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_rate": 1.0, "train_accuracy_recipient_token_rate_total": 1, "train_accuracy_recipient_token_re": 0.0, "train_accuracy_recipient_token_re_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.01383399209486166, "step": 7, "train_accuracy_recipient_token": 0.5625, "train_accuracy_recipient_token_all": 0.6, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_fruit": 1.0, "train_accuracy_recipient_token_fruit_total": 3, "train_accuracy_recipient_token_ident": 0.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_select": 0.0, "train_accuracy_recipient_token_select_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.015810276679841896, "grad_norm": 2.9210762729436683, "learning_rate": 2.5e-06, "loss": 0.346, "step": 8 }, { "epoch": 0.015810276679841896, "step": 8, "train_accuracy_recipient_token": 0.6666666666666666, "train_accuracy_recipient_token_all": 0.6, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_accuracy_recipient_token_ident": 0.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.015810276679841896, "step": 8, "train_accuracy_recipient_token": 0.7368421052631579, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compose": 0.0, "train_accuracy_recipient_token_compose_total": 1, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 1, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_g": 1.0, "train_accuracy_recipient_token_g_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.017786561264822136, "grad_norm": 2.4385850326725835, "learning_rate": 2.8125e-06, "loss": 0.2706, "step": 9 }, { "epoch": 0.017786561264822136, "step": 9, "train_accuracy_recipient_token": 0.6111111111111112, "train_accuracy_recipient_token_all": 0.6363636363636364, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_validate": 0.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.017786561264822136, "step": 9, "train_accuracy_recipient_token": 0.4117647058823529, "train_accuracy_recipient_token_all": 0.5454545454545454, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_compare": 0.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 0.3333333333333333, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_ident": 0.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_validate": 0.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.019762845849802372, "grad_norm": 2.738667329727624, "learning_rate": 3.125e-06, "loss": 0.3037, "step": 10 }, { "epoch": 0.019762845849802372, "step": 10, "train_accuracy_recipient_token": 0.625, "train_accuracy_recipient_token_all": 0.75, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_evaluate": 0.0, "train_accuracy_recipient_token_evaluate_total": 1, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.019762845849802372, "step": 10, "train_accuracy_recipient_token": 0.7222222222222222, "train_accuracy_recipient_token_all": 0.6, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { "epoch": 0.021739130434782608, "grad_norm": 3.164747479850259, "learning_rate": 3.4375e-06, "loss": 0.4035, "step": 11 }, { "epoch": 0.021739130434782608, "step": 11, "train_accuracy_recipient_token": 0.625, "train_accuracy_recipient_token_all": 0.5, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 16 }, { "epoch": 0.021739130434782608, "step": 11, "train_accuracy_recipient_token": 0.7894736842105263, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.7777777777777778, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.023715415019762844, "grad_norm": 2.361065524720244, "learning_rate": 3.7500000000000005e-06, "loss": 0.3349, "step": 12 }, { "epoch": 0.023715415019762844, "step": 12, "train_accuracy_recipient_token": 0.6666666666666666, "train_accuracy_recipient_token_all": 0.6, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_ident": 0.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.023715415019762844, "step": 12, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.025691699604743084, "grad_norm": 2.0061236591150666, "learning_rate": 4.0625000000000005e-06, "loss": 0.2121, "step": 13 }, { "epoch": 0.025691699604743084, "step": 13, "train_accuracy_recipient_token": 0.7368421052631579, "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_describe": 1.0, "train_accuracy_recipient_token_describe_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_sum": 0.5, "train_accuracy_recipient_token_sum_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.025691699604743084, "step": 13, "train_accuracy_recipient_token": 0.6666666666666666, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_p": 1.0, "train_accuracy_recipient_token_p_total": 1, "train_accuracy_recipient_token_schedule": 0.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.02766798418972332, "grad_norm": 2.7988165456067735, "learning_rate": 4.3750000000000005e-06, "loss": 0.2798, "step": 14 }, { "epoch": 0.02766798418972332, "step": 14, "train_accuracy_recipient_token": 0.625, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 0.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_load": 1.0, "train_accuracy_recipient_token_load_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.02766798418972332, "step": 14, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_property": 1.0, "train_accuracy_recipient_token_property_total": 3, "train_accuracy_recipient_token_real": 1.0, "train_accuracy_recipient_token_real_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.029644268774703556, "grad_norm": 2.4311181509760647, "learning_rate": 4.6875000000000004e-06, "loss": 0.2332, "step": 15 }, { "epoch": 0.029644268774703556, "step": 15, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_as": 1.0, "train_accuracy_recipient_token_as_total": 1, "train_accuracy_recipient_token_clean": 1.0, "train_accuracy_recipient_token_clean_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_rep": 0.5, "train_accuracy_recipient_token_rep_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.029644268774703556, "step": 15, "train_accuracy_recipient_token": 0.7368421052631579, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 0.75, "train_accuracy_recipient_token_compare_total": 4, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_accuracy_recipient_token_search": 0.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.03162055335968379, "grad_norm": 2.6718500710390014, "learning_rate": 5e-06, "loss": 0.2296, "step": 16 }, { "epoch": 0.03162055335968379, "step": 16, "train_accuracy_recipient_token": 0.7222222222222222, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 0.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_s": 0.5, "train_accuracy_recipient_token_s_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.03162055335968379, "step": 16, "train_accuracy_recipient_token": 0.7894736842105263, "train_accuracy_recipient_token_add": 0.8, "train_accuracy_recipient_token_add_total": 5, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 0.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.03359683794466403, "grad_norm": 2.4193634654326193, "learning_rate": 4.999948617395916e-06, "loss": 0.2744, "step": 17 }, { "epoch": 0.03359683794466403, "step": 17, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_h": 1.0, "train_accuracy_recipient_token_h_total": 3, "train_total_number_recipient_token": 20 }, { "epoch": 0.03359683794466403, "step": 17, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_f": 1.0, "train_accuracy_recipient_token_f_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.03557312252964427, "grad_norm": 1.9915037614380675, "learning_rate": 4.9997944716957985e-06, "loss": 0.2374, "step": 18 }, { "epoch": 0.03557312252964427, "step": 18, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 5, "train_accuracy_recipient_token_find": 0.75, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 10 }, { "epoch": 0.03557312252964427, "step": 18, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 2, "train_accuracy_recipient_token_ex": 1.0, "train_accuracy_recipient_token_ex_total": 1, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_translate": 0.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.037549407114624504, "grad_norm": 13.63293630478699, "learning_rate": 4.999537569235975e-06, "loss": 0.575, "step": 19 }, { "epoch": 0.037549407114624504, "step": 19, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 0.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.037549407114624504, "step": 19, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_s": 0.0, "train_accuracy_recipient_token_s_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.039525691699604744, "grad_norm": 2.7743056709907594, "learning_rate": 4.9991779205767e-06, "loss": 0.2632, "step": 20 }, { "epoch": 0.039525691699604744, "step": 20, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_verify": 1.0, "train_accuracy_recipient_token_verify_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.039525691699604744, "step": 20, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_track": 1.0, "train_accuracy_recipient_token_track_total": 1, "train_accuracy_recipient_token_validate": 0.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.041501976284584984, "grad_norm": 2.867128204261161, "learning_rate": 4.99871554050172e-06, "loss": 0.2, "step": 21 }, { "epoch": 0.041501976284584984, "step": 21, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 0.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 18 }, { "epoch": 0.041501976284584984, "step": 21, "train_accuracy_recipient_token": 0.7058823529411765, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_buy": 0.0, "train_accuracy_recipient_token_buy_total": 1, "train_accuracy_recipient_token_exp": 0.0, "train_accuracy_recipient_token_exp_total": 1, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 17 }, { "epoch": 0.043478260869565216, "grad_norm": 2.8424629908120873, "learning_rate": 4.99815044801767e-06, "loss": 0.3206, "step": 22 }, { "epoch": 0.043478260869565216, "step": 22, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_filter": 0.0, "train_accuracy_recipient_token_filter_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_locate": 1.0, "train_accuracy_recipient_token_locate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.043478260869565216, "step": 22, "train_accuracy_recipient_token": 0.7894736842105263, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_clean": 1.0, "train_accuracy_recipient_token_clean_total": 1, "train_accuracy_recipient_token_generate": 0.6, "train_accuracy_recipient_token_generate_total": 5, "train_accuracy_recipient_token_remove": 0.0, "train_accuracy_recipient_token_remove_total": 1, "train_accuracy_recipient_token_replace": 1.0, "train_accuracy_recipient_token_replace_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.045454545454545456, "grad_norm": 3.269809875569231, "learning_rate": 4.997482666353287e-06, "loss": 0.2165, "step": 23 }, { "epoch": 0.045454545454545456, "step": 23, "train_accuracy_recipient_token": 0.6470588235294118, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_ident": 0.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_register": 0.0, "train_accuracy_recipient_token_register_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.045454545454545456, "step": 23, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_s": 1.0, "train_accuracy_recipient_token_s_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.04743083003952569, "grad_norm": 2.399757318906346, "learning_rate": 4.9967122229584614e-06, "loss": 0.1488, "step": 24 }, { "epoch": 0.04743083003952569, "step": 24, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 0.8, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.04743083003952569, "step": 24, "train_accuracy_recipient_token": 0.7894736842105263, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_ident": 0.3333333333333333, "train_accuracy_recipient_token_ident_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.04940711462450593, "grad_norm": 1.7504463644259465, "learning_rate": 4.995839149503103e-06, "loss": 0.1377, "step": 25 }, { "epoch": 0.04940711462450593, "step": 25, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_histor": 1.0, "train_accuracy_recipient_token_histor_total": 1, "train_accuracy_recipient_token_math": 1.0, "train_accuracy_recipient_token_math_total": 1, "train_accuracy_recipient_token_symbol": 1.0, "train_accuracy_recipient_token_symbol_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.04940711462450593, "step": 25, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 18 }, { "epoch": 0.05138339920948617, "grad_norm": 2.7693470356484107, "learning_rate": 4.994863481875842e-06, "loss": 0.2731, "step": 26 }, { "epoch": 0.05138339920948617, "step": 26, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.05138339920948617, "step": 26, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_nut": 1.0, "train_accuracy_recipient_token_nut_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.0533596837944664, "grad_norm": 2.9947438603207726, "learning_rate": 4.993785260182552e-06, "loss": 0.2053, "step": 27 }, { "epoch": 0.0533596837944664, "step": 27, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_building": 1.0, "train_accuracy_recipient_token_building_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_property": 1.0, "train_accuracy_recipient_token_property_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.0533596837944664, "step": 27, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.05533596837944664, "grad_norm": 2.8975618558727008, "learning_rate": 4.992604528744705e-06, "loss": 0.2712, "step": 28 }, { "epoch": 0.05533596837944664, "step": 28, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { "epoch": 0.05533596837944664, "step": 28, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_plan": 0.0, "train_accuracy_recipient_token_plan_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.05731225296442688, "grad_norm": 2.5657098582780398, "learning_rate": 4.991321336097546e-06, "loss": 0.2796, "step": 29 }, { "epoch": 0.05731225296442688, "step": 29, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.05731225296442688, "step": 29, "train_accuracy_recipient_token": 0.85, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_traffic": 1.0, "train_accuracy_recipient_token_traffic_total": 2, "train_accuracy_recipient_token_verify": 1.0, "train_accuracy_recipient_token_verify_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.05928853754940711, "grad_norm": 2.7207547834543275, "learning_rate": 4.989935734988098e-06, "loss": 0.2106, "step": 30 }, { "epoch": 0.05928853754940711, "step": 30, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_accuracy_recipient_token_s": 1.0, "train_accuracy_recipient_token_s_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.05928853754940711, "step": 30, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.06126482213438735, "grad_norm": 2.0609870060002202, "learning_rate": 4.988447782372996e-06, "loss": 0.2457, "step": 31 }, { "epoch": 0.06126482213438735, "step": 31, "train_accuracy_recipient_token": 0.6875, "train_accuracy_recipient_token_all": 0.5454545454545454, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_deploy": 1.0, "train_accuracy_recipient_token_deploy_total": 1, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_install": 1.0, "train_accuracy_recipient_token_install_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.06126482213438735, "step": 31, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_accuracy_recipient_token_p": 1.0, "train_accuracy_recipient_token_p_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.06324110671936758, "grad_norm": 2.6519364596617034, "learning_rate": 4.986857539416144e-06, "loss": 0.2742, "step": 32 }, { "epoch": 0.06324110671936758, "step": 32, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.06324110671936758, "step": 32, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.06521739130434782, "grad_norm": 3.3761072923479434, "learning_rate": 4.985165071486201e-06, "loss": 0.3322, "step": 33 }, { "epoch": 0.06521739130434782, "step": 33, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_style": 1.0, "train_accuracy_recipient_token_style_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.06521739130434782, "step": 33, "train_accuracy_recipient_token": 0.7333333333333333, "train_accuracy_recipient_token_all": 0.6, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 15 }, { "epoch": 0.06719367588932806, "grad_norm": 3.2773278792649094, "learning_rate": 4.983370448153896e-06, "loss": 0.4958, "step": 34 }, { "epoch": 0.06719367588932806, "step": 34, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_list": 0.0, "train_accuracy_recipient_token_list_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.06719367588932806, "step": 34, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 17 }, { "epoch": 0.0691699604743083, "grad_norm": 2.1733891801745626, "learning_rate": 4.981473743189163e-06, "loss": 0.1996, "step": 35 }, { "epoch": 0.0691699604743083, "step": 35, "train_accuracy_recipient_token": 0.75, "train_accuracy_recipient_token_all": 0.6363636363636364, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_track": 1.0, "train_accuracy_recipient_token_track_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.0691699604743083, "step": 35, "train_accuracy_recipient_token": 0.8125, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 16 }, { "epoch": 0.07114624505928854, "grad_norm": 2.8245080994396585, "learning_rate": 4.979475034558115e-06, "loss": 0.2105, "step": 36 }, { "epoch": 0.07114624505928854, "step": 36, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.07114624505928854, "step": 36, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.07312252964426877, "grad_norm": 1.6572734803173075, "learning_rate": 4.977374404419838e-06, "loss": 0.2089, "step": 37 }, { "epoch": 0.07312252964426877, "step": 37, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 1, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_accuracy_recipient_token_rate": 1.0, "train_accuracy_recipient_token_rate_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.07312252964426877, "step": 37, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_get": 0.7142857142857143, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { "epoch": 0.07509881422924901, "grad_norm": 2.053920013378982, "learning_rate": 4.9751719391230055e-06, "loss": 0.1193, "step": 38 }, { "epoch": 0.07509881422924901, "step": 38, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.875, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.07509881422924901, "step": 38, "train_accuracy_recipient_token": 0.7777777777777778, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_plan": 0.0, "train_accuracy_recipient_token_plan_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.07707509881422925, "grad_norm": 1.2542268957003366, "learning_rate": 4.9728677292023405e-06, "loss": 0.1507, "step": 39 }, { "epoch": 0.07707509881422925, "step": 39, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_animal": 0.0, "train_accuracy_recipient_token_animal_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_im": 1.0, "train_accuracy_recipient_token_im_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.07707509881422925, "step": 39, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_plan": 0.0, "train_accuracy_recipient_token_plan_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.07905138339920949, "grad_norm": 1.7847434934616466, "learning_rate": 4.97046186937489e-06, "loss": 0.1873, "step": 40 }, { "epoch": 0.07905138339920949, "step": 40, "train_accuracy_recipient_token": 0.8571428571428571, "train_accuracy_recipient_token_all": 0.9166666666666666, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 8, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 21 }, { "epoch": 0.07905138339920949, "step": 40, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.08102766798418973, "grad_norm": 2.2812627178319715, "learning_rate": 4.967954458536126e-06, "loss": 0.3651, "step": 41 }, { "epoch": 0.08102766798418973, "step": 41, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_draw": 0.8, "train_accuracy_recipient_token_draw_total": 5, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.08102766798418973, "step": 41, "train_accuracy_recipient_token": 0.8, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 20 }, { "epoch": 0.08300395256916997, "grad_norm": 2.184012199227446, "learning_rate": 4.965345599755888e-06, "loss": 0.2094, "step": 42 }, { "epoch": 0.08300395256916997, "step": 42, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.7692307692307693, "train_accuracy_recipient_token_all_total": 13, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.08300395256916997, "step": 42, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.08498023715415019, "grad_norm": 2.2059440276239743, "learning_rate": 4.9626354002741424e-06, "loss": 0.2288, "step": 43 }, { "epoch": 0.08498023715415019, "step": 43, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_rec": 1.0, "train_accuracy_recipient_token_rec_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.08498023715415019, "step": 43, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 18 }, { "epoch": 0.08695652173913043, "grad_norm": 2.04652434683755, "learning_rate": 4.959823971496575e-06, "loss": 0.2356, "step": 44 }, { "epoch": 0.08695652173913043, "step": 44, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.08695652173913043, "step": 44, "train_accuracy_recipient_token": 0.8666666666666667, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_im": 1.0, "train_accuracy_recipient_token_im_total": 1, "train_accuracy_recipient_token_install": 0.0, "train_accuracy_recipient_token_install_total": 1, "train_total_number_recipient_token": 15 }, { "epoch": 0.08893280632411067, "grad_norm": 2.367283225731573, "learning_rate": 4.95691142899001e-06, "loss": 0.2289, "step": 45 }, { "epoch": 0.08893280632411067, "step": 45, "train_accuracy_recipient_token": 0.8125, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_event": 1.0, "train_accuracy_recipient_token_event_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 16 }, { "epoch": 0.08893280632411067, "step": 45, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.75, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.09090909090909091, "grad_norm": 2.580407364292025, "learning_rate": 4.953897892477664e-06, "loss": 0.4122, "step": 46 }, { "epoch": 0.09090909090909091, "step": 46, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_hire": 1.0, "train_accuracy_recipient_token_hire_total": 1, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.09090909090909091, "step": 46, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_convert": 0.8, "train_accuracy_recipient_token_convert_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.09288537549407115, "grad_norm": 2.516544751703755, "learning_rate": 4.950783485834218e-06, "loss": 0.2026, "step": 47 }, { "epoch": 0.09288537549407115, "step": 47, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_evaluate": 1.0, "train_accuracy_recipient_token_evaluate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_stat": 1.0, "train_accuracy_recipient_token_stat_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.09288537549407115, "step": 47, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_filter": 0.8, "train_accuracy_recipient_token_filter_total": 5, "train_accuracy_recipient_token_food": 1.0, "train_accuracy_recipient_token_food_total": 4, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.09486166007905138, "grad_norm": 2.1047249182046577, "learning_rate": 4.947568337080733e-06, "loss": 0.2317, "step": 48 }, { "epoch": 0.09486166007905138, "step": 48, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 0.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.09486166007905138, "step": 48, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.09683794466403162, "grad_norm": 2.1380480543246896, "learning_rate": 4.944252578379379e-06, "loss": 0.2049, "step": 49 }, { "epoch": 0.09683794466403162, "step": 49, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_s": 1.0, "train_accuracy_recipient_token_s_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.09683794466403162, "step": 49, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { "epoch": 0.09881422924901186, "grad_norm": 1.887568271122836, "learning_rate": 4.940836346028011e-06, "loss": 0.3046, "step": 50 }, { "epoch": 0.09881422924901186, "step": 50, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_assert": 0.8, "train_accuracy_recipient_token_assert_total": 5, "train_accuracy_recipient_token_describe": 1.0, "train_accuracy_recipient_token_describe_total": 1, "train_accuracy_recipient_token_mount": 1.0, "train_accuracy_recipient_token_mount_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.09881422924901186, "step": 50, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.8333333333333334, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { "epoch": 0.1007905138339921, "grad_norm": 1.5031262286748293, "learning_rate": 4.937319780454559e-06, "loss": 0.2287, "step": 51 }, { "epoch": 0.1007905138339921, "step": 51, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.1007905138339921, "step": 51, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.10276679841897234, "grad_norm": 2.1419992157156957, "learning_rate": 4.933703026211262e-06, "loss": 0.2536, "step": 52 }, { "epoch": 0.10276679841897234, "step": 52, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.10276679841897234, "step": 52, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.10474308300395258, "grad_norm": 2.61662813745291, "learning_rate": 4.92998623196872e-06, "loss": 0.2495, "step": 53 }, { "epoch": 0.10474308300395258, "step": 53, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.9, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 20 }, { "epoch": 0.10474308300395258, "step": 53, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 3, "train_accuracy_recipient_token_ins": 1.0, "train_accuracy_recipient_token_ins_total": 1, "train_accuracy_recipient_token_rep": 1.0, "train_accuracy_recipient_token_rep_total": 1, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.1067193675889328, "grad_norm": 2.206222398671073, "learning_rate": 4.926169550509787e-06, "loss": 0.2654, "step": 54 }, { "epoch": 0.1067193675889328, "step": 54, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 20 }, { "epoch": 0.1067193675889328, "step": 54, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.10869565217391304, "grad_norm": 1.7197195903001685, "learning_rate": 4.9222531387232885e-06, "loss": 0.2115, "step": 55 }, { "epoch": 0.10869565217391304, "step": 55, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 0.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.10869565217391304, "step": 55, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.11067193675889328, "grad_norm": 2.1768195760337314, "learning_rate": 4.918237157597574e-06, "loss": 0.2403, "step": 56 }, { "epoch": 0.11067193675889328, "step": 56, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_track": 1.0, "train_accuracy_recipient_token_track_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.11067193675889328, "step": 56, "train_accuracy_recipient_token": 0.9523809523809523, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 21 }, { "epoch": 0.11264822134387352, "grad_norm": 1.969156469443854, "learning_rate": 4.914121772213898e-06, "loss": 0.28, "step": 57 }, { "epoch": 0.11264822134387352, "step": 57, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_log": 1.0, "train_accuracy_recipient_token_log_total": 1, "train_accuracy_recipient_token_plant": 1.0, "train_accuracy_recipient_token_plant_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.11264822134387352, "step": 57, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_plan": 1.0, "train_accuracy_recipient_token_plan_total": 1, "train_accuracy_recipient_token_urban": 1.0, "train_accuracy_recipient_token_urban_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.11462450592885376, "grad_norm": 1.8351868382101044, "learning_rate": 4.909907151739634e-06, "loss": 0.2386, "step": 58 }, { "epoch": 0.11462450592885376, "step": 58, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_upgrade": 1.0, "train_accuracy_recipient_token_upgrade_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.11462450592885376, "step": 58, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 17 }, { "epoch": 0.116600790513834, "grad_norm": 3.0727021394959544, "learning_rate": 4.905593469421323e-06, "loss": 0.3147, "step": 59 }, { "epoch": 0.116600790513834, "step": 59, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 2, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.116600790513834, "step": 59, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_track": 1.0, "train_accuracy_recipient_token_track_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.11857707509881422, "grad_norm": 2.181993064766873, "learning_rate": 4.901180902577549e-06, "loss": 0.2224, "step": 60 }, { "epoch": 0.11857707509881422, "step": 60, "train_accuracy_recipient_token": 0.85, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 2, "train_accuracy_recipient_token_translate": 0.5, "train_accuracy_recipient_token_translate_total": 2, "train_total_number_recipient_token": 20 }, { "epoch": 0.11857707509881422, "step": 60, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.9, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 20 }, { "epoch": 0.12055335968379446, "grad_norm": 2.1374748114773303, "learning_rate": 4.896669632591652e-06, "loss": 0.1559, "step": 61 }, { "epoch": 0.12055335968379446, "step": 61, "train_accuracy_recipient_token": 0.8, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 15, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_setup": 0.0, "train_accuracy_recipient_token_setup_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.12055335968379446, "step": 61, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 0.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.1225296442687747, "grad_norm": 2.467435650737327, "learning_rate": 4.892059844904273e-06, "loss": 0.2946, "step": 62 }, { "epoch": 0.1225296442687747, "step": 62, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 1, "train_accuracy_recipient_token_get": 0.875, "train_accuracy_recipient_token_get_total": 8, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.1225296442687747, "step": 62, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 1, "train_accuracy_recipient_token_merge": 1.0, "train_accuracy_recipient_token_merge_total": 1, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.12450592885375494, "grad_norm": 3.655476447614754, "learning_rate": 4.8873517290057265e-06, "loss": 0.2478, "step": 63 }, { "epoch": 0.12450592885375494, "step": 63, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compile": 1.0, "train_accuracy_recipient_token_compile_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.12450592885375494, "step": 63, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 2, "train_total_number_recipient_token": 16 }, { "epoch": 0.12648221343873517, "grad_norm": 2.202751393561073, "learning_rate": 4.882545478428219e-06, "loss": 0.2624, "step": 64 }, { "epoch": 0.12648221343873517, "step": 64, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.12648221343873517, "step": 64, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 16 }, { "epoch": 0.12845849802371542, "grad_norm": 2.26174226561165, "learning_rate": 4.8776412907378845e-06, "loss": 0.3378, "step": 65 }, { "epoch": 0.12845849802371542, "step": 65, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9230769230769231, "train_accuracy_recipient_token_all_total": 13, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.12845849802371542, "step": 65, "train_accuracy_recipient_token": 0.8666666666666667, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 15 }, { "epoch": 0.13043478260869565, "grad_norm": 2.6607823334587644, "learning_rate": 4.872639367526672e-06, "loss": 0.3342, "step": 66 }, { "epoch": 0.13043478260869565, "step": 66, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.13043478260869565, "step": 66, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 3, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 20 }, { "epoch": 0.1324110671936759, "grad_norm": 2.3676089848491957, "learning_rate": 4.8675399144040535e-06, "loss": 0.4039, "step": 67 }, { "epoch": 0.1324110671936759, "step": 67, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_build": 1.0, "train_accuracy_recipient_token_build_total": 1, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 4, "train_accuracy_recipient_token_fire": 1.0, "train_accuracy_recipient_token_fire_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.1324110671936759, "step": 67, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 0.5, "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_find": 0.8, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.13438735177865613, "grad_norm": 2.1673908938305124, "learning_rate": 4.862343140988573e-06, "loss": 0.1757, "step": 68 }, { "epoch": 0.13438735177865613, "step": 68, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_ex": 0.6666666666666666, "train_accuracy_recipient_token_ex_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.13438735177865613, "step": 68, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_log": 1.0, "train_accuracy_recipient_token_log_total": 1, "train_accuracy_recipient_token_review": 1.0, "train_accuracy_recipient_token_review_total": 2, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.13636363636363635, "grad_norm": 1.8900015178295946, "learning_rate": 4.857049260899233e-06, "loss": 0.2066, "step": 69 }, { "epoch": 0.13636363636363635, "step": 69, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_filter": 1.0, "train_accuracy_recipient_token_filter_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.13636363636363635, "step": 69, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.1383399209486166, "grad_norm": 2.536038040555671, "learning_rate": 4.851658491746707e-06, "loss": 0.2677, "step": 70 }, { "epoch": 0.1383399209486166, "step": 70, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_city": 1.0, "train_accuracy_recipient_token_city_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.1383399209486166, "step": 70, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { "epoch": 0.14031620553359683, "grad_norm": 1.7887214738919732, "learning_rate": 4.846171055124401e-06, "loss": 0.1541, "step": 71 }, { "epoch": 0.14031620553359683, "step": 71, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 0.5, "train_accuracy_recipient_token_analy_total": 2, "train_accuracy_recipient_token_detect": 1.0, "train_accuracy_recipient_token_detect_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_ident": 0.5, "train_accuracy_recipient_token_ident_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.14031620553359683, "step": 71, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.8333333333333334, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_calculate": 0.5, "train_accuracy_recipient_token_calculate_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.1422924901185771, "grad_norm": 3.415062910528756, "learning_rate": 4.8405871765993435e-06, "loss": 0.3752, "step": 72 }, { "epoch": 0.1422924901185771, "step": 72, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_organ": 1.0, "train_accuracy_recipient_token_organ_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.1422924901185771, "step": 72, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 2, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_d": 1.0, "train_accuracy_recipient_token_d_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.1442687747035573, "grad_norm": 2.100833710936171, "learning_rate": 4.834907085702909e-06, "loss": 0.2565, "step": 73 }, { "epoch": 0.1442687747035573, "step": 73, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 2, "train_total_number_recipient_token": 20 }, { "epoch": 0.1442687747035573, "step": 73, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.875, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 20 }, { "epoch": 0.14624505928853754, "grad_norm": 1.8239921154406062, "learning_rate": 4.829131015921386e-06, "loss": 0.1924, "step": 74 }, { "epoch": 0.14624505928853754, "step": 74, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 8, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_setup": 0.0, "train_accuracy_recipient_token_setup_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.14624505928853754, "step": 74, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_g": 1.0, "train_accuracy_recipient_token_g_total": 3, "train_accuracy_recipient_token_track": 1.0, "train_accuracy_recipient_token_track_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.1482213438735178, "grad_norm": 2.1914980081973794, "learning_rate": 4.82325920468638e-06, "loss": 0.1502, "step": 75 }, { "epoch": 0.1482213438735178, "step": 75, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_report": 1.0, "train_accuracy_recipient_token_report_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.1482213438735178, "step": 75, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 16 }, { "epoch": 0.15019762845849802, "grad_norm": 2.315784596207221, "learning_rate": 4.817291893365055e-06, "loss": 0.1811, "step": 76 }, { "epoch": 0.15019762845849802, "step": 76, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 0.5, "train_accuracy_recipient_token_analy_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_predict": 1.0, "train_accuracy_recipient_token_predict_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.15019762845849802, "step": 76, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_plan": 0.0, "train_accuracy_recipient_token_plan_total": 1, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.15217391304347827, "grad_norm": 3.9917762289000542, "learning_rate": 4.811229327250204e-06, "loss": 0.2349, "step": 77 }, { "epoch": 0.15217391304347827, "step": 77, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_max": 1.0, "train_accuracy_recipient_token_max_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.15217391304347827, "step": 77, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.1541501976284585, "grad_norm": 2.2903178058367315, "learning_rate": 4.805071755550177e-06, "loss": 0.3338, "step": 78 }, { "epoch": 0.1541501976284585, "step": 78, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.1541501976284585, "step": 78, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_evaluate": 1.0, "train_accuracy_recipient_token_evaluate_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_im": 1.0, "train_accuracy_recipient_token_im_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.15612648221343872, "grad_norm": 2.526628452772315, "learning_rate": 4.7988194313786275e-06, "loss": 0.2483, "step": 79 }, { "epoch": 0.15612648221343872, "step": 79, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_display": 1.0, "train_accuracy_recipient_token_display_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_h": 1.0, "train_accuracy_recipient_token_h_total": 3, "train_accuracy_recipient_token_s": 1.0, "train_accuracy_recipient_token_s_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.15612648221343872, "step": 79, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_calculate": 0.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_check": 0.5, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.15810276679841898, "grad_norm": 1.3952343413076715, "learning_rate": 4.7924726117441135e-06, "loss": 0.1396, "step": 80 }, { "epoch": 0.15810276679841898, "step": 80, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.15810276679841898, "step": 80, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.1600790513833992, "grad_norm": 1.8632305360852095, "learning_rate": 4.786031557539532e-06, "loss": 0.2135, "step": 81 }, { "epoch": 0.1600790513833992, "step": 81, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_flag": 0.6666666666666666, "train_accuracy_recipient_token_flag_total": 3, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 18 }, { "epoch": 0.1600790513833992, "step": 81, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { "epoch": 0.16205533596837945, "grad_norm": 2.2968333669223, "learning_rate": 4.779496533531393e-06, "loss": 0.3168, "step": 82 }, { "epoch": 0.16205533596837945, "step": 82, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_log": 0.0, "train_accuracy_recipient_token_log_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.16205533596837945, "step": 82, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 1, "train_accuracy_recipient_token_draw": 1.0, "train_accuracy_recipient_token_draw_total": 1, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.16403162055335968, "grad_norm": 2.249381257252995, "learning_rate": 4.7728678083489375e-06, "loss": 0.3738, "step": 83 }, { "epoch": 0.16403162055335968, "step": 83, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_build": 0.5, "train_accuracy_recipient_token_build_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_paint": 1.0, "train_accuracy_recipient_token_paint_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.16403162055335968, "step": 83, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_fetch": 0.0, "train_accuracy_recipient_token_fetch_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { "epoch": 0.16600790513833993, "grad_norm": 1.8402947400679557, "learning_rate": 4.766145654473096e-06, "loss": 0.1812, "step": 84 }, { "epoch": 0.16600790513833993, "step": 84, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_monitor": 0.0, "train_accuracy_recipient_token_monitor_total": 1, "train_accuracy_recipient_token_remove": 1.0, "train_accuracy_recipient_token_remove_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.16600790513833993, "step": 84, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 1, "train_accuracy_recipient_token_ex": 1.0, "train_accuracy_recipient_token_ex_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_number": 1.0, "train_accuracy_recipient_token_number_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.16798418972332016, "grad_norm": 1.9475841569891188, "learning_rate": 4.7593303482252835e-06, "loss": 0.2456, "step": 85 }, { "epoch": 0.16798418972332016, "step": 85, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 20 }, { "epoch": 0.16798418972332016, "step": 85, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_hire": 1.0, "train_accuracy_recipient_token_hire_total": 1, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.16996047430830039, "grad_norm": 2.2098134551670943, "learning_rate": 4.752422169756048e-06, "loss": 0.2644, "step": 86 }, { "epoch": 0.16996047430830039, "step": 86, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.16996047430830039, "step": 86, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_ex": 1.0, "train_accuracy_recipient_token_ex_total": 3, "train_accuracy_recipient_token_format": 0.0, "train_accuracy_recipient_token_format_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.17193675889328064, "grad_norm": 1.9178477232600728, "learning_rate": 4.745421403033548e-06, "loss": 0.2145, "step": 87 }, { "epoch": 0.17193675889328064, "step": 87, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.9, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 20 }, { "epoch": 0.17193675889328064, "step": 87, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_f": 1.0, "train_accuracy_recipient_token_f_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 1, "train_accuracy_recipient_token_window": 1.0, "train_accuracy_recipient_token_window_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.17391304347826086, "grad_norm": 1.4953922375484354, "learning_rate": 4.738328335831883e-06, "loss": 0.1738, "step": 88 }, { "epoch": 0.17391304347826086, "step": 88, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_arch": 1.0, "train_accuracy_recipient_token_arch_total": 1, "train_accuracy_recipient_token_building": 1.0, "train_accuracy_recipient_token_building_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 18 }, { "epoch": 0.17391304347826086, "step": 88, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_detect": 1.0, "train_accuracy_recipient_token_detect_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_tr": 1.0, "train_accuracy_recipient_token_tr_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.17588932806324112, "grad_norm": 1.9501250918192976, "learning_rate": 4.7311432597192655e-06, "loss": 0.224, "step": 89 }, { "epoch": 0.17588932806324112, "step": 89, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.17588932806324112, "step": 89, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 20 }, { "epoch": 0.17786561264822134, "grad_norm": 1.7132307123231727, "learning_rate": 4.72386647004603e-06, "loss": 0.1349, "step": 90 }, { "epoch": 0.17786561264822134, "step": 90, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_door": 1.0, "train_accuracy_recipient_token_door_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_invest": 1.0, "train_accuracy_recipient_token_invest_total": 1, "train_accuracy_recipient_token_trans": 1.0, "train_accuracy_recipient_token_trans_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.17786561264822134, "step": 90, "train_accuracy_recipient_token": 0.8125, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 16 }, { "epoch": 0.17984189723320157, "grad_norm": 1.862599179005162, "learning_rate": 4.716498265932501e-06, "loss": 0.2707, "step": 91 }, { "epoch": 0.17984189723320157, "step": 91, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 2, "train_total_number_recipient_token": 20 }, { "epoch": 0.17984189723320157, "step": 91, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_animal": 1.0, "train_accuracy_recipient_token_animal_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_store": 1.0, "train_accuracy_recipient_token_store_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.18181818181818182, "grad_norm": 1.5411465996887044, "learning_rate": 4.7090389502566884e-06, "loss": 0.1157, "step": 92 }, { "epoch": 0.18181818181818182, "step": 92, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_plant": 1.0, "train_accuracy_recipient_token_plant_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.18181818181818182, "step": 92, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 17 }, { "epoch": 0.18379446640316205, "grad_norm": 2.100818863534559, "learning_rate": 4.701488829641845e-06, "loss": 0.2582, "step": 93 }, { "epoch": 0.18379446640316205, "step": 93, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.7777777777777778, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 16 }, { "epoch": 0.18379446640316205, "step": 93, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_par": 1.0, "train_accuracy_recipient_token_par_total": 2, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.1857707509881423, "grad_norm": 2.5798405707090852, "learning_rate": 4.693848214443858e-06, "loss": 0.4023, "step": 94 }, { "epoch": 0.1857707509881423, "step": 94, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_flower": 0.6666666666666666, "train_accuracy_recipient_token_flower_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.1857707509881423, "step": 94, "train_accuracy_recipient_token": 0.7647058823529411, "train_accuracy_recipient_token_all": 0.6666666666666666, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 2, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.18774703557312253, "grad_norm": 2.1919145048059283, "learning_rate": 4.686117418738489e-06, "loss": 0.2804, "step": 95 }, { "epoch": 0.18774703557312253, "step": 95, "train_accuracy_recipient_token": 0.7894736842105263, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_em": 1.0, "train_accuracy_recipient_token_em_total": 1, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_first": 1.0, "train_accuracy_recipient_token_first_total": 1, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 19 }, { "epoch": 0.18774703557312253, "step": 95, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_ex": 1.0, "train_accuracy_recipient_token_ex_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.18972332015810275, "grad_norm": 2.276698142449688, "learning_rate": 4.678296760308474e-06, "loss": 0.2569, "step": 96 }, { "epoch": 0.18972332015810275, "step": 96, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { "epoch": 0.18972332015810275, "step": 96, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.191699604743083, "grad_norm": 2.2258003018865185, "learning_rate": 4.670386560630446e-06, "loss": 0.2672, "step": 97 }, { "epoch": 0.191699604743083, "step": 97, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { "epoch": 0.191699604743083, "step": 97, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_gr": 1.0, "train_accuracy_recipient_token_gr_total": 5, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.19367588932806323, "grad_norm": 2.1162036419771804, "learning_rate": 4.6623871448617345e-06, "loss": 0.2356, "step": 98 }, { "epoch": 0.19367588932806323, "step": 98, "train_accuracy_recipient_token": 0.9333333333333333, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_license": 1.0, "train_accuracy_recipient_token_license_total": 1, "train_total_number_recipient_token": 15 }, { "epoch": 0.19367588932806323, "step": 98, "train_accuracy_recipient_token": 0.7777777777777778, "train_accuracy_recipient_token_all": 0.6666666666666666, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_compose": 1.0, "train_accuracy_recipient_token_compose_total": 4, "train_accuracy_recipient_token_crop": 1.0, "train_accuracy_recipient_token_crop_total": 1, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.1956521739130435, "grad_norm": 2.752408100243012, "learning_rate": 4.654298841826988e-06, "loss": 0.276, "step": 99 }, { "epoch": 0.1956521739130435, "step": 99, "train_accuracy_recipient_token": 0.7647058823529411, "train_accuracy_recipient_token_all": 0.7692307692307693, "train_accuracy_recipient_token_all_total": 13, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_schedule": 0.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.1956521739130435, "step": 99, "train_accuracy_recipient_token": 0.8, "train_accuracy_recipient_token_add": 0.5, "train_accuracy_recipient_token_add_total": 2, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 0.3333333333333333, "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_m": 0.5, "train_accuracy_recipient_token_m_total": 2, "train_total_number_recipient_token": 20 }, { "epoch": 0.1976284584980237, "grad_norm": 2.6892659000959456, "learning_rate": 4.646121984004666e-06, "loss": 0.2694, "step": 100 }, { "epoch": 0.1976284584980237, "step": 100, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { "epoch": 0.1976284584980237, "step": 100, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_get": 0.875, "train_accuracy_recipient_token_get_total": 8, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.19960474308300397, "grad_norm": 1.8294943574863658, "learning_rate": 4.637856907513366e-06, "loss": 0.1149, "step": 101 }, { "epoch": 0.19960474308300397, "step": 101, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.19960474308300397, "step": 101, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_build": 1.0, "train_accuracy_recipient_token_build_total": 1, "train_accuracy_recipient_token_con": 0.0, "train_accuracy_recipient_token_con_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_max": 1.0, "train_accuracy_recipient_token_max_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.2015810276679842, "grad_norm": 2.408233398782672, "learning_rate": 4.629503952098011e-06, "loss": 0.1773, "step": 102 }, { "epoch": 0.2015810276679842, "step": 102, "train_accuracy_recipient_token": 0.6875, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_be": 0.0, "train_accuracy_recipient_token_be_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.2015810276679842, "step": 102, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.20355731225296442, "grad_norm": 3.07819121819936, "learning_rate": 4.621063461115882e-06, "loss": 0.2781, "step": 103 }, { "epoch": 0.20355731225296442, "step": 103, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_choose": 1.0, "train_accuracy_recipient_token_choose_total": 1, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_order": 0.0, "train_accuracy_recipient_token_order_total": 1, "train_accuracy_recipient_token_plan": 1.0, "train_accuracy_recipient_token_plan_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.20355731225296442, "step": 103, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_armor": 1.0, "train_accuracy_recipient_token_armor_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.20553359683794467, "grad_norm": 2.3567297901560926, "learning_rate": 4.612535781522504e-06, "loss": 0.2303, "step": 104 }, { "epoch": 0.20553359683794467, "step": 104, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9230769230769231, "train_accuracy_recipient_token_all_total": 13, "train_accuracy_recipient_token_custom": 1.0, "train_accuracy_recipient_token_custom_total": 2, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.20553359683794467, "step": 104, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { "epoch": 0.2075098814229249, "grad_norm": 2.007174401193821, "learning_rate": 4.6039212638573835e-06, "loss": 0.2304, "step": 105 }, { "epoch": 0.2075098814229249, "step": 105, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 2, "train_accuracy_recipient_token_convert": 1.0, "train_accuracy_recipient_token_convert_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_render": 1.0, "train_accuracy_recipient_token_render_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.2075098814229249, "step": 105, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_mon": 0.75, "train_accuracy_recipient_token_mon_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.20948616600790515, "grad_norm": 1.8768688322152753, "learning_rate": 4.5952202622296015e-06, "loss": 0.1971, "step": 106 }, { "epoch": 0.20948616600790515, "step": 106, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.20948616600790515, "step": 106, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.5714285714285714, "train_accuracy_recipient_token_get_total": 7, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.21146245059288538, "grad_norm": 2.341750694388968, "learning_rate": 4.586433134303257e-06, "loss": 0.2459, "step": 107 }, { "epoch": 0.21146245059288538, "step": 107, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_flower": 0.75, "train_accuracy_recipient_token_flower_total": 4, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.21146245059288538, "step": 107, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.2134387351778656, "grad_norm": 2.1944375595196965, "learning_rate": 4.57756024128276e-06, "loss": 0.199, "step": 108 }, { "epoch": 0.2134387351778656, "step": 108, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 20 }, { "epoch": 0.2134387351778656, "step": 108, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_draw": 1.0, "train_accuracy_recipient_token_draw_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.21541501976284586, "grad_norm": 2.5062076998625566, "learning_rate": 4.5686019478979915e-06, "loss": 0.3176, "step": 109 }, { "epoch": 0.21541501976284586, "step": 109, "train_accuracy_recipient_token": 0.8, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_p": 1.0, "train_accuracy_recipient_token_p_total": 1, "train_accuracy_recipient_token_report": 0.0, "train_accuracy_recipient_token_report_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_accuracy_recipient_token_store": 1.0, "train_accuracy_recipient_token_store_total": 2, "train_total_number_recipient_token": 15 }, { "epoch": 0.21541501976284586, "step": 109, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.21739130434782608, "grad_norm": 1.746999082447299, "learning_rate": 4.559558622389304e-06, "loss": 0.1385, "step": 110 }, { "epoch": 0.21739130434782608, "step": 110, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_log": 1.0, "train_accuracy_recipient_token_log_total": 2, "train_total_number_recipient_token": 16 }, { "epoch": 0.21739130434782608, "step": 110, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.7777777777777778, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_hire": 1.0, "train_accuracy_recipient_token_hire_total": 1, "train_accuracy_recipient_token_track": 1.0, "train_accuracy_recipient_token_track_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.21936758893280633, "grad_norm": 2.7961010968564795, "learning_rate": 4.55043063649239e-06, "loss": 0.3181, "step": 111 }, { "epoch": 0.21936758893280633, "step": 111, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_project": 1.0, "train_accuracy_recipient_token_project_total": 1, "train_accuracy_recipient_token_stat": 1.0, "train_accuracy_recipient_token_stat_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.21936758893280633, "step": 111, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 2, "train_accuracy_recipient_token_s": 1.0, "train_accuracy_recipient_token_s_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.22134387351778656, "grad_norm": 1.9989358452463166, "learning_rate": 4.541218365422997e-06, "loss": 0.2531, "step": 112 }, { "epoch": 0.22134387351778656, "step": 112, "train_accuracy_recipient_token": 0.7647058823529411, "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_ch": 1.0, "train_accuracy_recipient_token_ch_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 17 }, { "epoch": 0.22134387351778656, "step": 112, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_f": 0.0, "train_accuracy_recipient_token_f_total": 1, "train_accuracy_recipient_token_plant": 1.0, "train_accuracy_recipient_token_plant_total": 1, "train_accuracy_recipient_token_se": 1.0, "train_accuracy_recipient_token_se_total": 1, "train_accuracy_recipient_token_trim": 1.0, "train_accuracy_recipient_token_trim_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.22332015810276679, "grad_norm": 2.571319112719595, "learning_rate": 4.531922187861507e-06, "loss": 0.2898, "step": 113 }, { "epoch": 0.22332015810276679, "step": 113, "train_accuracy_recipient_token": 0.9285714285714286, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_f": 1.0, "train_accuracy_recipient_token_f_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_restore": 1.0, "train_accuracy_recipient_token_restore_total": 1, "train_total_number_recipient_token": 14 }, { "epoch": 0.22332015810276679, "step": 113, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_gr": 1.0, "train_accuracy_recipient_token_gr_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.22529644268774704, "grad_norm": 1.8363624831376564, "learning_rate": 4.522542485937369e-06, "loss": 0.2103, "step": 114 }, { "epoch": 0.22529644268774704, "step": 114, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 0.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.22529644268774704, "step": 114, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.22727272727272727, "grad_norm": 2.6890155172544046, "learning_rate": 4.513079645213391e-06, "loss": 0.3333, "step": 115 }, { "epoch": 0.22727272727272727, "step": 115, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.22727272727272727, "step": 115, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { "epoch": 0.22924901185770752, "grad_norm": 3.4159406545291486, "learning_rate": 4.5035340546698915e-06, "loss": 0.2674, "step": 116 }, { "epoch": 0.22924901185770752, "step": 116, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_flower": 1.0, "train_accuracy_recipient_token_flower_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.22924901185770752, "step": 116, "train_accuracy_recipient_token": 0.9333333333333333, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 15 }, { "epoch": 0.23122529644268774, "grad_norm": 2.099346426532093, "learning_rate": 4.493906106688712e-06, "loss": 0.3201, "step": 117 }, { "epoch": 0.23122529644268774, "step": 117, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_add": 1.0, "train_accuracy_recipient_token_add_total": 1, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 19 }, { "epoch": 0.23122529644268774, "step": 117, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 0.8888888888888888, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 20 }, { "epoch": 0.233201581027668, "grad_norm": 1.9975724683282075, "learning_rate": 4.484196197037082e-06, "loss": 0.2154, "step": 118 }, { "epoch": 0.233201581027668, "step": 118, "train_accuracy_recipient_token": 0.9047619047619048, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_extract": 1.0, "train_accuracy_recipient_token_extract_total": 1, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 2, "train_accuracy_recipient_token_sum": 1.0, "train_accuracy_recipient_token_sum_total": 1, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 2, "train_total_number_recipient_token": 21 }, { "epoch": 0.233201581027668, "step": 118, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.23517786561264822, "grad_norm": 1.8241360509276188, "learning_rate": 4.474404724851356e-06, "loss": 0.2666, "step": 119 }, { "epoch": 0.23517786561264822, "step": 119, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_arc": 1.0, "train_accuracy_recipient_token_arc_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.23517786561264822, "step": 119, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 20 }, { "epoch": 0.23715415019762845, "grad_norm": 2.374987760349219, "learning_rate": 4.464532092620607e-06, "loss": 0.3071, "step": 120 }, { "epoch": 0.23715415019762845, "step": 120, "train_accuracy_recipient_token": 0.8125, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_property": 1.0, "train_accuracy_recipient_token_property_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.23715415019762845, "step": 120, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.2391304347826087, "grad_norm": 1.5783968329951958, "learning_rate": 4.454578706170075e-06, "loss": 0.1226, "step": 121 }, { "epoch": 0.2391304347826087, "step": 121, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.2391304347826087, "step": 121, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 2, "train_accuracy_recipient_token_ex": 1.0, "train_accuracy_recipient_token_ex_total": 1, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_sc": 1.0, "train_accuracy_recipient_token_sc_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.24110671936758893, "grad_norm": 1.848245708230007, "learning_rate": 4.444544974644493e-06, "loss": 0.2494, "step": 122 }, { "epoch": 0.24110671936758893, "step": 122, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.24110671936758893, "step": 122, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_box": 1.0, "train_accuracy_recipient_token_box_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.24308300395256918, "grad_norm": 1.747036665792108, "learning_rate": 4.434431310491267e-06, "loss": 0.1616, "step": 123 }, { "epoch": 0.24308300395256918, "step": 123, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.24308300395256918, "step": 123, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_ins": 1.0, "train_accuracy_recipient_token_ins_total": 1, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.2450592885375494, "grad_norm": 2.235367946517307, "learning_rate": 4.424238129443515e-06, "loss": 0.2711, "step": 124 }, { "epoch": 0.2450592885375494, "step": 124, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { "epoch": 0.2450592885375494, "step": 124, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.24703557312252963, "grad_norm": 1.8124306151814813, "learning_rate": 4.413965850502987e-06, "loss": 0.1279, "step": 125 }, { "epoch": 0.24703557312252963, "step": 125, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.24703557312252963, "step": 125, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compile": 0.5, "train_accuracy_recipient_token_compile_total": 2, "train_accuracy_recipient_token_compress": 1.0, "train_accuracy_recipient_token_compress_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.2490118577075099, "grad_norm": 2.3030811699911142, "learning_rate": 4.4036148959228365e-06, "loss": 0.3385, "step": 126 }, { "epoch": 0.2490118577075099, "step": 126, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.2490118577075099, "step": 126, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_research": 1.0, "train_accuracy_recipient_token_research_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.2509881422924901, "grad_norm": 1.7077330162259072, "learning_rate": 4.3931856911902635e-06, "loss": 0.1785, "step": 127 }, { "epoch": 0.2509881422924901, "step": 127, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compose": 1.0, "train_accuracy_recipient_token_compose_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.2509881422924901, "step": 127, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_write": 1.0, "train_accuracy_recipient_token_write_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.25296442687747034, "grad_norm": 2.2097826093919095, "learning_rate": 4.382678665009028e-06, "loss": 0.2991, "step": 128 }, { "epoch": 0.25296442687747034, "step": 128, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.25296442687747034, "step": 128, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_ex": 1.0, "train_accuracy_recipient_token_ex_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.2549407114624506, "grad_norm": 2.2527852681173846, "learning_rate": 4.37209424928182e-06, "loss": 0.1948, "step": 129 }, { "epoch": 0.2549407114624506, "step": 129, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_flag": 0.75, "train_accuracy_recipient_token_flag_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.2549407114624506, "step": 129, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9230769230769231, "train_accuracy_recipient_token_all_total": 13, "train_accuracy_recipient_token_compare": 0.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_describe": 1.0, "train_accuracy_recipient_token_describe_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.25691699604743085, "grad_norm": 2.248286132307088, "learning_rate": 4.361432879092518e-06, "loss": 0.345, "step": 130 }, { "epoch": 0.25691699604743085, "step": 130, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_ass": 0.5, "train_accuracy_recipient_token_ass_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_plan": 1.0, "train_accuracy_recipient_token_plan_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.25691699604743085, "step": 130, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.875, "train_accuracy_recipient_token_get_total": 8, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.25889328063241107, "grad_norm": 2.7851285907266368, "learning_rate": 4.350694992688289e-06, "loss": 0.4034, "step": 131 }, { "epoch": 0.25889328063241107, "step": 131, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.75, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_cross": 1.0, "train_accuracy_recipient_token_cross_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.25889328063241107, "step": 131, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 18 }, { "epoch": 0.2608695652173913, "grad_norm": 2.0403309814666697, "learning_rate": 4.339881031461588e-06, "loss": 0.2414, "step": 132 }, { "epoch": 0.2608695652173913, "step": 132, "train_accuracy_recipient_token": 0.8125, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 16 }, { "epoch": 0.2608695652173913, "step": 132, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_track": 1.0, "train_accuracy_recipient_token_track_total": 1, "train_accuracy_recipient_token_write": 1.0, "train_accuracy_recipient_token_write_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.2628458498023715, "grad_norm": 2.697975383814613, "learning_rate": 4.328991439932003e-06, "loss": 0.313, "step": 133 }, { "epoch": 0.2628458498023715, "step": 133, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 1, "train_accuracy_recipient_token_rent": 1.0, "train_accuracy_recipient_token_rent_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.2628458498023715, "step": 133, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_ex": 1.0, "train_accuracy_recipient_token_ex_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_track": 1.0, "train_accuracy_recipient_token_track_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.2648221343873518, "grad_norm": 2.5542065987019664, "learning_rate": 4.318026665727993e-06, "loss": 0.2338, "step": 134 }, { "epoch": 0.2648221343873518, "step": 134, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_evaluate": 1.0, "train_accuracy_recipient_token_evaluate_total": 1, "train_accuracy_recipient_token_fetch": 1.0, "train_accuracy_recipient_token_fetch_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_sport": 1.0, "train_accuracy_recipient_token_sport_total": 3, "train_total_number_recipient_token": 20 }, { "epoch": 0.2648221343873518, "step": 134, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_add": 1.0, "train_accuracy_recipient_token_add_total": 1, "train_accuracy_recipient_token_all": 0.8461538461538461, "train_accuracy_recipient_token_all_total": 13, "train_accuracy_recipient_token_apply": 1.0, "train_accuracy_recipient_token_apply_total": 1, "train_accuracy_recipient_token_crop": 1.0, "train_accuracy_recipient_token_crop_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.26679841897233203, "grad_norm": 2.4386171945089004, "learning_rate": 4.3069871595684795e-06, "loss": 0.1782, "step": 135 }, { "epoch": 0.26679841897233203, "step": 135, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.26679841897233203, "step": 135, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_m": 1.0, "train_accuracy_recipient_token_m_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.26877470355731226, "grad_norm": 1.9475949445800667, "learning_rate": 4.295873375244319e-06, "loss": 0.2018, "step": 136 }, { "epoch": 0.26877470355731226, "step": 136, "train_accuracy_recipient_token": 0.75, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 16 }, { "epoch": 0.26877470355731226, "step": 136, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_ele": 1.0, "train_accuracy_recipient_token_ele_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 17 }, { "epoch": 0.2707509881422925, "grad_norm": 1.932745049719722, "learning_rate": 4.284685769599658e-06, "loss": 0.3062, "step": 137 }, { "epoch": 0.2707509881422925, "step": 137, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 1, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 16 }, { "epoch": 0.2707509881422925, "step": 137, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 2, "train_accuracy_recipient_token_object": 1.0, "train_accuracy_recipient_token_object_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.2727272727272727, "grad_norm": 2.5725457529065006, "learning_rate": 4.273424802513145e-06, "loss": 0.2588, "step": 138 }, { "epoch": 0.2727272727272727, "step": 138, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_deploy": 1.0, "train_accuracy_recipient_token_deploy_total": 1, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_plan": 1.0, "train_accuracy_recipient_token_plan_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.2727272727272727, "step": 138, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_convert": 1.0, "train_accuracy_recipient_token_convert_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.274703557312253, "grad_norm": 1.86126996030237, "learning_rate": 4.26209093687903e-06, "loss": 0.205, "step": 139 }, { "epoch": 0.274703557312253, "step": 139, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_detect": 0.0, "train_accuracy_recipient_token_detect_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.274703557312253, "step": 139, "train_accuracy_recipient_token": 0.7894736842105263, "train_accuracy_recipient_token_add": 1.0, "train_accuracy_recipient_token_add_total": 1, "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 2, "train_accuracy_recipient_token_decode": 0.75, "train_accuracy_recipient_token_decode_total": 4, "train_accuracy_recipient_token_subtract": 1.0, "train_accuracy_recipient_token_subtract_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.2766798418972332, "grad_norm": 3.716815765106261, "learning_rate": 4.2506846385881375e-06, "loss": 0.4463, "step": 140 }, { "epoch": 0.2766798418972332, "step": 140, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.2766798418972332, "step": 140, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_study": 1.0, "train_accuracy_recipient_token_study_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.27865612648221344, "grad_norm": 1.8259425268376976, "learning_rate": 4.239206376508716e-06, "loss": 0.2031, "step": 141 }, { "epoch": 0.27865612648221344, "step": 141, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_d": 0.5, "train_accuracy_recipient_token_d_total": 2, "train_accuracy_recipient_token_fly": 1.0, "train_accuracy_recipient_token_fly_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.27865612648221344, "step": 141, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.28063241106719367, "grad_norm": 1.9633184170686586, "learning_rate": 4.227656622467162e-06, "loss": 0.2176, "step": 142 }, { "epoch": 0.28063241106719367, "step": 142, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_horse": 1.0, "train_accuracy_recipient_token_horse_total": 2, "train_accuracy_recipient_token_invest": 1.0, "train_accuracy_recipient_token_invest_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.28063241106719367, "step": 142, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_arr": 1.0, "train_accuracy_recipient_token_arr_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_accuracy_recipient_token_s": 1.0, "train_accuracy_recipient_token_s_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.2826086956521739, "grad_norm": 2.1587360641406623, "learning_rate": 4.2160358512286266e-06, "loss": 0.3984, "step": 143 }, { "epoch": 0.2826086956521739, "step": 143, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_f": 1.0, "train_accuracy_recipient_token_f_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_im": 1.0, "train_accuracy_recipient_token_im_total": 1, "train_accuracy_recipient_token_schedule": 0.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.2826086956521739, "step": 143, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_sh": 1.0, "train_accuracy_recipient_token_sh_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.2845849802371542, "grad_norm": 2.0510066673643084, "learning_rate": 4.204344540477499e-06, "loss": 0.2632, "step": 144 }, { "epoch": 0.2845849802371542, "step": 144, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 18 }, { "epoch": 0.2845849802371542, "step": 144, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_accuracy_recipient_token_verify": 1.0, "train_accuracy_recipient_token_verify_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.2865612648221344, "grad_norm": 2.0576342231497513, "learning_rate": 4.192583170797775e-06, "loss": 0.163, "step": 145 }, { "epoch": 0.2865612648221344, "step": 145, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.2865612648221344, "step": 145, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_enh": 1.0, "train_accuracy_recipient_token_enh_total": 1, "train_accuracy_recipient_token_ex": 1.0, "train_accuracy_recipient_token_ex_total": 1, "train_accuracy_recipient_token_public": 0.0, "train_accuracy_recipient_token_public_total": 1, "train_accuracy_recipient_token_street": 1.0, "train_accuracy_recipient_token_street_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.2885375494071146, "grad_norm": 1.7728828833133476, "learning_rate": 4.1807522256532925e-06, "loss": 0.2257, "step": 146 }, { "epoch": 0.2885375494071146, "step": 146, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8333333333333334, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 17 }, { "epoch": 0.2885375494071146, "step": 146, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 17 }, { "epoch": 0.29051383399209485, "grad_norm": 2.6572528287169903, "learning_rate": 4.1688521913678706e-06, "loss": 0.3191, "step": 147 }, { "epoch": 0.29051383399209485, "step": 147, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 0.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_c": 0.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.29051383399209485, "step": 147, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 0.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_assign": 1.0, "train_accuracy_recipient_token_assign_total": 1, "train_accuracy_recipient_token_configure": 1.0, "train_accuracy_recipient_token_configure_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 16 }, { "epoch": 0.2924901185770751, "grad_norm": 2.487728724207803, "learning_rate": 4.156883557105308e-06, "loss": 0.2116, "step": 148 }, { "epoch": 0.2924901185770751, "step": 148, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.2924901185770751, "step": 148, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 20 }, { "epoch": 0.29446640316205536, "grad_norm": 1.8557301698416115, "learning_rate": 4.144846814849282e-06, "loss": 0.1994, "step": 149 }, { "epoch": 0.29446640316205536, "step": 149, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_clean": 1.0, "train_accuracy_recipient_token_clean_total": 1, "train_accuracy_recipient_token_install": 1.0, "train_accuracy_recipient_token_install_total": 1, "train_accuracy_recipient_token_remove": 0.0, "train_accuracy_recipient_token_remove_total": 1, "train_accuracy_recipient_token_replace": 1.0, "train_accuracy_recipient_token_replace_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.29446640316205536, "step": 149, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 4, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.2964426877470356, "grad_norm": 3.3999140933900542, "learning_rate": 4.132742459383122e-06, "loss": 0.2816, "step": 150 }, { "epoch": 0.2964426877470356, "step": 150, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_fac": 0.6666666666666666, "train_accuracy_recipient_token_fac_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.2964426877470356, "step": 150, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_mount": 1.0, "train_accuracy_recipient_token_mount_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.2984189723320158, "grad_norm": 1.6480321230907027, "learning_rate": 4.120570988269472e-06, "loss": 0.1801, "step": 151 }, { "epoch": 0.2984189723320158, "step": 151, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 0.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_park": 0.6666666666666666, "train_accuracy_recipient_token_park_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.2984189723320158, "step": 151, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_accuracy_recipient_token_schedule": 0.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.30039525691699603, "grad_norm": 1.2847579476886248, "learning_rate": 4.1083329018298356e-06, "loss": 0.1073, "step": 152 }, { "epoch": 0.30039525691699603, "step": 152, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_detect": 1.0, "train_accuracy_recipient_token_detect_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.30039525691699603, "step": 152, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 18 }, { "epoch": 0.30237154150197626, "grad_norm": 1.6318246749351943, "learning_rate": 4.096028703124014e-06, "loss": 0.1468, "step": 153 }, { "epoch": 0.30237154150197626, "step": 153, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_detect": 1.0, "train_accuracy_recipient_token_detect_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_sum": 0.0, "train_accuracy_recipient_token_sum_total": 1, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.30237154150197626, "step": 153, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_city": 1.0, "train_accuracy_recipient_token_city_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.30434782608695654, "grad_norm": 2.3337051073106223, "learning_rate": 4.083658897929425e-06, "loss": 0.3715, "step": 154 }, { "epoch": 0.30434782608695654, "step": 154, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_event": 1.0, "train_accuracy_recipient_token_event_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.30434782608695654, "step": 154, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_bike": 1.0, "train_accuracy_recipient_token_bike_total": 2, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.30632411067193677, "grad_norm": 1.9972989241470553, "learning_rate": 4.071223994720309e-06, "loss": 0.2331, "step": 155 }, { "epoch": 0.30632411067193677, "step": 155, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.8333333333333334, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_disc": 1.0, "train_accuracy_recipient_token_disc_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 16 }, { "epoch": 0.30632411067193677, "step": 155, "train_accuracy_recipient_token": 0.9047619047619048, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_convert": 1.0, "train_accuracy_recipient_token_convert_total": 2, "train_accuracy_recipient_token_count": 0.5, "train_accuracy_recipient_token_count_total": 2, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 21 }, { "epoch": 0.308300395256917, "grad_norm": 2.4018437558185077, "learning_rate": 4.058724504646834e-06, "loss": 0.1986, "step": 156 }, { "epoch": 0.308300395256917, "step": 156, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 2, "train_accuracy_recipient_token_interpret": 1.0, "train_accuracy_recipient_token_interpret_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.308300395256917, "step": 156, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 20 }, { "epoch": 0.3102766798418972, "grad_norm": 1.9392056531490096, "learning_rate": 4.046160941514079e-06, "loss": 0.1503, "step": 157 }, { "epoch": 0.3102766798418972, "step": 157, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 1, "train_accuracy_recipient_token_v": 1.0, "train_accuracy_recipient_token_v_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.3102766798418972, "step": 157, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_m": 1.0, "train_accuracy_recipient_token_m_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.31225296442687744, "grad_norm": 2.00451491556449, "learning_rate": 4.033533821760917e-06, "loss": 0.3046, "step": 158 }, { "epoch": 0.31225296442687744, "step": 158, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 1, "train_accuracy_recipient_token_sum": 1.0, "train_accuracy_recipient_token_sum_total": 1, "train_accuracy_recipient_token_verify": 1.0, "train_accuracy_recipient_token_verify_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.31225296442687744, "step": 158, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.3142292490118577, "grad_norm": 3.024122625725331, "learning_rate": 4.020843664438783e-06, "loss": 0.3273, "step": 159 }, { "epoch": 0.3142292490118577, "step": 159, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.3142292490118577, "step": 159, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_report": 0.0, "train_accuracy_recipient_token_report_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.31620553359683795, "grad_norm": 1.8881036988631603, "learning_rate": 4.008090991190341e-06, "loss": 0.1621, "step": 160 }, { "epoch": 0.31620553359683795, "step": 160, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_interpret": 1.0, "train_accuracy_recipient_token_interpret_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.31620553359683795, "step": 160, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_d": 1.0, "train_accuracy_recipient_token_d_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.3181818181818182, "grad_norm": 3.3688529657875326, "learning_rate": 3.99527632622804e-06, "loss": 0.3867, "step": 161 }, { "epoch": 0.3181818181818182, "step": 161, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_audit": 1.0, "train_accuracy_recipient_token_audit_total": 1, "train_accuracy_recipient_token_environment": 1.0, "train_accuracy_recipient_token_environment_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_public": 0.0, "train_accuracy_recipient_token_public_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.3181818181818182, "step": 161, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_f": 1.0, "train_accuracy_recipient_token_f_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_search": 0.6666666666666666, "train_accuracy_recipient_token_search_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.3201581027667984, "grad_norm": 2.245641056003419, "learning_rate": 3.982400196312565e-06, "loss": 0.2937, "step": 162 }, { "epoch": 0.3201581027667984, "step": 162, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_check": 0.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.3201581027667984, "step": 162, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_histor": 1.0, "train_accuracy_recipient_token_histor_total": 2, "train_total_number_recipient_token": 16 }, { "epoch": 0.3221343873517787, "grad_norm": 2.548113126420555, "learning_rate": 3.969463130731183e-06, "loss": 0.2783, "step": 163 }, { "epoch": 0.3221343873517787, "step": 163, "train_accuracy_recipient_token": 0.5625, "train_accuracy_recipient_token_all": 0.5454545454545454, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_detect": 0.0, "train_accuracy_recipient_token_detect_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 2, "train_total_number_recipient_token": 16 }, { "epoch": 0.3221343873517787, "step": 163, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_plan": 1.0, "train_accuracy_recipient_token_plan_total": 1, "train_accuracy_recipient_token_s": 1.0, "train_accuracy_recipient_token_s_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.3241106719367589, "grad_norm": 2.7435239695326823, "learning_rate": 3.9564656612759904e-06, "loss": 0.2899, "step": 164 }, { "epoch": 0.3241106719367589, "step": 164, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 17 }, { "epoch": 0.3241106719367589, "step": 164, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.32608695652173914, "grad_norm": 2.2167192768597896, "learning_rate": 3.943408322222049e-06, "loss": 0.2021, "step": 165 }, { "epoch": 0.32608695652173914, "step": 165, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8461538461538461, "train_accuracy_recipient_token_all_total": 13, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_detect": 1.0, "train_accuracy_recipient_token_detect_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.32608695652173914, "step": 165, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_race": 1.0, "train_accuracy_recipient_token_race_total": 2, "train_total_number_recipient_token": 16 }, { "epoch": 0.32806324110671936, "grad_norm": 2.363046339246385, "learning_rate": 3.930291650305424e-06, "loss": 0.3441, "step": 166 }, { "epoch": 0.32806324110671936, "step": 166, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9166666666666666, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_evaluate": 1.0, "train_accuracy_recipient_token_evaluate_total": 1, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 20 }, { "epoch": 0.32806324110671936, "step": 166, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_add": 1.0, "train_accuracy_recipient_token_add_total": 1, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_change": 0.0, "train_accuracy_recipient_token_change_total": 1, "train_accuracy_recipient_token_highlight": 1.0, "train_accuracy_recipient_token_highlight_total": 2, "train_accuracy_recipient_token_remove": 1.0, "train_accuracy_recipient_token_remove_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.3300395256916996, "grad_norm": 2.6127483795239708, "learning_rate": 3.917116184701125e-06, "loss": 0.2668, "step": 167 }, { "epoch": 0.3300395256916996, "step": 167, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_configure": 1.0, "train_accuracy_recipient_token_configure_total": 1, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.3300395256916996, "step": 167, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 2, "train_total_number_recipient_token": 20 }, { "epoch": 0.33201581027667987, "grad_norm": 1.9279791632118746, "learning_rate": 3.903882467000938e-06, "loss": 0.1514, "step": 168 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 8, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 14 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.8, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.6, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_ident": 0.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 2, "train_total_number_recipient_token": 16 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_histor": 1.0, "train_accuracy_recipient_token_histor_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 4, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 17 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.6, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_ident": 0.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_interpret": 1.0, "train_accuracy_recipient_token_interpret_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.8666666666666667, "train_accuracy_recipient_token_all_total": 15, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.75, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.25, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_tour": 0.5, "train_accuracy_recipient_token_tour_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 0.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 19 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 17 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.85, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_hang": 0.0, "train_accuracy_recipient_token_hang_total": 1, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_accuracy_recipient_token_remove": 1.0, "train_accuracy_recipient_token_remove_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_accuracy_recipient_token_invest": 1.0, "train_accuracy_recipient_token_invest_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_add": 1.0, "train_accuracy_recipient_token_add_total": 1, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_construct": 0.0, "train_accuracy_recipient_token_construct_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_select": 1.0, "train_accuracy_recipient_token_select_total": 2, "train_total_number_recipient_token": 16 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_histor": 0.0, "train_accuracy_recipient_token_histor_total": 1, "train_accuracy_recipient_token_m": 1.0, "train_accuracy_recipient_token_m_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_gr": 1.0, "train_accuracy_recipient_token_gr_total": 3, "train_accuracy_recipient_token_ident": 0.0, "train_accuracy_recipient_token_ident_total": 2, "train_total_number_recipient_token": 20 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.8571428571428571, "train_accuracy_recipient_token_all": 0.8888888888888888, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_draw": 1.0, "train_accuracy_recipient_token_draw_total": 1, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 14 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_review": 0.5, "train_accuracy_recipient_token_review_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_create": 0.5, "train_accuracy_recipient_token_create_total": 2, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.875, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 0.875, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.7894736842105263, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_custom": 0.5, "train_accuracy_recipient_token_custom_total": 4, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 19 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_eng": 0.5, "train_accuracy_recipient_token_eng_total": 2, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 1, "train_accuracy_recipient_token_lock": 1.0, "train_accuracy_recipient_token_lock_total": 1, "train_accuracy_recipient_token_tem": 1.0, "train_accuracy_recipient_token_tem_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 16 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_car": 1.0, "train_accuracy_recipient_token_car_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.7222222222222222, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 0.4, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_invest": 1.0, "train_accuracy_recipient_token_invest_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_bird": 1.0, "train_accuracy_recipient_token_bird_total": 2, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_histor": 1.0, "train_accuracy_recipient_token_histor_total": 1, "train_accuracy_recipient_token_symbol": 1.0, "train_accuracy_recipient_token_symbol_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_organ": 0.0, "train_accuracy_recipient_token_organ_total": 1, "train_accuracy_recipient_token_rent": 1.0, "train_accuracy_recipient_token_rent_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_ident": 0.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_predict": 1.0, "train_accuracy_recipient_token_predict_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_evaluate": 0.0, "train_accuracy_recipient_token_evaluate_total": 1, "train_accuracy_recipient_token_find": 0.3333333333333333, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.7777777777777778, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_locate": 0.0, "train_accuracy_recipient_token_locate_total": 1, "train_accuracy_recipient_token_schedule": 0.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_park": 1.0, "train_accuracy_recipient_token_park_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.7777777777777778, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_art": 0.6666666666666666, "train_accuracy_recipient_token_art_total": 3, "train_accuracy_recipient_token_artist": 0.0, "train_accuracy_recipient_token_artist_total": 1, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 0.5, "train_accuracy_recipient_token_analy_total": 2, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_object": 0.6666666666666666, "train_accuracy_recipient_token_object_total": 3, "train_accuracy_recipient_token_transfer": 0.0, "train_accuracy_recipient_token_transfer_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_fruit": 1.0, "train_accuracy_recipient_token_fruit_total": 4, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_organ": 1.0, "train_accuracy_recipient_token_organ_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_interpret": 1.0, "train_accuracy_recipient_token_interpret_total": 1, "train_accuracy_recipient_token_stage": 1.0, "train_accuracy_recipient_token_stage_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.7647058823529411, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_rent": 1.0, "train_accuracy_recipient_token_rent_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 4, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 18 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.7777777777777778, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 5, "train_accuracy_recipient_token_f": 0.0, "train_accuracy_recipient_token_f_total": 1, "train_accuracy_recipient_token_pest": 0.0, "train_accuracy_recipient_token_pest_total": 1, "train_accuracy_recipient_token_plant": 1.0, "train_accuracy_recipient_token_plant_total": 1, "train_accuracy_recipient_token_sun": 1.0, "train_accuracy_recipient_token_sun_total": 1, "train_total_number_recipient_token": 9 }, { "epoch": 0.33201581027667987, "eval_loss": 0.218434140086174, "eval_runtime": 425.3868, "eval_samples_per_second": 0.237, "eval_steps_per_second": 0.12, "step": 168 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_dis": 1.0, "train_accuracy_recipient_token_dis_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_up": 1.0, "train_accuracy_recipient_token_up_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.33201581027667987, "step": 168, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 17 }, { "epoch": 0.3339920948616601, "grad_norm": 2.3704926233947448, "learning_rate": 3.890591041191162e-06, "loss": 0.2582, "step": 169 }, { "epoch": 0.3339920948616601, "step": 169, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.875, "train_accuracy_recipient_token_get_total": 8, "train_accuracy_recipient_token_verify": 1.0, "train_accuracy_recipient_token_verify_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.3339920948616601, "step": 169, "train_accuracy_recipient_token": 0.8, "train_accuracy_recipient_token_add": 1.0, "train_accuracy_recipient_token_add_total": 1, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_change": 1.0, "train_accuracy_recipient_token_change_total": 1, "train_accuracy_recipient_token_get": 0.4, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_replace": 0.5, "train_accuracy_recipient_token_replace_total": 2, "train_total_number_recipient_token": 20 }, { "epoch": 0.3359683794466403, "grad_norm": 2.4827135380795813, "learning_rate": 3.8772424536302565e-06, "loss": 0.227, "step": 170 }, { "epoch": 0.3359683794466403, "step": 170, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 2, "train_accuracy_recipient_token_bird": 1.0, "train_accuracy_recipient_token_bird_total": 4, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.3359683794466403, "step": 170, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.7777777777777778, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.33794466403162055, "grad_norm": 1.626312276221932, "learning_rate": 3.863837253026372e-06, "loss": 0.2166, "step": 171 }, { "epoch": 0.33794466403162055, "step": 171, "train_accuracy_recipient_token": 0.7894736842105263, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_arch": 0.0, "train_accuracy_recipient_token_arch_total": 1, "train_accuracy_recipient_token_building": 1.0, "train_accuracy_recipient_token_building_total": 1, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_histor": 0.0, "train_accuracy_recipient_token_histor_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.33794466403162055, "step": 171, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_ident": 0.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_im": 1.0, "train_accuracy_recipient_token_im_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.33992094861660077, "grad_norm": 1.5592305643249844, "learning_rate": 3.8503759904148005e-06, "loss": 0.183, "step": 172 }, { "epoch": 0.33992094861660077, "step": 172, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9166666666666666, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { "epoch": 0.33992094861660077, "step": 172, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 0.8, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_organ": 1.0, "train_accuracy_recipient_token_organ_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.34189723320158105, "grad_norm": 2.1789375117907976, "learning_rate": 3.8368592191353246e-06, "loss": 0.256, "step": 173 }, { "epoch": 0.34189723320158105, "step": 173, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.34189723320158105, "step": 173, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_bookmark": 1.0, "train_accuracy_recipient_token_bookmark_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_refresh": 1.0, "train_accuracy_recipient_token_refresh_total": 1, "train_accuracy_recipient_token_switch": 0.5, "train_accuracy_recipient_token_switch_total": 2, "train_total_number_recipient_token": 20 }, { "epoch": 0.3438735177865613, "grad_norm": 2.171729945066953, "learning_rate": 3.823287494809469e-06, "loss": 0.2111, "step": 174 }, { "epoch": 0.3438735177865613, "step": 174, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_remove": 1.0, "train_accuracy_recipient_token_remove_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.3438735177865613, "step": 174, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_s": 1.0, "train_accuracy_recipient_token_s_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.3458498023715415, "grad_norm": 2.962276041812103, "learning_rate": 3.8096613753176635e-06, "loss": 0.3309, "step": 175 }, { "epoch": 0.3458498023715415, "step": 175, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 2, "train_accuracy_recipient_token_design": 1.0, "train_accuracy_recipient_token_design_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.3458498023715415, "step": 175, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 16 }, { "epoch": 0.34782608695652173, "grad_norm": 1.9766749939858486, "learning_rate": 3.7959814207763134e-06, "loss": 0.3564, "step": 176 }, { "epoch": 0.34782608695652173, "step": 176, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.34782608695652173, "step": 176, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_setup": 1.0, "train_accuracy_recipient_token_setup_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.34980237154150196, "grad_norm": 3.077796900581793, "learning_rate": 3.782248193514766e-06, "loss": 0.2989, "step": 177 }, { "epoch": 0.34980237154150196, "step": 177, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.34980237154150196, "step": 177, "train_accuracy_recipient_token": 0.7647058823529411, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_project": 0.0, "train_accuracy_recipient_token_project_total": 1, "train_accuracy_recipient_token_sc": 0.5, "train_accuracy_recipient_token_sc_total": 2, "train_accuracy_recipient_token_verify": 1.0, "train_accuracy_recipient_token_verify_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.35177865612648224, "grad_norm": 2.055425382576303, "learning_rate": 3.7684622580522057e-06, "loss": 0.3784, "step": 178 }, { "epoch": 0.35177865612648224, "step": 178, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.35177865612648224, "step": 178, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.9, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 20 }, { "epoch": 0.35375494071146246, "grad_norm": 1.638839258863397, "learning_rate": 3.7546241810744444e-06, "loss": 0.1502, "step": 179 }, { "epoch": 0.35375494071146246, "step": 179, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_fin": 0.75, "train_accuracy_recipient_token_fin_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_invest": 1.0, "train_accuracy_recipient_token_invest_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.35375494071146246, "step": 179, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { "epoch": 0.3557312252964427, "grad_norm": 1.6977698069216782, "learning_rate": 3.740734531410626e-06, "loss": 0.1859, "step": 180 }, { "epoch": 0.3557312252964427, "step": 180, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_detect": 1.0, "train_accuracy_recipient_token_detect_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.3557312252964427, "step": 180, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 2, "train_accuracy_recipient_token_buy": 1.0, "train_accuracy_recipient_token_buy_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_s": 1.0, "train_accuracy_recipient_token_s_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.3577075098814229, "grad_norm": 2.104044600552162, "learning_rate": 3.7267938800098454e-06, "loss": 0.1879, "step": 181 }, { "epoch": 0.3577075098814229, "step": 181, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.3577075098814229, "step": 181, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.35968379446640314, "grad_norm": 2.3821501005519194, "learning_rate": 3.71280279991768e-06, "loss": 0.256, "step": 182 }, { "epoch": 0.35968379446640314, "step": 182, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_cl": 1.0, "train_accuracy_recipient_token_cl_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_im": 1.0, "train_accuracy_recipient_token_im_total": 1, "train_accuracy_recipient_token_search": 0.5, "train_accuracy_recipient_token_search_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.35968379446640314, "step": 182, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.3616600790513834, "grad_norm": 1.7412853263181982, "learning_rate": 3.698761866252635e-06, "loss": 0.2009, "step": 183 }, { "epoch": 0.3616600790513834, "step": 183, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.3616600790513834, "step": 183, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 3, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.36363636363636365, "grad_norm": 1.6453660919523958, "learning_rate": 3.684671656182497e-06, "loss": 0.1797, "step": 184 }, { "epoch": 0.36363636363636365, "step": 184, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.36363636363636365, "step": 184, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_arch": 1.0, "train_accuracy_recipient_token_arch_total": 2, "train_accuracy_recipient_token_build": 1.0, "train_accuracy_recipient_token_build_total": 3, "train_accuracy_recipient_token_histor": 0.0, "train_accuracy_recipient_token_histor_total": 1, "train_accuracy_recipient_token_s": 1.0, "train_accuracy_recipient_token_s_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.36561264822134387, "grad_norm": 1.7214718462583107, "learning_rate": 3.670532748900615e-06, "loss": 0.251, "step": 185 }, { "epoch": 0.36561264822134387, "step": 185, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9166666666666666, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.36561264822134387, "step": 185, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_land": 1.0, "train_accuracy_recipient_token_land_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.3675889328063241, "grad_norm": 1.9383253652427745, "learning_rate": 3.656345725602089e-06, "loss": 0.213, "step": 186 }, { "epoch": 0.3675889328063241, "step": 186, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 16 }, { "epoch": 0.3675889328063241, "step": 186, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { "epoch": 0.3695652173913043, "grad_norm": 2.5997132280472526, "learning_rate": 3.642111169459879e-06, "loss": 0.3241, "step": 187 }, { "epoch": 0.3695652173913043, "step": 187, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_evaluate": 1.0, "train_accuracy_recipient_token_evaluate_total": 1, "train_accuracy_recipient_token_phrase": 1.0, "train_accuracy_recipient_token_phrase_total": 4, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.3695652173913043, "step": 187, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_plan": 0.0, "train_accuracy_recipient_token_plan_total": 1, "train_accuracy_recipient_token_s": 1.0, "train_accuracy_recipient_token_s_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.3715415019762846, "grad_norm": 1.9124092057739746, "learning_rate": 3.6278296656008366e-06, "loss": 0.1795, "step": 188 }, { "epoch": 0.3715415019762846, "step": 188, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.3715415019762846, "step": 188, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_compare": 0.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_s": 1.0, "train_accuracy_recipient_token_s_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.37351778656126483, "grad_norm": 2.4304939026300847, "learning_rate": 3.613501801081648e-06, "loss": 0.3143, "step": 189 }, { "epoch": 0.37351778656126483, "step": 189, "train_accuracy_recipient_token": 0.9333333333333333, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_organ": 1.0, "train_accuracy_recipient_token_organ_total": 1, "train_total_number_recipient_token": 15 }, { "epoch": 0.37351778656126483, "step": 189, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_compare": 0.5, "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.37549407114624506, "grad_norm": 1.8544268929291403, "learning_rate": 3.599128164864706e-06, "loss": 0.1948, "step": 190 }, { "epoch": 0.37549407114624506, "step": 190, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.37549407114624506, "step": 190, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 18 }, { "epoch": 0.3774703557312253, "grad_norm": 2.7234941325542286, "learning_rate": 3.5847093477938955e-06, "loss": 0.3302, "step": 191 }, { "epoch": 0.3774703557312253, "step": 191, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 0.5, "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.3774703557312253, "step": 191, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8888888888888888, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.3794466403162055, "grad_norm": 1.7644071316796994, "learning_rate": 3.5702459425703146e-06, "loss": 0.198, "step": 192 }, { "epoch": 0.3794466403162055, "step": 192, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.3794466403162055, "step": 192, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_clean": 1.0, "train_accuracy_recipient_token_clean_total": 1, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 3, "train_accuracy_recipient_token_rep": 0.5, "train_accuracy_recipient_token_rep_total": 2, "train_accuracy_recipient_token_restore": 1.0, "train_accuracy_recipient_token_restore_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.3814229249011858, "grad_norm": 2.610658057678565, "learning_rate": 3.5557385437279e-06, "loss": 0.3898, "step": 193 }, { "epoch": 0.3814229249011858, "step": 193, "train_accuracy_recipient_token": 0.7894736842105263, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 0.0, "train_accuracy_recipient_token_book_total": 2, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_send": 1.0, "train_accuracy_recipient_token_send_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.3814229249011858, "step": 193, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.383399209486166, "grad_norm": 2.4079072059937383, "learning_rate": 3.541187747608998e-06, "loss": 0.1893, "step": 194 }, { "epoch": 0.383399209486166, "step": 194, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.383399209486166, "step": 194, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_arch": 1.0, "train_accuracy_recipient_token_arch_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 17 }, { "epoch": 0.38537549407114624, "grad_norm": 1.9829071189672327, "learning_rate": 3.5265941523398455e-06, "loss": 0.2634, "step": 195 }, { "epoch": 0.38537549407114624, "step": 195, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 17 }, { "epoch": 0.38537549407114624, "step": 195, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_log": 1.0, "train_accuracy_recipient_token_log_total": 4, "train_total_number_recipient_token": 17 }, { "epoch": 0.38735177865612647, "grad_norm": 2.237176244877257, "learning_rate": 3.5119583578059845e-06, "loss": 0.2343, "step": 196 }, { "epoch": 0.38735177865612647, "step": 196, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 0.5, "train_accuracy_recipient_token_book_total": 2, "train_accuracy_recipient_token_find": 0.75, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.38735177865612647, "step": 196, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 20 }, { "epoch": 0.3893280632411067, "grad_norm": 1.5354145070510907, "learning_rate": 3.4972809656276047e-06, "loss": 0.1565, "step": 197 }, { "epoch": 0.3893280632411067, "step": 197, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 0.3333333333333333, "train_accuracy_recipient_token_analy_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_log": 1.0, "train_accuracy_recipient_token_log_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.3893280632411067, "step": 197, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 1, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.391304347826087, "grad_norm": 1.8679250394533993, "learning_rate": 3.4825625791348093e-06, "loss": 0.2068, "step": 198 }, { "epoch": 0.391304347826087, "step": 198, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_trans": 1.0, "train_accuracy_recipient_token_trans_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.391304347826087, "step": 198, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 3, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_histor": 1.0, "train_accuracy_recipient_token_histor_total": 3, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.3932806324110672, "grad_norm": 2.1127643306550565, "learning_rate": 3.467803803342821e-06, "loss": 0.2574, "step": 199 }, { "epoch": 0.3932806324110672, "step": 199, "train_accuracy_recipient_token": 0.8571428571428571, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.7777777777777778, "train_accuracy_recipient_token_get_total": 9, "train_accuracy_recipient_token_order": 0.0, "train_accuracy_recipient_token_order_total": 1, "train_total_number_recipient_token": 21 }, { "epoch": 0.3932806324110672, "step": 199, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_ident": 0.0, "train_accuracy_recipient_token_ident_total": 2, "train_accuracy_recipient_token_translate": 0.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.3952569169960474, "grad_norm": 2.750136117135516, "learning_rate": 3.4530052449271044e-06, "loss": 0.2443, "step": 200 }, { "epoch": 0.3952569169960474, "step": 200, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_add": 1.0, "train_accuracy_recipient_token_add_total": 4, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_construct": 1.0, "train_accuracy_recipient_token_construct_total": 1, "train_accuracy_recipient_token_remove": 1.0, "train_accuracy_recipient_token_remove_total": 1, "train_accuracy_recipient_token_select": 1.0, "train_accuracy_recipient_token_select_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.3952569169960474, "step": 200, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.39723320158102765, "grad_norm": 2.4254009242612904, "learning_rate": 3.438167512198436e-06, "loss": 0.2524, "step": 201 }, { "epoch": 0.39723320158102765, "step": 201, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.39723320158102765, "step": 201, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compose": 1.0, "train_accuracy_recipient_token_compose_total": 1, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 1, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.39920948616600793, "grad_norm": 1.9697723841835133, "learning_rate": 3.4232912150778914e-06, "loss": 0.292, "step": 202 }, { "epoch": 0.39920948616600793, "step": 202, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.39920948616600793, "step": 202, "train_accuracy_recipient_token": 0.85, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 0.7777777777777778, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 20 }, { "epoch": 0.40118577075098816, "grad_norm": 1.9505870591219885, "learning_rate": 3.408376965071779e-06, "loss": 0.2301, "step": 203 }, { "epoch": 0.40118577075098816, "step": 203, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 20 }, { "epoch": 0.40118577075098816, "step": 203, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_convert": 1.0, "train_accuracy_recipient_token_convert_total": 1, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 17 }, { "epoch": 0.4031620553359684, "grad_norm": 2.259780090042466, "learning_rate": 3.393425375246503e-06, "loss": 0.2284, "step": 204 }, { "epoch": 0.4031620553359684, "step": 204, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.6, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_locate": 1.0, "train_accuracy_recipient_token_locate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.4031620553359684, "step": 204, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_invest": 1.0, "train_accuracy_recipient_token_invest_total": 1, "train_accuracy_recipient_token_tractor": 1.0, "train_accuracy_recipient_token_tractor_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.4051383399209486, "grad_norm": 1.654459204817194, "learning_rate": 3.3784370602033572e-06, "loss": 0.1279, "step": 205 }, { "epoch": 0.4051383399209486, "step": 205, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9166666666666666, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_detect": 1.0, "train_accuracy_recipient_token_detect_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 20 }, { "epoch": 0.4051383399209486, "step": 205, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_pay": 0.5, "train_accuracy_recipient_token_pay_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.40711462450592883, "grad_norm": 1.7599338692834208, "learning_rate": 3.3634126360532694e-06, "loss": 0.1629, "step": 206 }, { "epoch": 0.40711462450592883, "step": 206, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_make": 1.0, "train_accuracy_recipient_token_make_total": 4, "train_total_number_recipient_token": 17 }, { "epoch": 0.40711462450592883, "step": 206, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { "epoch": 0.4090909090909091, "grad_norm": 1.881346517215222, "learning_rate": 3.3483527203914694e-06, "loss": 0.2584, "step": 207 }, { "epoch": 0.4090909090909091, "step": 207, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_bird": 1.0, "train_accuracy_recipient_token_bird_total": 4, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_in": 1.0, "train_accuracy_recipient_token_in_total": 3, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.4090909090909091, "step": 207, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 18 }, { "epoch": 0.41106719367588934, "grad_norm": 2.0484839728339317, "learning_rate": 3.333257932272105e-06, "loss": 0.2313, "step": 208 }, { "epoch": 0.41106719367588934, "step": 208, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 17 }, { "epoch": 0.41106719367588934, "step": 208, "train_accuracy_recipient_token": 0.7777777777777778, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_pest": 0.0, "train_accuracy_recipient_token_pest_total": 1, "train_accuracy_recipient_token_sun": 1.0, "train_accuracy_recipient_token_sun_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.41304347826086957, "grad_norm": 2.473490384710613, "learning_rate": 3.3181288921827925e-06, "loss": 0.439, "step": 209 }, { "epoch": 0.41304347826086957, "step": 209, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_evaluate": 1.0, "train_accuracy_recipient_token_evaluate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_provide": 1.0, "train_accuracy_recipient_token_provide_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.41304347826086957, "step": 209, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 20 }, { "epoch": 0.4150197628458498, "grad_norm": 1.8750180775014957, "learning_rate": 3.3029662220191146e-06, "loss": 0.1752, "step": 210 }, { "epoch": 0.4150197628458498, "step": 210, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.4150197628458498, "step": 210, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 20 }, { "epoch": 0.41699604743083, "grad_norm": 2.2988992335970173, "learning_rate": 3.2877705450590525e-06, "loss": 0.2718, "step": 211 }, { "epoch": 0.41699604743083, "step": 211, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_ex": 1.0, "train_accuracy_recipient_token_ex_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_invest": 1.0, "train_accuracy_recipient_token_invest_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.41699604743083, "step": 211, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_locate": 1.0, "train_accuracy_recipient_token_locate_total": 1, "train_accuracy_recipient_token_p": 1.0, "train_accuracy_recipient_token_p_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.4189723320158103, "grad_norm": 2.0126487070423162, "learning_rate": 3.272542485937369e-06, "loss": 0.2419, "step": 212 }, { "epoch": 0.4189723320158103, "step": 212, "train_accuracy_recipient_token": 0.9523809523809523, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 2, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 21 }, { "epoch": 0.4189723320158103, "step": 212, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 3, "train_accuracy_recipient_token_verify": 1.0, "train_accuracy_recipient_token_verify_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.4209486166007905, "grad_norm": 2.030818509388425, "learning_rate": 3.2572826706199304e-06, "loss": 0.2074, "step": 213 }, { "epoch": 0.4209486166007905, "step": 213, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.4209486166007905, "step": 213, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_l": 1.0, "train_accuracy_recipient_token_l_total": 3, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.42292490118577075, "grad_norm": 1.9072544731043124, "learning_rate": 3.2419917263779765e-06, "loss": 0.235, "step": 214 }, { "epoch": 0.42292490118577075, "step": 214, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_event": 0.6666666666666666, "train_accuracy_recipient_token_event_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.42292490118577075, "step": 214, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_clean": 1.0, "train_accuracy_recipient_token_clean_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_paint": 0.5, "train_accuracy_recipient_token_paint_total": 2, "train_accuracy_recipient_token_remove": 1.0, "train_accuracy_recipient_token_remove_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.424901185770751, "grad_norm": 2.623452228952221, "learning_rate": 3.2266702817623348e-06, "loss": 0.2415, "step": 215 }, { "epoch": 0.424901185770751, "step": 215, "train_accuracy_recipient_token": 0.8, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 2, "train_accuracy_recipient_token_device": 1.0, "train_accuracy_recipient_token_device_total": 3, "train_accuracy_recipient_token_extract": 1.0, "train_accuracy_recipient_token_extract_total": 1, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 2, "train_total_number_recipient_token": 20 }, { "epoch": 0.424901185770751, "step": 215, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.4268774703557312, "grad_norm": 2.303051674442213, "learning_rate": 3.2113189665775812e-06, "loss": 0.2183, "step": 216 }, { "epoch": 0.4268774703557312, "step": 216, "train_accuracy_recipient_token": 0.7894736842105263, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_accuracy_recipient_token_translate": 0.75, "train_accuracy_recipient_token_translate_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.4268774703557312, "step": 216, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_opt": 1.0, "train_accuracy_recipient_token_opt_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.4288537549407115, "grad_norm": 1.626960528946486, "learning_rate": 3.1959384118561587e-06, "loss": 0.2717, "step": 217 }, { "epoch": 0.4288537549407115, "step": 217, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.4288537549407115, "step": 217, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_deploy": 1.0, "train_accuracy_recipient_token_deploy_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 17 }, { "epoch": 0.4308300395256917, "grad_norm": 1.985201264804771, "learning_rate": 3.180529249832428e-06, "loss": 0.1797, "step": 218 }, { "epoch": 0.4308300395256917, "step": 218, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_car": 1.0, "train_accuracy_recipient_token_car_total": 1, "train_accuracy_recipient_token_evaluate": 1.0, "train_accuracy_recipient_token_evaluate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.4308300395256917, "step": 218, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_accuracy_recipient_token_schedule": 0.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.43280632411067194, "grad_norm": 1.8348360216552162, "learning_rate": 3.165092113916688e-06, "loss": 0.2101, "step": 219 }, { "epoch": 0.43280632411067194, "step": 219, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 1, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_accuracy_recipient_token_stat": 1.0, "train_accuracy_recipient_token_stat_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.43280632411067194, "step": 219, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 20 }, { "epoch": 0.43478260869565216, "grad_norm": 1.9783954952399618, "learning_rate": 3.1496276386691327e-06, "loss": 0.2739, "step": 220 }, { "epoch": 0.43478260869565216, "step": 220, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.43478260869565216, "step": 220, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_ex": 1.0, "train_accuracy_recipient_token_ex_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_traffic": 1.0, "train_accuracy_recipient_token_traffic_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.4367588932806324, "grad_norm": 2.4700705976410786, "learning_rate": 3.1341364597737684e-06, "loss": 0.4093, "step": 221 }, { "epoch": 0.4367588932806324, "step": 221, "train_accuracy_recipient_token": 0.9523809523809523, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 21 }, { "epoch": 0.4367588932806324, "step": 221, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_arr": 1.0, "train_accuracy_recipient_token_arr_total": 1, "train_accuracy_recipient_token_book": 0.5, "train_accuracy_recipient_token_book_total": 2, "train_accuracy_recipient_token_hire": 1.0, "train_accuracy_recipient_token_hire_total": 1, "train_accuracy_recipient_token_plan": 1.0, "train_accuracy_recipient_token_plan_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.43873517786561267, "grad_norm": 2.0704854982802368, "learning_rate": 3.1186192140122863e-06, "loss": 0.2264, "step": 222 }, { "epoch": 0.43873517786561267, "step": 222, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.43873517786561267, "step": 222, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_stamp": 1.0, "train_accuracy_recipient_token_stamp_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.4407114624505929, "grad_norm": 1.654154021035547, "learning_rate": 3.1030765392378813e-06, "loss": 0.1653, "step": 223 }, { "epoch": 0.4407114624505929, "step": 223, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_display": 1.0, "train_accuracy_recipient_token_display_total": 1, "train_accuracy_recipient_token_find": 0.75, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.4407114624505929, "step": 223, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.4426877470355731, "grad_norm": 1.8368273428097381, "learning_rate": 3.0875090743490383e-06, "loss": 0.2253, "step": 224 }, { "epoch": 0.4426877470355731, "step": 224, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_s": 1.0, "train_accuracy_recipient_token_s_total": 2, "train_total_number_recipient_token": 20 }, { "epoch": 0.4426877470355731, "step": 224, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.44466403162055335, "grad_norm": 1.950504550895436, "learning_rate": 3.071917459263264e-06, "loss": 0.2345, "step": 225 }, { "epoch": 0.44466403162055335, "step": 225, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_dis": 1.0, "train_accuracy_recipient_token_dis_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 18 }, { "epoch": 0.44466403162055335, "step": 225, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.44664031620553357, "grad_norm": 2.1454666355505174, "learning_rate": 3.056302334890786e-06, "loss": 0.3298, "step": 226 }, { "epoch": 0.44664031620553357, "step": 226, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_stamp": 1.0, "train_accuracy_recipient_token_stamp_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.44664031620553357, "step": 226, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { "epoch": 0.44861660079051385, "grad_norm": 2.377176691255679, "learning_rate": 3.0406643431082088e-06, "loss": 0.2614, "step": 227 }, { "epoch": 0.44861660079051385, "step": 227, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_locate": 1.0, "train_accuracy_recipient_token_locate_total": 1, "train_accuracy_recipient_token_provide": 1.0, "train_accuracy_recipient_token_provide_total": 1, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.44861660079051385, "step": 227, "train_accuracy_recipient_token": 0.8125, "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.4505928853754941, "grad_norm": 1.5943078704191154, "learning_rate": 3.0250041267321234e-06, "loss": 0.148, "step": 228 }, { "epoch": 0.4505928853754941, "step": 228, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_configure": 1.0, "train_accuracy_recipient_token_configure_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_track": 0.5, "train_accuracy_recipient_token_track_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.4505928853754941, "step": 228, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.4525691699604743, "grad_norm": 2.166389899834986, "learning_rate": 3.009322329492689e-06, "loss": 0.1702, "step": 229 }, { "epoch": 0.4525691699604743, "step": 229, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_download": 1.0, "train_accuracy_recipient_token_download_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_rec": 1.0, "train_accuracy_recipient_token_rec_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.4525691699604743, "step": 229, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 3, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.45454545454545453, "grad_norm": 2.2170367848067176, "learning_rate": 2.993619596007168e-06, "loss": 0.2933, "step": 230 }, { "epoch": 0.45454545454545453, "step": 230, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.8333333333333334, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_describe": 1.0, "train_accuracy_recipient_token_describe_total": 1, "train_accuracy_recipient_token_find": 0.75, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_organ": 1.0, "train_accuracy_recipient_token_organ_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.45454545454545453, "step": 230, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 20 }, { "epoch": 0.45652173913043476, "grad_norm": 1.4984470237108378, "learning_rate": 2.9778965717534314e-06, "loss": 0.1063, "step": 231 }, { "epoch": 0.45652173913043476, "step": 231, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_f": 1.0, "train_accuracy_recipient_token_f_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.45652173913043476, "step": 231, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9166666666666666, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_m": 0.75, "train_accuracy_recipient_token_m_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.45849802371541504, "grad_norm": 2.713980767721741, "learning_rate": 2.9621539030434223e-06, "loss": 0.3257, "step": 232 }, { "epoch": 0.45849802371541504, "step": 232, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.875, "train_accuracy_recipient_token_all_total": 8, "train_accuracy_recipient_token_check": 0.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_histor": 1.0, "train_accuracy_recipient_token_histor_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.45849802371541504, "step": 232, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9166666666666666, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.46047430830039526, "grad_norm": 1.9449680707287624, "learning_rate": 2.946392236996592e-06, "loss": 0.1973, "step": 233 }, { "epoch": 0.46047430830039526, "step": 233, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8888888888888888, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_estimate": 0.5, "train_accuracy_recipient_token_estimate_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.46047430830039526, "step": 233, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9166666666666666, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_translate": 0.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.4624505928853755, "grad_norm": 2.428900846914706, "learning_rate": 2.9306122215132974e-06, "loss": 0.2668, "step": 234 }, { "epoch": 0.4624505928853755, "step": 234, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.4624505928853755, "step": 234, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 19 }, { "epoch": 0.4644268774703557, "grad_norm": 2.3084940546349375, "learning_rate": 2.91481450524817e-06, "loss": 0.2538, "step": 235 }, { "epoch": 0.4644268774703557, "step": 235, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 20 }, { "epoch": 0.4644268774703557, "step": 235, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_sum": 1.0, "train_accuracy_recipient_token_sum_total": 1, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.466403162055336, "grad_norm": 1.7162275334301005, "learning_rate": 2.8989997375834485e-06, "loss": 0.2411, "step": 236 }, { "epoch": 0.466403162055336, "step": 236, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_plan": 1.0, "train_accuracy_recipient_token_plan_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.466403162055336, "step": 236, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.3333333333333333, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_p": 1.0, "train_accuracy_recipient_token_p_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.4683794466403162, "grad_norm": 1.597659173981818, "learning_rate": 2.8831685686022897e-06, "loss": 0.1603, "step": 237 }, { "epoch": 0.4683794466403162, "step": 237, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.4683794466403162, "step": 237, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_shop": 1.0, "train_accuracy_recipient_token_shop_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.47035573122529645, "grad_norm": 1.9065156428525225, "learning_rate": 2.8673216490620453e-06, "loss": 0.1811, "step": 238 }, { "epoch": 0.47035573122529645, "step": 238, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.47035573122529645, "step": 238, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.4723320158102767, "grad_norm": 2.176658370568172, "learning_rate": 2.8514596303675073e-06, "loss": 0.1601, "step": 239 }, { "epoch": 0.4723320158102767, "step": 239, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_invest": 1.0, "train_accuracy_recipient_token_invest_total": 1, "train_accuracy_recipient_token_vehicle": 1.0, "train_accuracy_recipient_token_vehicle_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.4723320158102767, "step": 239, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 19 }, { "epoch": 0.4743083003952569, "grad_norm": 1.768442218413903, "learning_rate": 2.835583164544139e-06, "loss": 0.1682, "step": 240 }, { "epoch": 0.4743083003952569, "step": 240, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.4743083003952569, "step": 240, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 3, "train_accuracy_recipient_token_artist": 0.0, "train_accuracy_recipient_token_artist_total": 1, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_accuracy_recipient_token_water": 1.0, "train_accuracy_recipient_token_water_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.4762845849802372, "grad_norm": 1.725662364780488, "learning_rate": 2.8196929042112652e-06, "loss": 0.1773, "step": 241 }, { "epoch": 0.4762845849802372, "step": 241, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.4762845849802372, "step": 241, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.4782608695652174, "grad_norm": 2.278665078686508, "learning_rate": 2.8037895025552513e-06, "loss": 0.2275, "step": 242 }, { "epoch": 0.4782608695652174, "step": 242, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 0.5, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.4782608695652174, "step": 242, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 2, "train_total_number_recipient_token": 16 }, { "epoch": 0.48023715415019763, "grad_norm": 2.4836024942390087, "learning_rate": 2.787873613302649e-06, "loss": 0.1997, "step": 243 }, { "epoch": 0.48023715415019763, "step": 243, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_detect": 1.0, "train_accuracy_recipient_token_detect_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 2, "train_total_number_recipient_token": 21 }, { "epoch": 0.48023715415019763, "step": 243, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 0.5, "train_accuracy_recipient_token_calculate_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.48221343873517786, "grad_norm": 2.3999235961249066, "learning_rate": 2.7719458906933277e-06, "loss": 0.1463, "step": 244 }, { "epoch": 0.48221343873517786, "step": 244, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 5, "train_total_number_recipient_token": 18 }, { "epoch": 0.48221343873517786, "step": 244, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 3, "train_total_number_recipient_token": 20 }, { "epoch": 0.4841897233201581, "grad_norm": 1.7608203760764525, "learning_rate": 2.7560069894535783e-06, "loss": 0.2735, "step": 245 }, { "epoch": 0.4841897233201581, "step": 245, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_color": 1.0, "train_accuracy_recipient_token_color_total": 1, "train_accuracy_recipient_token_draw": 1.0, "train_accuracy_recipient_token_draw_total": 1, "train_accuracy_recipient_token_object": 1.0, "train_accuracy_recipient_token_object_total": 1, "train_accuracy_recipient_token_scene": 1.0, "train_accuracy_recipient_token_scene_total": 2, "train_total_number_recipient_token": 20 }, { "epoch": 0.4841897233201581, "step": 245, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.48616600790513836, "grad_norm": 1.6831661789344432, "learning_rate": 2.7400575647692046e-06, "loss": 0.1882, "step": 246 }, { "epoch": 0.48616600790513836, "step": 246, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_ins": 1.0, "train_accuracy_recipient_token_ins_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.48616600790513836, "step": 246, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_review": 1.0, "train_accuracy_recipient_token_review_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.4881422924901186, "grad_norm": 1.8826315438523655, "learning_rate": 2.724098272258584e-06, "loss": 0.2473, "step": 247 }, { "epoch": 0.4881422924901186, "step": 247, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_learn": 1.0, "train_accuracy_recipient_token_learn_total": 1, "train_accuracy_recipient_token_research": 1.0, "train_accuracy_recipient_token_research_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.4881422924901186, "step": 247, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_s": 1.0, "train_accuracy_recipient_token_s_total": 1, "train_accuracy_recipient_token_stat": 1.0, "train_accuracy_recipient_token_stat_total": 1, "train_total_number_recipient_token": 15 }, { "epoch": 0.4901185770750988, "grad_norm": 1.9713625520958284, "learning_rate": 2.7081297679457238e-06, "loss": 0.3234, "step": 248 }, { "epoch": 0.4901185770750988, "step": 248, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_st": 1.0, "train_accuracy_recipient_token_st_total": 3, "train_total_number_recipient_token": 20 }, { "epoch": 0.4901185770750988, "step": 248, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 2, "train_accuracy_recipient_token_flag": 1.0, "train_accuracy_recipient_token_flag_total": 1, "train_accuracy_recipient_token_pl": 0.0, "train_accuracy_recipient_token_pl_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.49209486166007904, "grad_norm": 2.4921370963085328, "learning_rate": 2.692152708233292e-06, "loss": 0.4432, "step": 249 }, { "epoch": 0.49209486166007904, "step": 249, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.49209486166007904, "step": 249, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 2, "train_accuracy_recipient_token_design": 1.0, "train_accuracy_recipient_token_design_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.49407114624505927, "grad_norm": 2.402509089657591, "learning_rate": 2.676167749875635e-06, "loss": 0.2032, "step": 250 }, { "epoch": 0.49407114624505927, "step": 250, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.49407114624505927, "step": 250, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 18 }, { "epoch": 0.49604743083003955, "grad_norm": 2.717753245253635, "learning_rate": 2.6601755499517826e-06, "loss": 0.315, "step": 251 }, { "epoch": 0.49604743083003955, "step": 251, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_compile": 1.0, "train_accuracy_recipient_token_compile_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.49604743083003955, "step": 251, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.4980237154150198, "grad_norm": 1.8526560236000496, "learning_rate": 2.6441767658384363e-06, "loss": 0.2212, "step": 252 }, { "epoch": 0.4980237154150198, "step": 252, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.8888888888888888, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_setup": 1.0, "train_accuracy_recipient_token_setup_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.4980237154150198, "step": 252, "train_accuracy_recipient_token": 0.75, "train_accuracy_recipient_token_all": 0.6666666666666666, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 12 }, { "epoch": 0.5, "grad_norm": 2.0997848100152594, "learning_rate": 2.628172055182948e-06, "loss": 0.188, "step": 253 }, { "epoch": 0.5, "step": 253, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.5, "step": 253, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.5019762845849802, "grad_norm": 2.296699723342699, "learning_rate": 2.6121620758762877e-06, "loss": 0.3644, "step": 254 }, { "epoch": 0.5019762845849802, "step": 254, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_camera": 1.0, "train_accuracy_recipient_token_camera_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_invest": 1.0, "train_accuracy_recipient_token_invest_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.5019762845849802, "step": 254, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.5039525691699605, "grad_norm": 1.7053263617954735, "learning_rate": 2.596147486025996e-06, "loss": 0.2601, "step": 255 }, { "epoch": 0.5039525691699605, "step": 255, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_m": 1.0, "train_accuracy_recipient_token_m_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.5039525691699605, "step": 255, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { "epoch": 0.5059288537549407, "grad_norm": 1.6057303693260692, "learning_rate": 2.580128943929139e-06, "loss": 0.1021, "step": 256 }, { "epoch": 0.5059288537549407, "step": 256, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_extract": 1.0, "train_accuracy_recipient_token_extract_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_recogn": 0.0, "train_accuracy_recipient_token_recogn_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.5059288537549407, "step": 256, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 21 }, { "epoch": 0.5079051383399209, "grad_norm": 2.160182965266891, "learning_rate": 2.564107108045239e-06, "loss": 0.3013, "step": 257 }, { "epoch": 0.5079051383399209, "step": 257, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.5079051383399209, "step": 257, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 5, "train_total_number_recipient_token": 18 }, { "epoch": 0.5098814229249012, "grad_norm": 2.540894830262886, "learning_rate": 2.5480826369692178e-06, "loss": 0.2537, "step": 258 }, { "epoch": 0.5098814229249012, "step": 258, "train_accuracy_recipient_token": 0.8666666666666667, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 15 }, { "epoch": 0.5098814229249012, "step": 258, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_histor": 0.5, "train_accuracy_recipient_token_histor_total": 2, "train_accuracy_recipient_token_organ": 1.0, "train_accuracy_recipient_token_organ_total": 1, "train_accuracy_recipient_token_public": 1.0, "train_accuracy_recipient_token_public_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.5118577075098815, "grad_norm": 1.6078898635903305, "learning_rate": 2.532056189404318e-06, "loss": 0.1394, "step": 259 }, { "epoch": 0.5118577075098815, "step": 259, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.5118577075098815, "step": 259, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_plant": 1.0, "train_accuracy_recipient_token_plant_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.5138339920948617, "grad_norm": 4.285683109324995, "learning_rate": 2.516028424135028e-06, "loss": 0.2524, "step": 260 }, { "epoch": 0.5138339920948617, "step": 260, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_evaluate": 1.0, "train_accuracy_recipient_token_evaluate_total": 1, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_plan": 1.0, "train_accuracy_recipient_token_plan_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.5138339920948617, "step": 260, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_update": 1.0, "train_accuracy_recipient_token_update_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.5158102766798419, "grad_norm": 1.4223559498348044, "learning_rate": 2.5e-06, "loss": 0.1352, "step": 261 }, { "epoch": 0.5158102766798419, "step": 261, "train_accuracy_recipient_token": 0.9333333333333333, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_int": 1.0, "train_accuracy_recipient_token_int_total": 1, "train_total_number_recipient_token": 15 }, { "epoch": 0.5158102766798419, "step": 261, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.5177865612648221, "grad_norm": 2.2628735436679763, "learning_rate": 2.4839715758649724e-06, "loss": 0.3725, "step": 262 }, { "epoch": 0.5177865612648221, "step": 262, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.5177865612648221, "step": 262, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_ca": 1.0, "train_accuracy_recipient_token_ca_total": 4, "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 20 }, { "epoch": 0.5197628458498024, "grad_norm": 2.2947519367621347, "learning_rate": 2.467943810595682e-06, "loss": 0.4472, "step": 263 }, { "epoch": 0.5197628458498024, "step": 263, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 2, "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 20 }, { "epoch": 0.5197628458498024, "step": 263, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 16 }, { "epoch": 0.5217391304347826, "grad_norm": 2.106955012004411, "learning_rate": 2.4519173630307826e-06, "loss": 0.3995, "step": 264 }, { "epoch": 0.5217391304347826, "step": 264, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 7, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.5217391304347826, "step": 264, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 1, "train_accuracy_recipient_token_draw": 1.0, "train_accuracy_recipient_token_draw_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 17 }, { "epoch": 0.5237154150197628, "grad_norm": 3.6280464608191956, "learning_rate": 2.4358928919547616e-06, "loss": 0.3171, "step": 265 }, { "epoch": 0.5237154150197628, "step": 265, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_phone": 1.0, "train_accuracy_recipient_token_phone_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.5237154150197628, "step": 265, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.8333333333333334, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_generate": 0.5, "train_accuracy_recipient_token_generate_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.525691699604743, "grad_norm": 1.7300477450689355, "learning_rate": 2.4198710560708623e-06, "loss": 0.2653, "step": 266 }, { "epoch": 0.525691699604743, "step": 266, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 0.5, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_sum": 1.0, "train_accuracy_recipient_token_sum_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.525691699604743, "step": 266, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compress": 1.0, "train_accuracy_recipient_token_compress_total": 1, "train_accuracy_recipient_token_crop": 1.0, "train_accuracy_recipient_token_crop_total": 1, "train_accuracy_recipient_token_d": 1.0, "train_accuracy_recipient_token_d_total": 2, "train_accuracy_recipient_token_enh": 1.0, "train_accuracy_recipient_token_enh_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.5276679841897233, "grad_norm": 2.345904530746849, "learning_rate": 2.403852513974004e-06, "loss": 0.2374, "step": 267 }, { "epoch": 0.5276679841897233, "step": 267, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_custom": 0.75, "train_accuracy_recipient_token_custom_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.5276679841897233, "step": 267, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 0.7142857142857143, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { "epoch": 0.5296442687747036, "grad_norm": 2.7558657907586137, "learning_rate": 2.3878379241237136e-06, "loss": 0.3184, "step": 268 }, { "epoch": 0.5296442687747036, "step": 268, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.5296442687747036, "step": 268, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 5, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 10 }, { "epoch": 0.5316205533596838, "grad_norm": 7.208775244573762, "learning_rate": 2.3718279448170527e-06, "loss": 0.8561, "step": 269 }, { "epoch": 0.5316205533596838, "step": 269, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 0.7142857142857143, "train_accuracy_recipient_token_get_total": 7, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.5316205533596838, "step": 269, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_describe": 1.0, "train_accuracy_recipient_token_describe_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.5335968379446641, "grad_norm": 2.6145328310474136, "learning_rate": 2.3558232341615645e-06, "loss": 0.2586, "step": 270 }, { "epoch": 0.5335968379446641, "step": 270, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { "epoch": 0.5335968379446641, "step": 270, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_ex": 1.0, "train_accuracy_recipient_token_ex_total": 3, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.5355731225296443, "grad_norm": 2.3232906112218323, "learning_rate": 2.339824450048218e-06, "loss": 0.2427, "step": 271 }, { "epoch": 0.5355731225296443, "step": 271, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_fish": 1.0, "train_accuracy_recipient_token_fish_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_purchase": 1.0, "train_accuracy_recipient_token_purchase_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.5355731225296443, "step": 271, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.9, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 20 }, { "epoch": 0.5375494071146245, "grad_norm": 2.208719311446786, "learning_rate": 2.323832250124365e-06, "loss": 0.2572, "step": 272 }, { "epoch": 0.5375494071146245, "step": 272, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.5375494071146245, "step": 272, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 2, "train_accuracy_recipient_token_extract": 1.0, "train_accuracy_recipient_token_extract_total": 1, "train_accuracy_recipient_token_monitor": 1.0, "train_accuracy_recipient_token_monitor_total": 1, "train_accuracy_recipient_token_recogn": 0.0, "train_accuracy_recipient_token_recogn_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.5395256916996047, "grad_norm": 1.8998607337146658, "learning_rate": 2.307847291766709e-06, "loss": 0.335, "step": 273 }, { "epoch": 0.5395256916996047, "step": 273, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_notify": 1.0, "train_accuracy_recipient_token_notify_total": 1, "train_accuracy_recipient_token_set": 1.0, "train_accuracy_recipient_token_set_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.5395256916996047, "step": 273, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.9166666666666666, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_s": 0.0, "train_accuracy_recipient_token_s_total": 1, "train_accuracy_recipient_token_vol": 1.0, "train_accuracy_recipient_token_vol_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.541501976284585, "grad_norm": 2.3240212915101277, "learning_rate": 2.291870232054277e-06, "loss": 0.2411, "step": 274 }, { "epoch": 0.541501976284585, "step": 274, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_book": 0.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.541501976284585, "step": 274, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_accuracy_recipient_token_schedule": 0.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.5434782608695652, "grad_norm": 2.1736866226516485, "learning_rate": 2.2759017277414165e-06, "loss": 0.2978, "step": 275 }, { "epoch": 0.5434782608695652, "step": 275, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.5434782608695652, "step": 275, "train_accuracy_recipient_token": 0.8, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 15 }, { "epoch": 0.5454545454545454, "grad_norm": 2.2088882652148123, "learning_rate": 2.2599424352307958e-06, "loss": 0.2148, "step": 276 }, { "epoch": 0.5454545454545454, "step": 276, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_log": 1.0, "train_accuracy_recipient_token_log_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.5454545454545454, "step": 276, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_interpret": 1.0, "train_accuracy_recipient_token_interpret_total": 1, "train_accuracy_recipient_token_stage": 1.0, "train_accuracy_recipient_token_stage_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.5474308300395256, "grad_norm": 1.9068759419051124, "learning_rate": 2.243993010546422e-06, "loss": 0.2597, "step": 277 }, { "epoch": 0.5474308300395256, "step": 277, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.5474308300395256, "step": 277, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.8333333333333334, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 3, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.549407114624506, "grad_norm": 1.8438582147731286, "learning_rate": 2.228054109306673e-06, "loss": 0.196, "step": 278 }, { "epoch": 0.549407114624506, "step": 278, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_update": 0.0, "train_accuracy_recipient_token_update_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.549407114624506, "step": 278, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_ident": 0.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.5513833992094862, "grad_norm": 1.7961818288907425, "learning_rate": 2.212126386697352e-06, "loss": 0.1863, "step": 279 }, { "epoch": 0.5513833992094862, "step": 279, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.5513833992094862, "step": 279, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.5533596837944664, "grad_norm": 2.0438685228831366, "learning_rate": 2.196210497444749e-06, "loss": 0.3187, "step": 280 }, { "epoch": 0.5533596837944664, "step": 280, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_describe": 1.0, "train_accuracy_recipient_token_describe_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.5533596837944664, "step": 280, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.5553359683794467, "grad_norm": 2.026195406795956, "learning_rate": 2.1803070957887348e-06, "loss": 0.219, "step": 281 }, { "epoch": 0.5553359683794467, "step": 281, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_convert": 1.0, "train_accuracy_recipient_token_convert_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 17 }, { "epoch": 0.5553359683794467, "step": 281, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { "epoch": 0.5573122529644269, "grad_norm": 2.4977490050823454, "learning_rate": 2.1644168354558623e-06, "loss": 0.1723, "step": 282 }, { "epoch": 0.5573122529644269, "step": 282, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_ex": 1.0, "train_accuracy_recipient_token_ex_total": 1, "train_accuracy_recipient_token_gest": 1.0, "train_accuracy_recipient_token_gest_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 20 }, { "epoch": 0.5573122529644269, "step": 282, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.7692307692307693, "train_accuracy_recipient_token_all_total": 13, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 18 }, { "epoch": 0.5592885375494071, "grad_norm": 2.362592143856781, "learning_rate": 2.1485403696324935e-06, "loss": 0.3041, "step": 283 }, { "epoch": 0.5592885375494071, "step": 283, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_add": 1.0, "train_accuracy_recipient_token_add_total": 1, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 1, "train_accuracy_recipient_token_update": 1.0, "train_accuracy_recipient_token_update_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.5592885375494071, "step": 283, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_product": 0.75, "train_accuracy_recipient_token_product_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.5612648221343873, "grad_norm": 2.234906036510673, "learning_rate": 2.1326783509379555e-06, "loss": 0.1854, "step": 284 }, { "epoch": 0.5612648221343873, "step": 284, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 4, "train_accuracy_recipient_token_hang": 1.0, "train_accuracy_recipient_token_hang_total": 1, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_accuracy_recipient_token_remove": 1.0, "train_accuracy_recipient_token_remove_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.5612648221343873, "step": 284, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_schedule": 0.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.5632411067193676, "grad_norm": 2.0687918980973863, "learning_rate": 2.11683143139771e-06, "loss": 0.1854, "step": 285 }, { "epoch": 0.5632411067193676, "step": 285, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.875, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.5632411067193676, "step": 285, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.5652173913043478, "grad_norm": 3.474374141128837, "learning_rate": 2.1010002624165528e-06, "loss": 0.4115, "step": 286 }, { "epoch": 0.5652173913043478, "step": 286, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.5652173913043478, "step": 286, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compose": 1.0, "train_accuracy_recipient_token_compose_total": 1, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 2, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.567193675889328, "grad_norm": 2.803107135984778, "learning_rate": 2.085185494751831e-06, "loss": 0.2326, "step": 287 }, { "epoch": 0.567193675889328, "step": 287, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 2, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.567193675889328, "step": 287, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 2, "train_accuracy_recipient_token_draw": 1.0, "train_accuracy_recipient_token_draw_total": 4, "train_accuracy_recipient_token_s": 1.0, "train_accuracy_recipient_token_s_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.5691699604743083, "grad_norm": 2.1007745780069653, "learning_rate": 2.069387778486703e-06, "loss": 0.2901, "step": 288 }, { "epoch": 0.5691699604743083, "step": 288, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 16 }, { "epoch": 0.5691699604743083, "step": 288, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 5, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.5711462450592886, "grad_norm": 2.5164590329084304, "learning_rate": 2.053607763003409e-06, "loss": 0.1953, "step": 289 }, { "epoch": 0.5711462450592886, "step": 289, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8333333333333334, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 17 }, { "epoch": 0.5711462450592886, "step": 289, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_g": 1.0, "train_accuracy_recipient_token_g_total": 3, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.5731225296442688, "grad_norm": 2.1664077194429154, "learning_rate": 2.037846096956578e-06, "loss": 0.2971, "step": 290 }, { "epoch": 0.5731225296442688, "step": 290, "train_accuracy_recipient_token": 0.9333333333333333, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 1, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 15 }, { "epoch": 0.5731225296442688, "step": 290, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_energy": 1.0, "train_accuracy_recipient_token_energy_total": 1, "train_accuracy_recipient_token_lands": 1.0, "train_accuracy_recipient_token_lands_total": 1, "train_accuracy_recipient_token_s": 1.0, "train_accuracy_recipient_token_s_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.575098814229249, "grad_norm": 2.7683955116401697, "learning_rate": 2.02210342824657e-06, "loss": 0.3624, "step": 291 }, { "epoch": 0.575098814229249, "step": 291, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_accuracy_recipient_token_sort": 1.0, "train_accuracy_recipient_token_sort_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.575098814229249, "step": 291, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.5770750988142292, "grad_norm": 1.8345301561689498, "learning_rate": 2.0063804039928324e-06, "loss": 0.1363, "step": 292 }, { "epoch": 0.5770750988142292, "step": 292, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.5770750988142292, "step": 292, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 19 }, { "epoch": 0.5790513833992095, "grad_norm": 1.893006363064422, "learning_rate": 1.9906776705073113e-06, "loss": 0.1679, "step": 293 }, { "epoch": 0.5790513833992095, "step": 293, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { "epoch": 0.5790513833992095, "step": 293, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 1, "train_accuracy_recipient_token_stat": 1.0, "train_accuracy_recipient_token_stat_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.5810276679841897, "grad_norm": 1.8414367691181166, "learning_rate": 1.9749958732678766e-06, "loss": 0.197, "step": 294 }, { "epoch": 0.5810276679841897, "step": 294, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.5810276679841897, "step": 294, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8888888888888888, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.5830039525691699, "grad_norm": 2.06959947325046, "learning_rate": 1.9593356568917916e-06, "loss": 0.1392, "step": 295 }, { "epoch": 0.5830039525691699, "step": 295, "train_accuracy_recipient_token": 0.8571428571428571, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_monitor": 1.0, "train_accuracy_recipient_token_monitor_total": 1, "train_total_number_recipient_token": 14 }, { "epoch": 0.5830039525691699, "step": 295, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_construction": 1.0, "train_accuracy_recipient_token_construction_total": 2, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.5849802371541502, "grad_norm": 2.4681273751326116, "learning_rate": 1.9436976651092143e-06, "loss": 0.2056, "step": 296 }, { "epoch": 0.5849802371541502, "step": 296, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 17 }, { "epoch": 0.5849802371541502, "step": 296, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.5869565217391305, "grad_norm": 1.6965670348460065, "learning_rate": 1.928082540736737e-06, "loss": 0.1315, "step": 297 }, { "epoch": 0.5869565217391305, "step": 297, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.8, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_make": 1.0, "train_accuracy_recipient_token_make_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.5869565217391305, "step": 297, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_estimate": 0.5, "train_accuracy_recipient_token_estimate_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.5889328063241107, "grad_norm": 2.4962220449528836, "learning_rate": 1.912490925650962e-06, "loss": 0.1743, "step": 298 }, { "epoch": 0.5889328063241107, "step": 298, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 0.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_compile": 1.0, "train_accuracy_recipient_token_compile_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 20 }, { "epoch": 0.5889328063241107, "step": 298, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.875, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.5909090909090909, "grad_norm": 1.7917091655168156, "learning_rate": 1.8969234607621187e-06, "loss": 0.1669, "step": 299 }, { "epoch": 0.5909090909090909, "step": 299, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_accuracy_recipient_token_s": 1.0, "train_accuracy_recipient_token_s_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.5909090909090909, "step": 299, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_building": 1.0, "train_accuracy_recipient_token_building_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.5928853754940712, "grad_norm": 1.5595949160601696, "learning_rate": 1.8813807859877147e-06, "loss": 0.1331, "step": 300 }, { "epoch": 0.5928853754940712, "step": 300, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_air": 1.0, "train_accuracy_recipient_token_air_total": 1, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 16 }, { "epoch": 0.5928853754940712, "step": 300, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_average": 1.0, "train_accuracy_recipient_token_average_total": 1, "train_accuracy_recipient_token_compare": 0.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.5948616600790514, "grad_norm": 2.413658770037087, "learning_rate": 1.865863540226232e-06, "loss": 0.249, "step": 301 }, { "epoch": 0.5948616600790514, "step": 301, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_attach": 1.0, "train_accuracy_recipient_token_attach_total": 1, "train_accuracy_recipient_token_clean": 1.0, "train_accuracy_recipient_token_clean_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_remove": 1.0, "train_accuracy_recipient_token_remove_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.5948616600790514, "step": 301, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.5968379446640316, "grad_norm": 1.981013163078487, "learning_rate": 1.8503723613308683e-06, "loss": 0.2832, "step": 302 }, { "epoch": 0.5968379446640316, "step": 302, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_ext": 0.0, "train_accuracy_recipient_token_ext_total": 1, "train_accuracy_recipient_token_fire": 1.0, "train_accuracy_recipient_token_fire_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 20 }, { "epoch": 0.5968379446640316, "step": 302, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_d": 1.0, "train_accuracy_recipient_token_d_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.5988142292490118, "grad_norm": 1.8168717362636846, "learning_rate": 1.8349078860833125e-06, "loss": 0.1964, "step": 303 }, { "epoch": 0.5988142292490118, "step": 303, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 2, "train_accuracy_recipient_token_plant": 1.0, "train_accuracy_recipient_token_plant_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.5988142292490118, "step": 303, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_convert": 1.0, "train_accuracy_recipient_token_convert_total": 1, "train_accuracy_recipient_token_create": 0.0, "train_accuracy_recipient_token_create_total": 1, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.6007905138339921, "grad_norm": 1.9313706363255394, "learning_rate": 1.8194707501675724e-06, "loss": 0.2054, "step": 304 }, { "epoch": 0.6007905138339921, "step": 304, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_measure": 1.0, "train_accuracy_recipient_token_measure_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.6007905138339921, "step": 304, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.6027667984189723, "grad_norm": 2.503698222924907, "learning_rate": 1.8040615881438425e-06, "loss": 0.4178, "step": 305 }, { "epoch": 0.6027667984189723, "step": 305, "train_accuracy_recipient_token": 0.8125, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_d": 1.0, "train_accuracy_recipient_token_d_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 16 }, { "epoch": 0.6027667984189723, "step": 305, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9166666666666666, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_bridge": 1.0, "train_accuracy_recipient_token_bridge_total": 3, "train_accuracy_recipient_token_find": 0.75, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_invest": 1.0, "train_accuracy_recipient_token_invest_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.6047430830039525, "grad_norm": 2.0299144209716946, "learning_rate": 1.7886810334224192e-06, "loss": 0.2902, "step": 306 }, { "epoch": 0.6047430830039525, "step": 306, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 3, "train_accuracy_recipient_token_track": 1.0, "train_accuracy_recipient_token_track_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.6047430830039525, "step": 306, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.6067193675889329, "grad_norm": 1.5819437624234434, "learning_rate": 1.7733297182376663e-06, "loss": 0.1652, "step": 307 }, { "epoch": 0.6067193675889329, "step": 307, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_book": 0.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.6067193675889329, "step": 307, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_organ": 1.0, "train_accuracy_recipient_token_organ_total": 1, "train_accuracy_recipient_token_review": 0.5, "train_accuracy_recipient_token_review_total": 2, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.6086956521739131, "grad_norm": 1.805001545693983, "learning_rate": 1.7580082736220235e-06, "loss": 0.1703, "step": 308 }, { "epoch": 0.6086956521739131, "step": 308, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_parse": 0.5, "train_accuracy_recipient_token_parse_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.6086956521739131, "step": 308, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_convert": 1.0, "train_accuracy_recipient_token_convert_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.6106719367588933, "grad_norm": 2.3514653723969947, "learning_rate": 1.74271732938007e-06, "loss": 0.266, "step": 309 }, { "epoch": 0.6106719367588933, "step": 309, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 0.875, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.6106719367588933, "step": 309, "train_accuracy_recipient_token": 0.9047619047619048, "train_accuracy_recipient_token_all": 0.9285714285714286, "train_accuracy_recipient_token_all_total": 14, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 21 }, { "epoch": 0.6126482213438735, "grad_norm": 1.995728274275021, "learning_rate": 1.7274575140626318e-06, "loss": 0.1586, "step": 310 }, { "epoch": 0.6126482213438735, "step": 310, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_average": 1.0, "train_accuracy_recipient_token_average_total": 1, "train_accuracy_recipient_token_cat": 0.0, "train_accuracy_recipient_token_cat_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.6126482213438735, "step": 310, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.7857142857142857, "train_accuracy_recipient_token_all_total": 14, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.6146245059288538, "grad_norm": 2.072925082708649, "learning_rate": 1.7122294549409486e-06, "loss": 0.2363, "step": 311 }, { "epoch": 0.6146245059288538, "step": 311, "train_accuracy_recipient_token": 0.8, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 15 }, { "epoch": 0.6146245059288538, "step": 311, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_design": 1.0, "train_accuracy_recipient_token_design_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_monitor": 1.0, "train_accuracy_recipient_token_monitor_total": 1, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.616600790513834, "grad_norm": 2.2383854312267015, "learning_rate": 1.6970337779808862e-06, "loss": 0.2371, "step": 312 }, { "epoch": 0.616600790513834, "step": 312, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 0.5, "train_accuracy_recipient_token_calculate_total": 2, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_schedule": 0.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.616600790513834, "step": 312, "train_accuracy_recipient_token": 0.8125, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 16 }, { "epoch": 0.6185770750988142, "grad_norm": 1.9179139741219748, "learning_rate": 1.681871107817208e-06, "loss": 0.2351, "step": 313 }, { "epoch": 0.6185770750988142, "step": 313, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 15, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.6185770750988142, "step": 313, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_p": 1.0, "train_accuracy_recipient_token_p_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.6205533596837944, "grad_norm": 2.832295500440934, "learning_rate": 1.6667420677278962e-06, "loss": 0.2569, "step": 314 }, { "epoch": 0.6205533596837944, "step": 314, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.6205533596837944, "step": 314, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_locate": 1.0, "train_accuracy_recipient_token_locate_total": 1, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.6225296442687747, "grad_norm": 1.4578028303646424, "learning_rate": 1.6516472796085314e-06, "loss": 0.1809, "step": 315 }, { "epoch": 0.6225296442687747, "step": 315, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.6225296442687747, "step": 315, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.6245059288537549, "grad_norm": 2.1140362979737986, "learning_rate": 1.6365873639467314e-06, "loss": 0.2706, "step": 316 }, { "epoch": 0.6245059288537549, "step": 316, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.6245059288537549, "step": 316, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.6264822134387352, "grad_norm": 2.2633287960322597, "learning_rate": 1.6215629397966432e-06, "loss": 0.31, "step": 317 }, { "epoch": 0.6264822134387352, "step": 317, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_interpret": 1.0, "train_accuracy_recipient_token_interpret_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.6264822134387352, "step": 317, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.6284584980237155, "grad_norm": 2.5410331978041145, "learning_rate": 1.6065746247534986e-06, "loss": 0.23, "step": 318 }, { "epoch": 0.6284584980237155, "step": 318, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.6284584980237155, "step": 318, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 19 }, { "epoch": 0.6304347826086957, "grad_norm": 1.6183376573109411, "learning_rate": 1.5916230349282215e-06, "loss": 0.1793, "step": 319 }, { "epoch": 0.6304347826086957, "step": 319, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 0.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 0.75, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_locate": 0.0, "train_accuracy_recipient_token_locate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.6304347826086957, "step": 319, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_sum": 0.5, "train_accuracy_recipient_token_sum_total": 2, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.6324110671936759, "grad_norm": 1.526229611696862, "learning_rate": 1.5767087849221097e-06, "loss": 0.2264, "step": 320 }, { "epoch": 0.6324110671936759, "step": 320, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_register": 1.0, "train_accuracy_recipient_token_register_total": 1, "train_accuracy_recipient_token_verify": 1.0, "train_accuracy_recipient_token_verify_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.6324110671936759, "step": 320, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_filter": 1.0, "train_accuracy_recipient_token_filter_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_accuracy_recipient_token_ident": 0.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.6343873517786561, "grad_norm": 1.6433946920394795, "learning_rate": 1.561832487801565e-06, "loss": 0.199, "step": 321 }, { "epoch": 0.6343873517786561, "step": 321, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.6343873517786561, "step": 321, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_boat": 1.0, "train_accuracy_recipient_token_boat_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_s": 1.0, "train_accuracy_recipient_token_s_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.6363636363636364, "grad_norm": 1.9498159113092581, "learning_rate": 1.546994755072896e-06, "loss": 0.1708, "step": 322 }, { "epoch": 0.6363636363636364, "step": 322, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_h": 0.6666666666666666, "train_accuracy_recipient_token_h_total": 3, "train_accuracy_recipient_token_invest": 1.0, "train_accuracy_recipient_token_invest_total": 1, "train_accuracy_recipient_token_plane": 1.0, "train_accuracy_recipient_token_plane_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.6363636363636364, "step": 322, "train_accuracy_recipient_token": 0.6875, "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_update": 0.0, "train_accuracy_recipient_token_update_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.6383399209486166, "grad_norm": 2.3482135262198, "learning_rate": 1.5321961966571801e-06, "loss": 0.221, "step": 323 }, { "epoch": 0.6383399209486166, "step": 323, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_evaluate": 0.0, "train_accuracy_recipient_token_evaluate_total": 1, "train_accuracy_recipient_token_find": 0.3333333333333333, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_pack": 1.0, "train_accuracy_recipient_token_pack_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.6383399209486166, "step": 323, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_ind": 1.0, "train_accuracy_recipient_token_ind_total": 4, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.6403162055335968, "grad_norm": 2.265690935530158, "learning_rate": 1.5174374208651913e-06, "loss": 0.2203, "step": 324 }, { "epoch": 0.6403162055335968, "step": 324, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_arch": 1.0, "train_accuracy_recipient_token_arch_total": 1, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { "epoch": 0.6403162055335968, "step": 324, "train_accuracy_recipient_token": 0.8, "train_accuracy_recipient_token_all": 0.7692307692307693, "train_accuracy_recipient_token_all_total": 13, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 20 }, { "epoch": 0.642292490118577, "grad_norm": 1.6543940069908076, "learning_rate": 1.502719034372396e-06, "loss": 0.2158, "step": 325 }, { "epoch": 0.642292490118577, "step": 325, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.875, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 20 }, { "epoch": 0.642292490118577, "step": 325, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_compare": 0.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.6442687747035574, "grad_norm": 1.915721888237991, "learning_rate": 1.4880416421940155e-06, "loss": 0.3389, "step": 326 }, { "epoch": 0.6442687747035574, "step": 326, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.6442687747035574, "step": 326, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.8888888888888888, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_ins": 0.0, "train_accuracy_recipient_token_ins_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.6462450592885376, "grad_norm": 2.057668318543242, "learning_rate": 1.4734058476601553e-06, "loss": 0.1014, "step": 327 }, { "epoch": 0.6462450592885376, "step": 327, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_language": 1.0, "train_accuracy_recipient_token_language_total": 1, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.6462450592885376, "step": 327, "train_accuracy_recipient_token": 0.7647058823529411, "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.6482213438735178, "grad_norm": 1.286134009298201, "learning_rate": 1.4588122523910032e-06, "loss": 0.1612, "step": 328 }, { "epoch": 0.6482213438735178, "step": 328, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.6482213438735178, "step": 328, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.650197628458498, "grad_norm": 1.7384058205276167, "learning_rate": 1.4442614562721011e-06, "loss": 0.175, "step": 329 }, { "epoch": 0.650197628458498, "step": 329, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_ex": 1.0, "train_accuracy_recipient_token_ex_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 20 }, { "epoch": 0.650197628458498, "step": 329, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_b": 1.0, "train_accuracy_recipient_token_b_total": 1, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.6521739130434783, "grad_norm": 2.0702750470665947, "learning_rate": 1.429754057429687e-06, "loss": 0.1972, "step": 330 }, { "epoch": 0.6521739130434783, "step": 330, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_ins": 1.0, "train_accuracy_recipient_token_ins_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.6521739130434783, "step": 330, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 16 }, { "epoch": 0.6541501976284585, "grad_norm": 3.148476858942437, "learning_rate": 1.415290652206105e-06, "loss": 0.3255, "step": 331 }, { "epoch": 0.6541501976284585, "step": 331, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 19 }, { "epoch": 0.6541501976284585, "step": 331, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_invest": 1.0, "train_accuracy_recipient_token_invest_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.6561264822134387, "grad_norm": 1.4492697108896657, "learning_rate": 1.400871835135295e-06, "loss": 0.1334, "step": 332 }, { "epoch": 0.6561264822134387, "step": 332, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_load": 1.0, "train_accuracy_recipient_token_load_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.6561264822134387, "step": 332, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.658102766798419, "grad_norm": 1.8406241806364723, "learning_rate": 1.386498198918352e-06, "loss": 0.2765, "step": 333 }, { "epoch": 0.658102766798419, "step": 333, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_device": 1.0, "train_accuracy_recipient_token_device_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.658102766798419, "step": 333, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.9, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 20 }, { "epoch": 0.6600790513833992, "grad_norm": 1.716219108289217, "learning_rate": 1.3721703343991634e-06, "loss": 0.0999, "step": 334 }, { "epoch": 0.6600790513833992, "step": 334, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_park": 1.0, "train_accuracy_recipient_token_park_total": 2, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.6600790513833992, "step": 334, "train_accuracy_recipient_token": 0.8666666666666667, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 0.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 15 }, { "epoch": 0.6620553359683794, "grad_norm": 2.70419172920364, "learning_rate": 1.3578888305401208e-06, "loss": 0.2241, "step": 335 }, { "epoch": 0.6620553359683794, "step": 335, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_ex": 1.0, "train_accuracy_recipient_token_ex_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_ident": 0.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_translate": 0.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.6620553359683794, "step": 335, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_local": 1.0, "train_accuracy_recipient_token_local_total": 1, "train_accuracy_recipient_token_plan": 1.0, "train_accuracy_recipient_token_plan_total": 1, "train_accuracy_recipient_token_restaurant": 0.5, "train_accuracy_recipient_token_restaurant_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.6640316205533597, "grad_norm": 2.4216736231197733, "learning_rate": 1.3436542743979125e-06, "loss": 0.3036, "step": 336 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 0.9285714285714286, "train_accuracy_recipient_token_all": 0.875, "train_accuracy_recipient_token_all_total": 8, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 14 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.8, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_ident": 0.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 2, "train_total_number_recipient_token": 16 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_histor": 1.0, "train_accuracy_recipient_token_histor_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 17 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_interpret": 1.0, "train_accuracy_recipient_token_interpret_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.8666666666666667, "train_accuracy_recipient_token_all_total": 15, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_tour": 0.75, "train_accuracy_recipient_token_tour_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 0.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 19 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 17 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_hang": 1.0, "train_accuracy_recipient_token_hang_total": 1, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_accuracy_recipient_token_remove": 1.0, "train_accuracy_recipient_token_remove_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_accuracy_recipient_token_invest": 1.0, "train_accuracy_recipient_token_invest_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_add": 1.0, "train_accuracy_recipient_token_add_total": 1, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_construct": 1.0, "train_accuracy_recipient_token_construct_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_select": 1.0, "train_accuracy_recipient_token_select_total": 2, "train_total_number_recipient_token": 16 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_histor": 0.0, "train_accuracy_recipient_token_histor_total": 1, "train_accuracy_recipient_token_m": 1.0, "train_accuracy_recipient_token_m_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_gr": 1.0, "train_accuracy_recipient_token_gr_total": 3, "train_accuracy_recipient_token_ident": 0.5, "train_accuracy_recipient_token_ident_total": 2, "train_total_number_recipient_token": 20 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 0.9285714285714286, "train_accuracy_recipient_token_all": 0.8888888888888888, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_draw": 1.0, "train_accuracy_recipient_token_draw_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 14 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_review": 0.5, "train_accuracy_recipient_token_review_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_create": 0.5, "train_accuracy_recipient_token_create_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 0.875, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.8888888888888888, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_custom": 1.0, "train_accuracy_recipient_token_custom_total": 4, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 19 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_eng": 1.0, "train_accuracy_recipient_token_eng_total": 2, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 1, "train_accuracy_recipient_token_lock": 1.0, "train_accuracy_recipient_token_lock_total": 1, "train_accuracy_recipient_token_tem": 1.0, "train_accuracy_recipient_token_tem_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 16 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_car": 1.0, "train_accuracy_recipient_token_car_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_invest": 1.0, "train_accuracy_recipient_token_invest_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_bird": 1.0, "train_accuracy_recipient_token_bird_total": 2, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_histor": 1.0, "train_accuracy_recipient_token_histor_total": 1, "train_accuracy_recipient_token_symbol": 1.0, "train_accuracy_recipient_token_symbol_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 0.75, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_organ": 1.0, "train_accuracy_recipient_token_organ_total": 1, "train_accuracy_recipient_token_rent": 1.0, "train_accuracy_recipient_token_rent_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_predict": 1.0, "train_accuracy_recipient_token_predict_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_evaluate": 0.0, "train_accuracy_recipient_token_evaluate_total": 1, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_locate": 0.0, "train_accuracy_recipient_token_locate_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_park": 1.0, "train_accuracy_recipient_token_park_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 3, "train_accuracy_recipient_token_artist": 0.0, "train_accuracy_recipient_token_artist_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 2, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_object": 0.6666666666666666, "train_accuracy_recipient_token_object_total": 3, "train_accuracy_recipient_token_transfer": 1.0, "train_accuracy_recipient_token_transfer_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_fruit": 1.0, "train_accuracy_recipient_token_fruit_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_organ": 1.0, "train_accuracy_recipient_token_organ_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_interpret": 1.0, "train_accuracy_recipient_token_interpret_total": 1, "train_accuracy_recipient_token_stage": 1.0, "train_accuracy_recipient_token_stage_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_rent": 1.0, "train_accuracy_recipient_token_rent_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 18 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 5, "train_accuracy_recipient_token_f": 0.0, "train_accuracy_recipient_token_f_total": 1, "train_accuracy_recipient_token_pest": 1.0, "train_accuracy_recipient_token_pest_total": 1, "train_accuracy_recipient_token_plant": 1.0, "train_accuracy_recipient_token_plant_total": 1, "train_accuracy_recipient_token_sun": 1.0, "train_accuracy_recipient_token_sun_total": 1, "train_total_number_recipient_token": 9 }, { "epoch": 0.6640316205533597, "eval_loss": 0.1947811394929886, "eval_runtime": 411.4932, "eval_samples_per_second": 0.245, "eval_steps_per_second": 0.124, "step": 336 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_display": 1.0, "train_accuracy_recipient_token_display_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_s": 1.0, "train_accuracy_recipient_token_s_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.6640316205533597, "step": 336, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_class": 0.0, "train_accuracy_recipient_token_class_total": 1, "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 1, "train_accuracy_recipient_token_g": 1.0, "train_accuracy_recipient_token_g_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.66600790513834, "grad_norm": 1.8225855613190758, "learning_rate": 1.3294672510993862e-06, "loss": 0.173, "step": 337 }, { "epoch": 0.66600790513834, "step": 337, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.66600790513834, "step": 337, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 19 }, { "epoch": 0.6679841897233202, "grad_norm": 1.5905272243219477, "learning_rate": 1.3153283438175036e-06, "loss": 0.2181, "step": 338 }, { "epoch": 0.6679841897233202, "step": 338, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_ch": 1.0, "train_accuracy_recipient_token_ch_total": 1, "train_accuracy_recipient_token_ex": 0.0, "train_accuracy_recipient_token_ex_total": 1, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.6679841897233202, "step": 338, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compose": 1.0, "train_accuracy_recipient_token_compose_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.6699604743083004, "grad_norm": 1.7999227337628305, "learning_rate": 1.3012381337473656e-06, "loss": 0.3063, "step": 339 }, { "epoch": 0.6699604743083004, "step": 339, "train_accuracy_recipient_token": 0.8, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_configure": 0.0, "train_accuracy_recipient_token_configure_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 15 }, { "epoch": 0.6699604743083004, "step": 339, "train_accuracy_recipient_token": 0.8571428571428571, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 21 }, { "epoch": 0.6719367588932806, "grad_norm": 1.6464971197164584, "learning_rate": 1.2871972000823197e-06, "loss": 0.1449, "step": 340 }, { "epoch": 0.6719367588932806, "step": 340, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 20 }, { "epoch": 0.6719367588932806, "step": 340, "train_accuracy_recipient_token": 0.7777777777777778, "train_accuracy_recipient_token_all": 0.6666666666666666, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.6739130434782609, "grad_norm": 2.125692840238993, "learning_rate": 1.2732061199901563e-06, "loss": 0.2194, "step": 341 }, { "epoch": 0.6739130434782609, "step": 341, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_d": 1.0, "train_accuracy_recipient_token_d_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 14 }, { "epoch": 0.6739130434782609, "step": 341, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_custom": 1.0, "train_accuracy_recipient_token_custom_total": 1, "train_accuracy_recipient_token_draw": 1.0, "train_accuracy_recipient_token_draw_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.6758893280632411, "grad_norm": 2.5717224194827053, "learning_rate": 1.2592654685893757e-06, "loss": 0.4011, "step": 342 }, { "epoch": 0.6758893280632411, "step": 342, "train_accuracy_recipient_token": 0.7894736842105263, "train_accuracy_recipient_token_all": 0.7857142857142857, "train_accuracy_recipient_token_all_total": 14, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_event": 0.0, "train_accuracy_recipient_token_event_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.6758893280632411, "step": 342, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 4, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.6778656126482213, "grad_norm": 1.944985236425837, "learning_rate": 1.2453758189255568e-06, "loss": 0.2952, "step": 343 }, { "epoch": 0.6778656126482213, "step": 343, "train_accuracy_recipient_token": 0.8571428571428571, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_custom": 1.0, "train_accuracy_recipient_token_custom_total": 1, "train_accuracy_recipient_token_remove": 1.0, "train_accuracy_recipient_token_remove_total": 1, "train_accuracy_recipient_token_translate": 0.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 14 }, { "epoch": 0.6778656126482213, "step": 343, "train_accuracy_recipient_token": 0.8666666666666667, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_c": 0.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 15 }, { "epoch": 0.6798418972332015, "grad_norm": 2.939817997079809, "learning_rate": 1.231537741947795e-06, "loss": 0.5102, "step": 344 }, { "epoch": 0.6798418972332015, "step": 344, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 20 }, { "epoch": 0.6798418972332015, "step": 344, "train_accuracy_recipient_token": 0.8, "train_accuracy_recipient_token_all": 0.8333333333333334, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.6818181818181818, "grad_norm": 1.8184116146817706, "learning_rate": 1.217751806485235e-06, "loss": 0.184, "step": 345 }, { "epoch": 0.6818181818181818, "step": 345, "train_accuracy_recipient_token": 0.9285714285714286, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 14 }, { "epoch": 0.6818181818181818, "step": 345, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.6837944664031621, "grad_norm": 1.8206375079632138, "learning_rate": 1.2040185792236874e-06, "loss": 0.2078, "step": 346 }, { "epoch": 0.6837944664031621, "step": 346, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.6837944664031621, "step": 346, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.6857707509881423, "grad_norm": 2.216136798863525, "learning_rate": 1.1903386246823363e-06, "loss": 0.2829, "step": 347 }, { "epoch": 0.6857707509881423, "step": 347, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8666666666666667, "train_accuracy_recipient_token_all_total": 15, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.6857707509881423, "step": 347, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.6877470355731226, "grad_norm": 2.9404057405898345, "learning_rate": 1.1767125051905314e-06, "loss": 0.365, "step": 348 }, { "epoch": 0.6877470355731226, "step": 348, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_draw": 1.0, "train_accuracy_recipient_token_draw_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_ins": 1.0, "train_accuracy_recipient_token_ins_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.6877470355731226, "step": 348, "train_accuracy_recipient_token": 0.9333333333333333, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_dec": 0.0, "train_accuracy_recipient_token_dec_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_verify": 1.0, "train_accuracy_recipient_token_verify_total": 1, "train_total_number_recipient_token": 15 }, { "epoch": 0.6897233201581028, "grad_norm": 1.6928281922145976, "learning_rate": 1.1631407808646758e-06, "loss": 0.2359, "step": 349 }, { "epoch": 0.6897233201581028, "step": 349, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_convert": 1.0, "train_accuracy_recipient_token_convert_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.6897233201581028, "step": 349, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.691699604743083, "grad_norm": 3.6558475010700313, "learning_rate": 1.1496240095852002e-06, "loss": 0.197, "step": 350 }, { "epoch": 0.691699604743083, "step": 350, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.8571428571428571, "train_accuracy_recipient_token_find_total": 7, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.691699604743083, "step": 350, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 0.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_enh": 1.0, "train_accuracy_recipient_token_enh_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_opt": 1.0, "train_accuracy_recipient_token_opt_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.6936758893280632, "grad_norm": 1.4008973035356875, "learning_rate": 1.1361627469736286e-06, "loss": 0.1253, "step": 351 }, { "epoch": 0.6936758893280632, "step": 351, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.7857142857142857, "train_accuracy_recipient_token_all_total": 14, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.6936758893280632, "step": 351, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.8333333333333334, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { "epoch": 0.6956521739130435, "grad_norm": 1.8030546844266104, "learning_rate": 1.122757546369744e-06, "loss": 0.2569, "step": 352 }, { "epoch": 0.6956521739130435, "step": 352, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_detect": 1.0, "train_accuracy_recipient_token_detect_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_review": 1.0, "train_accuracy_recipient_token_review_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.6956521739130435, "step": 352, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_join": 1.0, "train_accuracy_recipient_token_join_total": 1, "train_accuracy_recipient_token_vol": 1.0, "train_accuracy_recipient_token_vol_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.6976284584980237, "grad_norm": 1.1757496987091924, "learning_rate": 1.1094089588088383e-06, "loss": 0.144, "step": 353 }, { "epoch": 0.6976284584980237, "step": 353, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 8, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.6976284584980237, "step": 353, "train_accuracy_recipient_token": 0.85, "train_accuracy_recipient_token_all": 0.7692307692307693, "train_accuracy_recipient_token_all_total": 13, "train_accuracy_recipient_token_detect": 1.0, "train_accuracy_recipient_token_detect_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.6996047430830039, "grad_norm": 1.7957489707100158, "learning_rate": 1.0961175329990631e-06, "loss": 0.2026, "step": 354 }, { "epoch": 0.6996047430830039, "step": 354, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_opt": 1.0, "train_accuracy_recipient_token_opt_total": 1, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.6996047430830039, "step": 354, "train_accuracy_recipient_token": 0.8571428571428571, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_design": 1.0, "train_accuracy_recipient_token_design_total": 1, "train_accuracy_recipient_token_im": 1.0, "train_accuracy_recipient_token_im_total": 1, "train_accuracy_recipient_token_install": 0.0, "train_accuracy_recipient_token_install_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 14 }, { "epoch": 0.7015810276679841, "grad_norm": 1.7932806243824873, "learning_rate": 1.082883815298876e-06, "loss": 0.2821, "step": 355 }, { "epoch": 0.7015810276679841, "step": 355, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.7015810276679841, "step": 355, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.7142857142857143, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.7035573122529645, "grad_norm": 2.0800060452488855, "learning_rate": 1.0697083496945766e-06, "loss": 0.2197, "step": 356 }, { "epoch": 0.7035573122529645, "step": 356, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 20 }, { "epoch": 0.7035573122529645, "step": 356, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.7055335968379447, "grad_norm": 1.460137074708585, "learning_rate": 1.0565916777779519e-06, "loss": 0.1778, "step": 357 }, { "epoch": 0.7055335968379447, "step": 357, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.7055335968379447, "step": 357, "train_accuracy_recipient_token": 0.9333333333333333, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_track": 1.0, "train_accuracy_recipient_token_track_total": 1, "train_total_number_recipient_token": 15 }, { "epoch": 0.7075098814229249, "grad_norm": 2.634085943263362, "learning_rate": 1.04353433872401e-06, "loss": 0.487, "step": 358 }, { "epoch": 0.7075098814229249, "step": 358, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 16 }, { "epoch": 0.7075098814229249, "step": 358, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { "epoch": 0.7094861660079052, "grad_norm": 1.699389503506117, "learning_rate": 1.0305368692688175e-06, "loss": 0.2116, "step": 359 }, { "epoch": 0.7094861660079052, "step": 359, "train_accuracy_recipient_token": 0.9333333333333333, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_merge": 1.0, "train_accuracy_recipient_token_merge_total": 1, "train_total_number_recipient_token": 15 }, { "epoch": 0.7094861660079052, "step": 359, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.7114624505928854, "grad_norm": 2.1791368035785204, "learning_rate": 1.0175998036874357e-06, "loss": 0.3231, "step": 360 }, { "epoch": 0.7114624505928854, "step": 360, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.7114624505928854, "step": 360, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_ins": 1.0, "train_accuracy_recipient_token_ins_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.7134387351778656, "grad_norm": 1.5860150416516434, "learning_rate": 1.00472367377196e-06, "loss": 0.1088, "step": 361 }, { "epoch": 0.7134387351778656, "step": 361, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_d": 1.0, "train_accuracy_recipient_token_d_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_vehicle": 0.6666666666666666, "train_accuracy_recipient_token_vehicle_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.7134387351778656, "step": 361, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.7154150197628458, "grad_norm": 2.178306730601671, "learning_rate": 9.91909008809659e-07, "loss": 0.2397, "step": 362 }, { "epoch": 0.7154150197628458, "step": 362, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.7154150197628458, "step": 362, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 2, "train_accuracy_recipient_token_building": 1.0, "train_accuracy_recipient_token_building_total": 3, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_object": 0.6666666666666666, "train_accuracy_recipient_token_object_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.717391304347826, "grad_norm": 2.3506607158415993, "learning_rate": 9.791563355612171e-07, "loss": 0.3363, "step": 363 }, { "epoch": 0.717391304347826, "step": 363, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8888888888888888, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.717391304347826, "step": 363, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_join": 0.0, "train_accuracy_recipient_token_join_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.7193675889328063, "grad_norm": 2.4377826724234533, "learning_rate": 9.66466178239084e-07, "loss": 0.3347, "step": 364 }, { "epoch": 0.7193675889328063, "step": 364, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compile": 1.0, "train_accuracy_recipient_token_compile_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.6, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 17 }, { "epoch": 0.7193675889328063, "step": 364, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.7142857142857143, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.7213438735177866, "grad_norm": 2.1195801120949644, "learning_rate": 9.538390584859213e-07, "loss": 0.2291, "step": 365 }, { "epoch": 0.7213438735177866, "step": 365, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.7213438735177866, "step": 365, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_exp": 1.0, "train_accuracy_recipient_token_exp_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.7233201581027668, "grad_norm": 1.4023845894942437, "learning_rate": 9.412754953531664e-07, "loss": 0.1807, "step": 366 }, { "epoch": 0.7233201581027668, "step": 366, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 1, "train_accuracy_recipient_token_design": 1.0, "train_accuracy_recipient_token_design_total": 2, "train_accuracy_recipient_token_draw": 1.0, "train_accuracy_recipient_token_draw_total": 1, "train_accuracy_recipient_token_evaluate": 0.0, "train_accuracy_recipient_token_evaluate_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.7233201581027668, "step": 366, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 3, "train_accuracy_recipient_token_use": 1.0, "train_accuracy_recipient_token_use_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.7252964426877471, "grad_norm": 1.6925202838713784, "learning_rate": 9.28776005279691e-07, "loss": 0.2561, "step": 367 }, { "epoch": 0.7252964426877471, "step": 367, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_design": 1.0, "train_accuracy_recipient_token_design_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.7252964426877471, "step": 367, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_create": 0.5, "train_accuracy_recipient_token_create_total": 2, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.7272727272727273, "grad_norm": 1.4692209621399164, "learning_rate": 9.163411020705762e-07, "loss": 0.2018, "step": 368 }, { "epoch": 0.7272727272727273, "step": 368, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { "epoch": 0.7272727272727273, "step": 368, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9166666666666666, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.7292490118577075, "grad_norm": 1.8701554324040663, "learning_rate": 9.039712968759864e-07, "loss": 0.2345, "step": 369 }, { "epoch": 0.7292490118577075, "step": 369, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 17 }, { "epoch": 0.7292490118577075, "step": 369, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 2, "train_accuracy_recipient_token_inventory": 0.0, "train_accuracy_recipient_token_inventory_total": 1, "train_accuracy_recipient_token_sales": 1.0, "train_accuracy_recipient_token_sales_total": 1, "train_accuracy_recipient_token_shelf": 1.0, "train_accuracy_recipient_token_shelf_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.7312252964426877, "grad_norm": 1.9983260849142814, "learning_rate": 8.916670981701656e-07, "loss": 0.2475, "step": 370 }, { "epoch": 0.7312252964426877, "step": 370, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_buy": 0.0, "train_accuracy_recipient_token_buy_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 7, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 20 }, { "epoch": 0.7312252964426877, "step": 370, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_locate": 1.0, "train_accuracy_recipient_token_locate_total": 1, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.733201581027668, "grad_norm": 1.6154324345581008, "learning_rate": 8.794290117305296e-07, "loss": 0.1644, "step": 371 }, { "epoch": 0.733201581027668, "step": 371, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_m": 1.0, "train_accuracy_recipient_token_m_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.733201581027668, "step": 371, "train_accuracy_recipient_token": 0.7777777777777778, "train_accuracy_recipient_token_all": 0.7857142857142857, "train_accuracy_recipient_token_all_total": 14, "train_accuracy_recipient_token_calculate": 0.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.7351778656126482, "grad_norm": 1.5991587060424923, "learning_rate": 8.672575406168782e-07, "loss": 0.1712, "step": 372 }, { "epoch": 0.7351778656126482, "step": 372, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.7351778656126482, "step": 372, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.6, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.7371541501976284, "grad_norm": 1.660305490888912, "learning_rate": 8.551531851507186e-07, "loss": 0.2554, "step": 373 }, { "epoch": 0.7371541501976284, "step": 373, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 14, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_translate": 0.0, "train_accuracy_recipient_token_translate_total": 1, "train_accuracy_recipient_token_vehicle": 1.0, "train_accuracy_recipient_token_vehicle_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.7371541501976284, "step": 373, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_construction": 1.0, "train_accuracy_recipient_token_construction_total": 2, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 16 }, { "epoch": 0.7391304347826086, "grad_norm": 2.4165564387939567, "learning_rate": 8.431164428946928e-07, "loss": 0.3098, "step": 374 }, { "epoch": 0.7391304347826086, "step": 374, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.7142857142857143, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.7391304347826086, "step": 374, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.741106719367589, "grad_norm": 1.801628808429806, "learning_rate": 8.311478086321301e-07, "loss": 0.1757, "step": 375 }, { "epoch": 0.741106719367589, "step": 375, "train_accuracy_recipient_token": 0.8, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 5, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 10 }, { "epoch": 0.741106719367589, "step": 375, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_animal": 1.0, "train_accuracy_recipient_token_animal_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_locate": 1.0, "train_accuracy_recipient_token_locate_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.7430830039525692, "grad_norm": 12.540187334844513, "learning_rate": 8.192477743467078e-07, "loss": 0.7105, "step": 376 }, { "epoch": 0.7430830039525692, "step": 376, "train_accuracy_recipient_token": 0.8571428571428571, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 14 }, { "epoch": 0.7430830039525692, "step": 376, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_compose": 1.0, "train_accuracy_recipient_token_compose_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.7450592885375494, "grad_norm": 1.9972903002700673, "learning_rate": 8.074168292022269e-07, "loss": 0.25, "step": 377 }, { "epoch": 0.7450592885375494, "step": 377, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.7450592885375494, "step": 377, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.7470355731225297, "grad_norm": 2.3216872504414847, "learning_rate": 7.956554595225017e-07, "loss": 0.2433, "step": 378 }, { "epoch": 0.7470355731225297, "step": 378, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8333333333333334, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.7470355731225297, "step": 378, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_evaluate": 0.6666666666666666, "train_accuracy_recipient_token_evaluate_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 20 }, { "epoch": 0.7490118577075099, "grad_norm": 1.7733742585866188, "learning_rate": 7.839641487713745e-07, "loss": 0.2184, "step": 379 }, { "epoch": 0.7490118577075099, "step": 379, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_te": 1.0, "train_accuracy_recipient_token_te_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.7490118577075099, "step": 379, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_evaluate": 0.0, "train_accuracy_recipient_token_evaluate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_s": 1.0, "train_accuracy_recipient_token_s_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.7509881422924901, "grad_norm": 1.5040141373589382, "learning_rate": 7.723433775328385e-07, "loss": 0.1427, "step": 380 }, { "epoch": 0.7509881422924901, "step": 380, "train_accuracy_recipient_token": 0.7777777777777778, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 0.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_review": 1.0, "train_accuracy_recipient_token_review_total": 1, "train_accuracy_recipient_token_stamp": 1.0, "train_accuracy_recipient_token_stamp_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.7509881422924901, "step": 380, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_convert": 1.0, "train_accuracy_recipient_token_convert_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_sum": 1.0, "train_accuracy_recipient_token_sum_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.7529644268774703, "grad_norm": 2.2982731810128127, "learning_rate": 7.607936234912841e-07, "loss": 0.3524, "step": 381 }, { "epoch": 0.7529644268774703, "step": 381, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.7529644268774703, "step": 381, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_adjust": 1.0, "train_accuracy_recipient_token_adjust_total": 1, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_dis": 1.0, "train_accuracy_recipient_token_dis_total": 1, "train_accuracy_recipient_token_enh": 1.0, "train_accuracy_recipient_token_enh_total": 1, "train_accuracy_recipient_token_event": 1.0, "train_accuracy_recipient_token_event_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.7549407114624506, "grad_norm": 2.126256397171694, "learning_rate": 7.493153614118634e-07, "loss": 0.1726, "step": 382 }, { "epoch": 0.7549407114624506, "step": 382, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_compress": 1.0, "train_accuracy_recipient_token_compress_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.7549407114624506, "step": 382, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_custom": 1.0, "train_accuracy_recipient_token_custom_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_style": 1.0, "train_accuracy_recipient_token_style_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.7569169960474308, "grad_norm": 1.8984072674752601, "learning_rate": 7.379090631209712e-07, "loss": 0.1575, "step": 383 }, { "epoch": 0.7569169960474308, "step": 383, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 18 }, { "epoch": 0.7569169960474308, "step": 383, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_rec": 1.0, "train_accuracy_recipient_token_rec_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.758893280632411, "grad_norm": 1.8794511099778521, "learning_rate": 7.265751974868554e-07, "loss": 0.18, "step": 384 }, { "epoch": 0.758893280632411, "step": 384, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.758893280632411, "step": 384, "train_accuracy_recipient_token": 0.8125, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 0.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_clean": 1.0, "train_accuracy_recipient_token_clean_total": 1, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_rent": 1.0, "train_accuracy_recipient_token_rent_total": 2, "train_total_number_recipient_token": 16 }, { "epoch": 0.7608695652173914, "grad_norm": 2.08074863120985, "learning_rate": 7.153142304003419e-07, "loss": 0.1936, "step": 385 }, { "epoch": 0.7608695652173914, "step": 385, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 20 }, { "epoch": 0.7608695652173914, "step": 385, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8888888888888888, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.7628458498023716, "grad_norm": 1.8910489893410165, "learning_rate": 7.041266247556814e-07, "loss": 0.1853, "step": 386 }, { "epoch": 0.7628458498023716, "step": 386, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.7628458498023716, "step": 386, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_calculate": 0.5, "train_accuracy_recipient_token_calculate_total": 2, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.7648221343873518, "grad_norm": 2.1682658052582444, "learning_rate": 6.930128404315214e-07, "loss": 0.1819, "step": 387 }, { "epoch": 0.7648221343873518, "step": 387, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_ap": 1.0, "train_accuracy_recipient_token_ap_total": 1, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.7648221343873518, "step": 387, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 1, "train_accuracy_recipient_token_dis": 1.0, "train_accuracy_recipient_token_dis_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_interpret": 1.0, "train_accuracy_recipient_token_interpret_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.766798418972332, "grad_norm": 1.581806277615435, "learning_rate": 6.819733342720067e-07, "loss": 0.1822, "step": 388 }, { "epoch": 0.766798418972332, "step": 388, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_event": 0.6666666666666666, "train_accuracy_recipient_token_event_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_s": 1.0, "train_accuracy_recipient_token_s_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.766798418972332, "step": 388, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.7687747035573123, "grad_norm": 1.2710521758130064, "learning_rate": 6.710085600679967e-07, "loss": 0.0792, "step": 389 }, { "epoch": 0.7687747035573123, "step": 389, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.7687747035573123, "step": 389, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 20 }, { "epoch": 0.7707509881422925, "grad_norm": 1.618155327367614, "learning_rate": 6.601189685384127e-07, "loss": 0.1764, "step": 390 }, { "epoch": 0.7707509881422925, "step": 390, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8888888888888888, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_accuracy_recipient_token_restore": 0.0, "train_accuracy_recipient_token_restore_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.7707509881422925, "step": 390, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_evaluate": 1.0, "train_accuracy_recipient_token_evaluate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.7727272727272727, "grad_norm": 1.5648952371083706, "learning_rate": 6.493050073117115e-07, "loss": 0.2366, "step": 391 }, { "epoch": 0.7727272727272727, "step": 391, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.7727272727272727, "step": 391, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.8, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_locate": 1.0, "train_accuracy_recipient_token_locate_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.7747035573122529, "grad_norm": 1.8885528316331135, "learning_rate": 6.385671209074829e-07, "loss": 0.2286, "step": 392 }, { "epoch": 0.7747035573122529, "step": 392, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_ass": 0.0, "train_accuracy_recipient_token_ass_total": 1, "train_accuracy_recipient_token_environment": 1.0, "train_accuracy_recipient_token_environment_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_report": 1.0, "train_accuracy_recipient_token_report_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.7747035573122529, "step": 392, "train_accuracy_recipient_token": 0.9523809523809523, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_f": 0.75, "train_accuracy_recipient_token_f_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_locate": 1.0, "train_accuracy_recipient_token_locate_total": 1, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 2, "train_total_number_recipient_token": 21 }, { "epoch": 0.7766798418972332, "grad_norm": 1.841125886164984, "learning_rate": 6.279057507181796e-07, "loss": 0.2481, "step": 393 }, { "epoch": 0.7766798418972332, "step": 393, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 17 }, { "epoch": 0.7766798418972332, "step": 393, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 16 }, { "epoch": 0.7786561264822134, "grad_norm": 1.3411412038428177, "learning_rate": 6.17321334990973e-07, "loss": 0.0975, "step": 394 }, { "epoch": 0.7786561264822134, "step": 394, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 2, "train_accuracy_recipient_token_enh": 1.0, "train_accuracy_recipient_token_enh_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.7786561264822134, "step": 394, "train_accuracy_recipient_token": 0.85, "train_accuracy_recipient_token_all": 0.7692307692307693, "train_accuracy_recipient_token_all_total": 13, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.7806324110671937, "grad_norm": 2.0638574101441773, "learning_rate": 6.068143088097372e-07, "loss": 0.2713, "step": 395 }, { "epoch": 0.7806324110671937, "step": 395, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_locate": 1.0, "train_accuracy_recipient_token_locate_total": 2, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 2, "train_total_number_recipient_token": 20 }, { "epoch": 0.7806324110671937, "step": 395, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 20 }, { "epoch": 0.782608695652174, "grad_norm": 1.5396226003765137, "learning_rate": 5.963851040771639e-07, "loss": 0.2094, "step": 396 }, { "epoch": 0.782608695652174, "step": 396, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_stamp": 1.0, "train_accuracy_recipient_token_stamp_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.782608695652174, "step": 396, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.7845849802371542, "grad_norm": 2.0476241421143033, "learning_rate": 5.860341494970132e-07, "loss": 0.3632, "step": 397 }, { "epoch": 0.7845849802371542, "step": 397, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_city": 1.0, "train_accuracy_recipient_token_city_total": 1, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 1, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_urban": 1.0, "train_accuracy_recipient_token_urban_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.7845849802371542, "step": 397, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_gift": 1.0, "train_accuracy_recipient_token_gift_total": 1, "train_accuracy_recipient_token_personal": 1.0, "train_accuracy_recipient_token_personal_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.7865612648221344, "grad_norm": 1.9696813260070691, "learning_rate": 5.757618705564849e-07, "loss": 0.2611, "step": 398 }, { "epoch": 0.7865612648221344, "step": 398, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_accuracy_recipient_token_initialize": 0.0, "train_accuracy_recipient_token_initialize_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.7865612648221344, "step": 398, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 0.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_verify": 1.0, "train_accuracy_recipient_token_verify_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.7885375494071146, "grad_norm": 2.175586246163446, "learning_rate": 5.655686895087328e-07, "loss": 0.1876, "step": 399 }, { "epoch": 0.7885375494071146, "step": 399, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_building": 1.0, "train_accuracy_recipient_token_building_total": 1, "train_accuracy_recipient_token_flag": 1.0, "train_accuracy_recipient_token_flag_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.7885375494071146, "step": 399, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_deb": 0.0, "train_accuracy_recipient_token_deb_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.7905138339920948, "grad_norm": 2.494414189208174, "learning_rate": 5.554550253555066e-07, "loss": 0.2566, "step": 400 }, { "epoch": 0.7905138339920948, "step": 400, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_building": 1.0, "train_accuracy_recipient_token_building_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.7905138339920948, "step": 400, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_configure": 1.0, "train_accuracy_recipient_token_configure_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.7924901185770751, "grad_norm": 1.953703115430371, "learning_rate": 5.454212938299256e-07, "loss": 0.1746, "step": 401 }, { "epoch": 0.7924901185770751, "step": 401, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.7924901185770751, "step": 401, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_accuracy_recipient_token_rel": 1.0, "train_accuracy_recipient_token_rel_total": 2, "train_accuracy_recipient_token_y": 1.0, "train_accuracy_recipient_token_y_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.7944664031620553, "grad_norm": 1.8516896605082545, "learning_rate": 5.354679073793942e-07, "loss": 0.1904, "step": 402 }, { "epoch": 0.7944664031620553, "step": 402, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 20 }, { "epoch": 0.7944664031620553, "step": 402, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 0.875, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.7964426877470355, "grad_norm": 1.8578121936874719, "learning_rate": 5.255952751486442e-07, "loss": 0.2239, "step": 403 }, { "epoch": 0.7964426877470355, "step": 403, "train_accuracy_recipient_token": 0.9523809523809523, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_locate": 1.0, "train_accuracy_recipient_token_locate_total": 1, "train_total_number_recipient_token": 21 }, { "epoch": 0.7964426877470355, "step": 403, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_short": 1.0, "train_accuracy_recipient_token_short_total": 1, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.7984189723320159, "grad_norm": 2.1579654219661144, "learning_rate": 5.158038029629195e-07, "loss": 0.2192, "step": 404 }, { "epoch": 0.7984189723320159, "step": 404, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_recipe": 1.0, "train_accuracy_recipient_token_recipe_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.7984189723320159, "step": 404, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.8003952569169961, "grad_norm": 1.5522372172075602, "learning_rate": 5.060938933112891e-07, "loss": 0.14, "step": 405 }, { "epoch": 0.8003952569169961, "step": 405, "train_accuracy_recipient_token": 0.75, "train_accuracy_recipient_token_all": 0.75, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_crop": 1.0, "train_accuracy_recipient_token_crop_total": 1, "train_accuracy_recipient_token_f": 1.0, "train_accuracy_recipient_token_f_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_har": 0.0, "train_accuracy_recipient_token_har_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.8003952569169961, "step": 405, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.8888888888888888, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_land": 0.5, "train_accuracy_recipient_token_land_total": 4, "train_accuracy_recipient_token_mount": 1.0, "train_accuracy_recipient_token_mount_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.8023715415019763, "grad_norm": 2.1036940569062943, "learning_rate": 4.964659453301088e-07, "loss": 0.2345, "step": 406 }, { "epoch": 0.8023715415019763, "step": 406, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 0.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_go": 1.0, "train_accuracy_recipient_token_go_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.8023715415019763, "step": 406, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { "epoch": 0.8043478260869565, "grad_norm": 1.6667772877286244, "learning_rate": 4.869203547866097e-07, "loss": 0.1708, "step": 407 }, { "epoch": 0.8043478260869565, "step": 407, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.8043478260869565, "step": 407, "train_accuracy_recipient_token": 0.7894736842105263, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_c": 0.5, "train_accuracy_recipient_token_c_total": 2, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_ident": 0.5, "train_accuracy_recipient_token_ident_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.8063241106719368, "grad_norm": 2.2458743380912662, "learning_rate": 4.774575140626317e-07, "loss": 0.2932, "step": 408 }, { "epoch": 0.8063241106719368, "step": 408, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 17 }, { "epoch": 0.8063241106719368, "step": 408, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.808300395256917, "grad_norm": 2.48794385983657, "learning_rate": 4.6807781213849354e-07, "loss": 0.2428, "step": 409 }, { "epoch": 0.808300395256917, "step": 409, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.808300395256917, "step": 409, "train_accuracy_recipient_token": 0.7647058823529411, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_convert": 1.0, "train_accuracy_recipient_token_convert_total": 3, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 17 }, { "epoch": 0.8102766798418972, "grad_norm": 2.7307307205035567, "learning_rate": 4.5878163457700327e-07, "loss": 0.27, "step": 410 }, { "epoch": 0.8102766798418972, "step": 410, "train_accuracy_recipient_token": 0.9285714285714286, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_data": 1.0, "train_accuracy_recipient_token_data_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 14 }, { "epoch": 0.8102766798418972, "step": 410, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 19 }, { "epoch": 0.8122529644268774, "grad_norm": 2.7985267824858187, "learning_rate": 4.495693635076101e-07, "loss": 0.2315, "step": 411 }, { "epoch": 0.8122529644268774, "step": 411, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_h": 1.0, "train_accuracy_recipient_token_h_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.8122529644268774, "step": 411, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 0.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { "epoch": 0.8142292490118577, "grad_norm": 1.2979415141987987, "learning_rate": 4.4044137761069584e-07, "loss": 0.0836, "step": 412 }, { "epoch": 0.8142292490118577, "step": 412, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_crop": 1.0, "train_accuracy_recipient_token_crop_total": 1, "train_accuracy_recipient_token_f": 1.0, "train_accuracy_recipient_token_f_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_har": 0.0, "train_accuracy_recipient_token_har_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.8142292490118577, "step": 412, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 20 }, { "epoch": 0.8162055335968379, "grad_norm": 2.3682354960537606, "learning_rate": 4.3139805210200917e-07, "loss": 0.1285, "step": 413 }, { "epoch": 0.8162055335968379, "step": 413, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 20 }, { "epoch": 0.8162055335968379, "step": 413, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_sum": 1.0, "train_accuracy_recipient_token_sum_total": 1, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.8181818181818182, "grad_norm": 2.1831995030818256, "learning_rate": 4.2243975871724017e-07, "loss": 0.2866, "step": 414 }, { "epoch": 0.8181818181818182, "step": 414, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 17 }, { "epoch": 0.8181818181818182, "step": 414, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_r": 1.0, "train_accuracy_recipient_token_r_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.8201581027667985, "grad_norm": 2.3802602702771556, "learning_rate": 4.1356686569674344e-07, "loss": 0.2559, "step": 415 }, { "epoch": 0.8201581027667985, "step": 415, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { "epoch": 0.8201581027667985, "step": 415, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 0.75, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.8221343873517787, "grad_norm": 2.0956524139072057, "learning_rate": 4.0477973777039854e-07, "loss": 0.177, "step": 416 }, { "epoch": 0.8221343873517787, "step": 416, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_motor": 1.0, "train_accuracy_recipient_token_motor_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.8221343873517787, "step": 416, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.8241106719367589, "grad_norm": 1.505715219382788, "learning_rate": 3.9607873614261717e-07, "loss": 0.2022, "step": 417 }, { "epoch": 0.8241106719367589, "step": 417, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 0.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 0.875, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.8241106719367589, "step": 417, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 17 }, { "epoch": 0.8260869565217391, "grad_norm": 1.7465367169322799, "learning_rate": 3.874642184774977e-07, "loss": 0.0952, "step": 418 }, { "epoch": 0.8260869565217391, "step": 418, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_be": 0.75, "train_accuracy_recipient_token_be_total": 4, "train_accuracy_recipient_token_find": 0.8, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_s": 1.0, "train_accuracy_recipient_token_s_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.8260869565217391, "step": 418, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_total_number_recipient_token": 18 }, { "epoch": 0.8280632411067194, "grad_norm": 1.4978654089504846, "learning_rate": 3.789365388841193e-07, "loss": 0.1888, "step": 419 }, { "epoch": 0.8280632411067194, "step": 419, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_gr": 1.0, "train_accuracy_recipient_token_gr_total": 2, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_interpret": 1.0, "train_accuracy_recipient_token_interpret_total": 1, "train_accuracy_recipient_token_remove": 1.0, "train_accuracy_recipient_token_remove_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.8280632411067194, "step": 419, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 0.5, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.8300395256916996, "grad_norm": 1.5323629508097336, "learning_rate": 3.7049604790198976e-07, "loss": 0.1212, "step": 420 }, { "epoch": 0.8300395256916996, "step": 420, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_f": 1.0, "train_accuracy_recipient_token_f_total": 2, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_port": 1.0, "train_accuracy_recipient_token_port_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.8300395256916996, "step": 420, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 2, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.8320158102766798, "grad_norm": 1.9086165088662692, "learning_rate": 3.621430924866348e-07, "loss": 0.2971, "step": 421 }, { "epoch": 0.8320158102766798, "step": 421, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_calculate": 0.6666666666666666, "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_convert": 1.0, "train_accuracy_recipient_token_convert_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_set": 1.0, "train_accuracy_recipient_token_set_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.8320158102766798, "step": 421, "train_accuracy_recipient_token": 0.8125, "train_accuracy_recipient_token_a": 1.0, "train_accuracy_recipient_token_a_total": 1, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_fish": 0.6666666666666666, "train_accuracy_recipient_token_fish_total": 3, "train_total_number_recipient_token": 16 }, { "epoch": 0.83399209486166, "grad_norm": 2.9759596008058633, "learning_rate": 3.538780159953348e-07, "loss": 0.212, "step": 422 }, { "epoch": 0.83399209486166, "step": 422, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9166666666666666, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_locate": 1.0, "train_accuracy_recipient_token_locate_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.83399209486166, "step": 422, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.9, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 20 }, { "epoch": 0.8359683794466403, "grad_norm": 1.8348800188829963, "learning_rate": 3.457011581730124e-07, "loss": 0.2975, "step": 423 }, { "epoch": 0.8359683794466403, "step": 423, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 18 }, { "epoch": 0.8359683794466403, "step": 423, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_pill": 1.0, "train_accuracy_recipient_token_pill_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.8379446640316206, "grad_norm": 1.7013901333390404, "learning_rate": 3.3761285513826627e-07, "loss": 0.2019, "step": 424 }, { "epoch": 0.8379446640316206, "step": 424, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_evaluate": 1.0, "train_accuracy_recipient_token_evaluate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_plan": 1.0, "train_accuracy_recipient_token_plan_total": 1, "train_accuracy_recipient_token_s": 1.0, "train_accuracy_recipient_token_s_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.8379446640316206, "step": 424, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 0.5, "train_accuracy_recipient_token_calculate_total": 2, "train_accuracy_recipient_token_estimate": 0.5, "train_accuracy_recipient_token_estimate_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_opt": 1.0, "train_accuracy_recipient_token_opt_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.8399209486166008, "grad_norm": 2.3984171711273907, "learning_rate": 3.2961343936955384e-07, "loss": 0.229, "step": 425 }, { "epoch": 0.8399209486166008, "step": 425, "train_accuracy_recipient_token": 0.8125, "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_con": 1.0, "train_accuracy_recipient_token_con_total": 1, "train_accuracy_recipient_token_d": 1.0, "train_accuracy_recipient_token_d_total": 1, "train_accuracy_recipient_token_lands": 1.0, "train_accuracy_recipient_token_lands_total": 1, "train_accuracy_recipient_token_organ": 1.0, "train_accuracy_recipient_token_organ_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.8399209486166008, "step": 425, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8888888888888888, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.841897233201581, "grad_norm": 2.1793174841282528, "learning_rate": 3.217032396915265e-07, "loss": 0.2292, "step": 426 }, { "epoch": 0.841897233201581, "step": 426, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_dec": 1.0, "train_accuracy_recipient_token_dec_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.841897233201581, "step": 426, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.8438735177865613, "grad_norm": 2.181267379011323, "learning_rate": 3.1388258126151093e-07, "loss": 0.177, "step": 427 }, { "epoch": 0.8438735177865613, "step": 427, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.8438735177865613, "step": 427, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_l": 1.0, "train_accuracy_recipient_token_l_total": 1, "train_accuracy_recipient_token_remove": 1.0, "train_accuracy_recipient_token_remove_total": 1, "train_accuracy_recipient_token_write": 1.0, "train_accuracy_recipient_token_write_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.8458498023715415, "grad_norm": 1.5941096991791754, "learning_rate": 3.06151785556143e-07, "loss": 0.0573, "step": 428 }, { "epoch": 0.8458498023715415, "step": 428, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_detect": 1.0, "train_accuracy_recipient_token_detect_total": 1, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 19 }, { "epoch": 0.8458498023715415, "step": 428, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_accuracy_recipient_token_write": 1.0, "train_accuracy_recipient_token_write_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.8478260869565217, "grad_norm": 1.5894279466737562, "learning_rate": 2.98511170358155e-07, "loss": 0.1393, "step": 429 }, { "epoch": 0.8478260869565217, "step": 429, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9166666666666666, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_paint": 1.0, "train_accuracy_recipient_token_paint_total": 1, "train_accuracy_recipient_token_rep": 1.0, "train_accuracy_recipient_token_rep_total": 1, "train_accuracy_recipient_token_surface": 1.0, "train_accuracy_recipient_token_surface_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.8478260869565217, "step": 429, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.8333333333333334, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.849802371541502, "grad_norm": 1.872429124870318, "learning_rate": 2.9096104974331186e-07, "loss": 0.2275, "step": 430 }, { "epoch": 0.849802371541502, "step": 430, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 20 }, { "epoch": 0.849802371541502, "step": 430, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.8517786561264822, "grad_norm": 1.9625847182922385, "learning_rate": 2.8350173406749975e-07, "loss": 0.1982, "step": 431 }, { "epoch": 0.8517786561264822, "step": 431, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.8517786561264822, "step": 431, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 16 }, { "epoch": 0.8537549407114624, "grad_norm": 2.051731787985664, "learning_rate": 2.761335299539708e-07, "loss": 0.249, "step": 432 }, { "epoch": 0.8537549407114624, "step": 432, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_f": 1.0, "train_accuracy_recipient_token_f_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_mul": 1.0, "train_accuracy_recipient_token_mul_total": 1, "train_accuracy_recipient_token_we": 1.0, "train_accuracy_recipient_token_we_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.8537549407114624, "step": 432, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_histor": 1.0, "train_accuracy_recipient_token_histor_total": 1, "train_accuracy_recipient_token_land": 1.0, "train_accuracy_recipient_token_land_total": 3, "train_total_number_recipient_token": 20 }, { "epoch": 0.8557312252964426, "grad_norm": 1.5632631855936685, "learning_rate": 2.688567402807357e-07, "loss": 0.1849, "step": 433 }, { "epoch": 0.8557312252964426, "step": 433, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.8333333333333334, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.8557312252964426, "step": 433, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_rep": 1.0, "train_accuracy_recipient_token_rep_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.857707509881423, "grad_norm": 1.8054699895841433, "learning_rate": 2.6167166416811745e-07, "loss": 0.1723, "step": 434 }, { "epoch": 0.857707509881423, "step": 434, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_design": 1.0, "train_accuracy_recipient_token_design_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_play": 1.0, "train_accuracy_recipient_token_play_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.857707509881423, "step": 434, "train_accuracy_recipient_token": 0.7894736842105263, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_train": 0.25, "train_accuracy_recipient_token_train_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.8596837944664032, "grad_norm": 1.723792192639983, "learning_rate": 2.5457859696645245e-07, "loss": 0.1502, "step": 435 }, { "epoch": 0.8596837944664032, "step": 435, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_mach": 1.0, "train_accuracy_recipient_token_mach_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.8596837944664032, "step": 435, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 20 }, { "epoch": 0.8616600790513834, "grad_norm": 1.5466863752221973, "learning_rate": 2.4757783024395244e-07, "loss": 0.1792, "step": 436 }, { "epoch": 0.8616600790513834, "step": 436, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.8616600790513834, "step": 436, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_p": 1.0, "train_accuracy_recipient_token_p_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.8636363636363636, "grad_norm": 2.6743224940140915, "learning_rate": 2.4066965177471645e-07, "loss": 0.3654, "step": 437 }, { "epoch": 0.8636363636363636, "step": 437, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_accuracy_recipient_token_track": 1.0, "train_accuracy_recipient_token_track_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.8636363636363636, "step": 437, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.8656126482213439, "grad_norm": 2.1181583086625415, "learning_rate": 2.338543455269046e-07, "loss": 0.3387, "step": 438 }, { "epoch": 0.8656126482213439, "step": 438, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_ins": 1.0, "train_accuracy_recipient_token_ins_total": 1, "train_accuracy_recipient_token_remove": 1.0, "train_accuracy_recipient_token_remove_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.8656126482213439, "step": 438, "train_accuracy_recipient_token": 0.9130434782608695, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 0.6666666666666666, "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_m": 1.0, "train_accuracy_recipient_token_m_total": 4, "train_accuracy_recipient_token_s": 1.0, "train_accuracy_recipient_token_s_total": 2, "train_total_number_recipient_token": 23 }, { "epoch": 0.8675889328063241, "grad_norm": 2.0841614154281136, "learning_rate": 2.271321916510627e-07, "loss": 0.1456, "step": 439 }, { "epoch": 0.8675889328063241, "step": 439, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 8, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 2, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_join": 0.0, "train_accuracy_recipient_token_join_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.8675889328063241, "step": 439, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_locate": 1.0, "train_accuracy_recipient_token_locate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.8695652173913043, "grad_norm": 1.2831293107819657, "learning_rate": 2.205034664686076e-07, "loss": 0.1333, "step": 440 }, { "epoch": 0.8695652173913043, "step": 440, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.8695652173913043, "step": 440, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_text": 1.0, "train_accuracy_recipient_token_text_total": 6, "train_total_number_recipient_token": 21 }, { "epoch": 0.8715415019762845, "grad_norm": 2.007799864507481, "learning_rate": 2.1396844246046904e-07, "loss": 0.1843, "step": 441 }, { "epoch": 0.8715415019762845, "step": 441, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.8715415019762845, "step": 441, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_draw": 1.0, "train_accuracy_recipient_token_draw_total": 5, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 20 }, { "epoch": 0.8735177865612648, "grad_norm": 2.045124495201404, "learning_rate": 2.075273882558873e-07, "loss": 0.3035, "step": 442 }, { "epoch": 0.8735177865612648, "step": 442, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 20 }, { "epoch": 0.8735177865612648, "step": 442, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_locate": 1.0, "train_accuracy_recipient_token_locate_total": 1, "train_accuracy_recipient_token_s": 1.0, "train_accuracy_recipient_token_s_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.8754940711462451, "grad_norm": 2.0761359917977593, "learning_rate": 2.0118056862137358e-07, "loss": 0.1576, "step": 443 }, { "epoch": 0.8754940711462451, "step": 443, "train_accuracy_recipient_token": 0.8666666666666667, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_design": 0.0, "train_accuracy_recipient_token_design_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_extract": 0.5, "train_accuracy_recipient_token_extract_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 15 }, { "epoch": 0.8754940711462451, "step": 443, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_att": 0.0, "train_accuracy_recipient_token_att_total": 1, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_ex": 1.0, "train_accuracy_recipient_token_ex_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.8774703557312253, "grad_norm": 2.2388845065518983, "learning_rate": 1.9492824444982378e-07, "loss": 0.3765, "step": 444 }, { "epoch": 0.8774703557312253, "step": 444, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_learn": 1.0, "train_accuracy_recipient_token_learn_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.8774703557312253, "step": 444, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.8794466403162056, "grad_norm": 1.1600613922342395, "learning_rate": 1.887706727497965e-07, "loss": 0.0662, "step": 445 }, { "epoch": 0.8794466403162056, "step": 445, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.75, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 17 }, { "epoch": 0.8794466403162056, "step": 445, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.8814229249011858, "grad_norm": 2.141770501735159, "learning_rate": 1.8270810663494591e-07, "loss": 0.2166, "step": 446 }, { "epoch": 0.8814229249011858, "step": 446, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_accuracy_recipient_token_sort": 0.0, "train_accuracy_recipient_token_sort_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.8814229249011858, "step": 446, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9230769230769231, "train_accuracy_recipient_token_all_total": 13, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_restaurant": 1.0, "train_accuracy_recipient_token_restaurant_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.883399209486166, "grad_norm": 1.668725614891495, "learning_rate": 1.767407953136202e-07, "loss": 0.2025, "step": 447 }, { "epoch": 0.883399209486166, "step": 447, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.883399209486166, "step": 447, "train_accuracy_recipient_token": 0.7777777777777778, "train_accuracy_recipient_token_all": 0.75, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_building": 0.5, "train_accuracy_recipient_token_building_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_stat": 1.0, "train_accuracy_recipient_token_stat_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.8853754940711462, "grad_norm": 2.0839098091259984, "learning_rate": 1.7086898407861486e-07, "loss": 0.3808, "step": 448 }, { "epoch": 0.8853754940711462, "step": 448, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compile": 1.0, "train_accuracy_recipient_token_compile_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_accuracy_recipient_token_restaurant": 1.0, "train_accuracy_recipient_token_restaurant_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.8853754940711462, "step": 448, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_draw": 1.0, "train_accuracy_recipient_token_draw_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 20 }, { "epoch": 0.8873517786561265, "grad_norm": 2.6050856048557867, "learning_rate": 1.6509291429709224e-07, "loss": 0.2345, "step": 449 }, { "epoch": 0.8873517786561265, "step": 449, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_exp": 1.0, "train_accuracy_recipient_token_exp_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.8873517786561265, "step": 449, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_plant": 1.0, "train_accuracy_recipient_token_plant_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.8893280632411067, "grad_norm": 1.8361106243035428, "learning_rate": 1.59412823400657e-07, "loss": 0.2602, "step": 450 }, { "epoch": 0.8893280632411067, "step": 450, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.8893280632411067, "step": 450, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_bird": 1.0, "train_accuracy_recipient_token_bird_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_ident": 0.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.8913043478260869, "grad_norm": 2.2065101723440863, "learning_rate": 1.538289448755989e-07, "loss": 0.2448, "step": 451 }, { "epoch": 0.8913043478260869, "step": 451, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { "epoch": 0.8913043478260869, "step": 451, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { "epoch": 0.8932806324110671, "grad_norm": 2.529268779693517, "learning_rate": 1.483415082532938e-07, "loss": 0.3118, "step": 452 }, { "epoch": 0.8932806324110671, "step": 452, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8333333333333334, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 18 }, { "epoch": 0.8932806324110671, "step": 452, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 1, "train_accuracy_recipient_token_stat": 1.0, "train_accuracy_recipient_token_stat_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.8952569169960475, "grad_norm": 1.9479548952801957, "learning_rate": 1.4295073910076757e-07, "loss": 0.3757, "step": 453 }, { "epoch": 0.8952569169960475, "step": 453, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_object": 1.0, "train_accuracy_recipient_token_object_total": 1, "train_accuracy_recipient_token_people": 0.5, "train_accuracy_recipient_token_people_total": 2, "train_total_number_recipient_token": 20 }, { "epoch": 0.8952569169960475, "step": 453, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_verify": 1.0, "train_accuracy_recipient_token_verify_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.8972332015810277, "grad_norm": 3.204511061904964, "learning_rate": 1.3765685901142718e-07, "loss": 0.2914, "step": 454 }, { "epoch": 0.8972332015810277, "step": 454, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_building": 1.0, "train_accuracy_recipient_token_building_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.8972332015810277, "step": 454, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 0.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.8992094861660079, "grad_norm": 1.3138994217003859, "learning_rate": 1.3246008559594708e-07, "loss": 0.1385, "step": 455 }, { "epoch": 0.8992094861660079, "step": 455, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.8992094861660079, "step": 455, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.9011857707509882, "grad_norm": 2.8924372432420418, "learning_rate": 1.273606324733284e-07, "loss": 0.3419, "step": 456 }, { "epoch": 0.9011857707509882, "step": 456, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 0.6666666666666666, "train_accuracy_recipient_token_check_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.9011857707509882, "step": 456, "train_accuracy_recipient_token": 0.9333333333333333, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_recogn": 1.0, "train_accuracy_recipient_token_recogn_total": 1, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 15 }, { "epoch": 0.9031620553359684, "grad_norm": 2.118134582049429, "learning_rate": 1.223587092621162e-07, "loss": 0.3318, "step": 457 }, { "epoch": 0.9031620553359684, "step": 457, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 17 }, { "epoch": 0.9031620553359684, "step": 457, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 20 }, { "epoch": 0.9051383399209486, "grad_norm": 1.7617993832600736, "learning_rate": 1.1745452157178206e-07, "loss": 0.1621, "step": 458 }, { "epoch": 0.9051383399209486, "step": 458, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_event": 1.0, "train_accuracy_recipient_token_event_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 16 }, { "epoch": 0.9051383399209486, "step": 458, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_car": 1.0, "train_accuracy_recipient_token_car_total": 1, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.9071146245059288, "grad_norm": 1.5552007622687527, "learning_rate": 1.1264827099427417e-07, "loss": 0.1899, "step": 459 }, { "epoch": 0.9071146245059288, "step": 459, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_stat": 1.0, "train_accuracy_recipient_token_stat_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.9071146245059288, "step": 459, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.9090909090909091, "grad_norm": 2.050523711336241, "learning_rate": 1.0794015509572819e-07, "loss": 0.171, "step": 460 }, { "epoch": 0.9090909090909091, "step": 460, "train_accuracy_recipient_token": 0.9047619047619048, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 2, "train_accuracy_recipient_token_search": 0.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 21 }, { "epoch": 0.9090909090909091, "step": 460, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 2, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.9110671936758893, "grad_norm": 2.0269935707637354, "learning_rate": 1.0333036740834857e-07, "loss": 0.2898, "step": 461 }, { "epoch": 0.9110671936758893, "step": 461, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_gr": 1.0, "train_accuracy_recipient_token_gr_total": 3, "train_accuracy_recipient_token_ident": 0.5, "train_accuracy_recipient_token_ident_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.9110671936758893, "step": 461, "train_accuracy_recipient_token": 0.9047619047619048, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.9090909090909091, "train_accuracy_recipient_token_get_total": 11, "train_total_number_recipient_token": 21 }, { "epoch": 0.9130434782608695, "grad_norm": 1.7100530929084616, "learning_rate": 9.881909742245177e-08, "loss": 0.2431, "step": 462 }, { "epoch": 0.9130434782608695, "step": 462, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_verify": 1.0, "train_accuracy_recipient_token_verify_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.9130434782608695, "step": 462, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_arr": 0.0, "train_accuracy_recipient_token_arr_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_house": 1.0, "train_accuracy_recipient_token_house_total": 3, "train_accuracy_recipient_token_lands": 1.0, "train_accuracy_recipient_token_lands_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.9150197628458498, "grad_norm": 1.5464732166200588, "learning_rate": 9.440653057867816e-08, "loss": 0.1198, "step": 463 }, { "epoch": 0.9150197628458498, "step": 463, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_market": 1.0, "train_accuracy_recipient_token_market_total": 3, "train_accuracy_recipient_token_verify": 1.0, "train_accuracy_recipient_token_verify_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.9150197628458498, "step": 463, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 0.8888888888888888, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 20 }, { "epoch": 0.9169960474308301, "grad_norm": 1.6792924700521263, "learning_rate": 9.00928482603669e-08, "loss": 0.1242, "step": 464 }, { "epoch": 0.9169960474308301, "step": 464, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 20 }, { "epoch": 0.9169960474308301, "step": 464, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_research": 1.0, "train_accuracy_recipient_token_research_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.9189723320158103, "grad_norm": 2.1453318249444084, "learning_rate": 8.587822778610284e-08, "loss": 0.2816, "step": 465 }, { "epoch": 0.9189723320158103, "step": 465, "train_accuracy_recipient_token": 0.9285714285714286, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_extract": 0.5, "train_accuracy_recipient_token_extract_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 14 }, { "epoch": 0.9189723320158103, "step": 465, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_ex": 1.0, "train_accuracy_recipient_token_ex_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_s": 1.0, "train_accuracy_recipient_token_s_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.9209486166007905, "grad_norm": 2.6960675173414663, "learning_rate": 8.176284240242638e-08, "loss": 0.4503, "step": 466 }, { "epoch": 0.9209486166007905, "step": 466, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 0.875, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.9209486166007905, "step": 466, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 16 }, { "epoch": 0.9229249011857708, "grad_norm": 1.7798090264319586, "learning_rate": 7.774686127671183e-08, "loss": 0.1686, "step": 467 }, { "epoch": 0.9229249011857708, "step": 467, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_property": 1.0, "train_accuracy_recipient_token_property_total": 1, "train_accuracy_recipient_token_schedule": 0.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.9229249011857708, "step": 467, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.924901185770751, "grad_norm": 2.545370291066633, "learning_rate": 7.383044949021339e-08, "loss": 0.2601, "step": 468 }, { "epoch": 0.924901185770751, "step": 468, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_ap": 1.0, "train_accuracy_recipient_token_ap_total": 1, "train_accuracy_recipient_token_ger": 1.0, "train_accuracy_recipient_token_ger_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.924901185770751, "step": 468, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_s": 1.0, "train_accuracy_recipient_token_s_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.9268774703557312, "grad_norm": 1.4300480959159698, "learning_rate": 7.001376803128041e-08, "loss": 0.162, "step": 469 }, { "epoch": 0.9268774703557312, "step": 469, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8888888888888888, "train_accuracy_recipient_token_get_total": 9, "train_accuracy_recipient_token_ident": 0.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.9268774703557312, "step": 469, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 16 }, { "epoch": 0.9288537549407114, "grad_norm": 2.7747664296518257, "learning_rate": 6.629697378873839e-08, "loss": 0.2523, "step": 470 }, { "epoch": 0.9288537549407114, "step": 470, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_arr": 0.0, "train_accuracy_recipient_token_arr_total": 1, "train_accuracy_recipient_token_mar": 1.0, "train_accuracy_recipient_token_mar_total": 1, "train_accuracy_recipient_token_o": 1.0, "train_accuracy_recipient_token_o_total": 1, "train_accuracy_recipient_token_under": 1.0, "train_accuracy_recipient_token_under_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.9288537549407114, "step": 470, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 0.5, "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.9308300395256917, "grad_norm": 1.8580136680680779, "learning_rate": 6.268021954544095e-08, "loss": 0.2213, "step": 471 }, { "epoch": 0.9308300395256917, "step": 471, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_plan": 0.0, "train_accuracy_recipient_token_plan_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.9308300395256917, "step": 471, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_stat": 1.0, "train_accuracy_recipient_token_stat_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.932806324110672, "grad_norm": 2.204067976920768, "learning_rate": 5.9163653971989754e-08, "loss": 0.3436, "step": 472 }, { "epoch": 0.932806324110672, "step": 472, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_arch": 1.0, "train_accuracy_recipient_token_arch_total": 2, "train_accuracy_recipient_token_building": 1.0, "train_accuracy_recipient_token_building_total": 3, "train_accuracy_recipient_token_gr": 1.0, "train_accuracy_recipient_token_gr_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.932806324110672, "step": 472, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_inter": 1.0, "train_accuracy_recipient_token_inter_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.9347826086956522, "grad_norm": 1.3085005755431722, "learning_rate": 5.574742162062163e-08, "loss": 0.193, "step": 473 }, { "epoch": 0.9347826086956522, "step": 473, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.9347826086956522, "step": 473, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_f": 1.0, "train_accuracy_recipient_token_f_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_send": 1.0, "train_accuracy_recipient_token_send_total": 1, "train_accuracy_recipient_token_set": 1.0, "train_accuracy_recipient_token_set_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.9367588932806324, "grad_norm": 2.537954422029753, "learning_rate": 5.2431662919267825e-08, "loss": 0.2243, "step": 474 }, { "epoch": 0.9367588932806324, "step": 474, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 19 }, { "epoch": 0.9367588932806324, "step": 474, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_format": 1.0, "train_accuracy_recipient_token_format_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_sum": 1.0, "train_accuracy_recipient_token_sum_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.9387351778656127, "grad_norm": 1.805153682895635, "learning_rate": 4.9216514165781885e-08, "loss": 0.145, "step": 475 }, { "epoch": 0.9387351778656127, "step": 475, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_arr": 1.0, "train_accuracy_recipient_token_arr_total": 1, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.9387351778656127, "step": 475, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.8333333333333334, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_compare": 0.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.9407114624505929, "grad_norm": 2.045602997144322, "learning_rate": 4.61021075223364e-08, "loss": 0.2222, "step": 476 }, { "epoch": 0.9407114624505929, "step": 476, "train_accuracy_recipient_token": 0.7894736842105263, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_ex": 1.0, "train_accuracy_recipient_token_ex_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_plane": 1.0, "train_accuracy_recipient_token_plane_total": 3, "train_accuracy_recipient_token_zi": 0.0, "train_accuracy_recipient_token_zi_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.9407114624505929, "step": 476, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_arr": 1.0, "train_accuracy_recipient_token_arr_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.9426877470355731, "grad_norm": 1.3694314670334982, "learning_rate": 4.308857100999042e-08, "loss": 0.1326, "step": 477 }, { "epoch": 0.9426877470355731, "step": 477, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_ident": 0.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.9426877470355731, "step": 477, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 18 }, { "epoch": 0.9446640316205533, "grad_norm": 3.126805689186157, "learning_rate": 4.017602850342584e-08, "loss": 0.3008, "step": 478 }, { "epoch": 0.9446640316205533, "step": 478, "train_accuracy_recipient_token": 0.85, "train_accuracy_recipient_token_all": 0.75, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.9446640316205533, "step": 478, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.9466403162055336, "grad_norm": 2.434210637252682, "learning_rate": 3.736459972585815e-08, "loss": 0.2991, "step": 479 }, { "epoch": 0.9466403162055336, "step": 479, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.75, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 2, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.9466403162055336, "step": 479, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_plant": 1.0, "train_accuracy_recipient_token_plant_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.9486166007905138, "grad_norm": 1.5988207864411084, "learning_rate": 3.4654400244112654e-08, "loss": 0.1611, "step": 480 }, { "epoch": 0.9486166007905138, "step": 480, "train_accuracy_recipient_token": 0.85, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 15, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 20 }, { "epoch": 0.9486166007905138, "step": 480, "train_accuracy_recipient_token": 0.8, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_predict": 0.0, "train_accuracy_recipient_token_predict_total": 1, "train_accuracy_recipient_token_stamp": 1.0, "train_accuracy_recipient_token_stamp_total": 2, "train_total_number_recipient_token": 15 }, { "epoch": 0.950592885375494, "grad_norm": 1.9240578271638804, "learning_rate": 3.204554146387456e-08, "loss": 0.3299, "step": 481 }, { "epoch": 0.950592885375494, "step": 481, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.950592885375494, "step": 481, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_opt": 0.0, "train_accuracy_recipient_token_opt_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.9525691699604744, "grad_norm": 2.5790139743271983, "learning_rate": 2.9538130625110796e-08, "loss": 0.2576, "step": 482 }, { "epoch": 0.9525691699604744, "step": 482, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.9525691699604744, "step": 482, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 16 }, { "epoch": 0.9545454545454546, "grad_norm": 2.8948073860814616, "learning_rate": 2.7132270797659567e-08, "loss": 0.3431, "step": 483 }, { "epoch": 0.9545454545454546, "step": 483, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_ident": 0.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_ne": 1.0, "train_accuracy_recipient_token_ne_total": 3, "train_total_number_recipient_token": 20 }, { "epoch": 0.9545454545454546, "step": 483, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_convert": 1.0, "train_accuracy_recipient_token_convert_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 15 }, { "epoch": 0.9565217391304348, "grad_norm": 3.1119629331028023, "learning_rate": 2.4828060876995462e-08, "loss": 0.2377, "step": 484 }, { "epoch": 0.9565217391304348, "step": 484, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_custom": 1.0, "train_accuracy_recipient_token_custom_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.9565217391304348, "step": 484, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_pool": 1.0, "train_accuracy_recipient_token_pool_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.958498023715415, "grad_norm": 1.7117541555790323, "learning_rate": 2.262559558016325e-08, "loss": 0.2084, "step": 485 }, { "epoch": 0.958498023715415, "step": 485, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_robot": 1.0, "train_accuracy_recipient_token_robot_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.958498023715415, "step": 485, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 0.875, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.9604743083003953, "grad_norm": 1.620668314091392, "learning_rate": 2.052496544188487e-08, "loss": 0.1256, "step": 486 }, { "epoch": 0.9604743083003953, "step": 486, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_learn": 1.0, "train_accuracy_recipient_token_learn_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.9604743083003953, "step": 486, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 0.8888888888888888, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 20 }, { "epoch": 0.9624505928853755, "grad_norm": 1.5982169963407395, "learning_rate": 1.852625681083742e-08, "loss": 0.112, "step": 487 }, { "epoch": 0.9624505928853755, "step": 487, "train_accuracy_recipient_token": 0.9523809523809523, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_subtract": 1.0, "train_accuracy_recipient_token_subtract_total": 1, "train_accuracy_recipient_token_track": 1.0, "train_accuracy_recipient_token_track_total": 1, "train_total_number_recipient_token": 21 }, { "epoch": 0.9624505928853755, "step": 487, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 0.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_compile": 1.0, "train_accuracy_recipient_token_compile_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.9644268774703557, "grad_norm": 1.8281750558489605, "learning_rate": 1.6629551846104874e-08, "loss": 0.1926, "step": 488 }, { "epoch": 0.9644268774703557, "step": 488, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_extract": 1.0, "train_accuracy_recipient_token_extract_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_recogn": 1.0, "train_accuracy_recipient_token_recogn_total": 1, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.9644268774703557, "step": 488, "train_accuracy_recipient_token": 0.9333333333333333, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_verify": 1.0, "train_accuracy_recipient_token_verify_total": 1, "train_total_number_recipient_token": 15 }, { "epoch": 0.9664031620553359, "grad_norm": 2.6298463108507355, "learning_rate": 1.483492851379914e-08, "loss": 0.4112, "step": 489 }, { "epoch": 0.9664031620553359, "step": 489, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_check": 0.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.9664031620553359, "step": 489, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compile": 1.0, "train_accuracy_recipient_token_compile_total": 1, "train_accuracy_recipient_token_find": 0.75, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.9683794466403162, "grad_norm": 2.433575683997582, "learning_rate": 1.3142460583856487e-08, "loss": 0.2541, "step": 490 }, { "epoch": 0.9683794466403162, "step": 490, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_download": 0.0, "train_accuracy_recipient_token_download_total": 1, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_search": 0.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.9683794466403162, "step": 490, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.9703557312252964, "grad_norm": 1.7477968227475646, "learning_rate": 1.1552217627004426e-08, "loss": 0.1936, "step": 491 }, { "epoch": 0.9703557312252964, "step": 491, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_deb": 0.0, "train_accuracy_recipient_token_deb_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_plan": 1.0, "train_accuracy_recipient_token_plan_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.9703557312252964, "step": 491, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.9723320158102767, "grad_norm": 1.8404798729311513, "learning_rate": 1.006426501190233e-08, "loss": 0.2403, "step": 492 }, { "epoch": 0.9723320158102767, "step": 492, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.9723320158102767, "step": 492, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.974308300395257, "grad_norm": 2.0079365969770246, "learning_rate": 8.67866390245442e-09, "loss": 0.216, "step": 493 }, { "epoch": 0.974308300395257, "step": 493, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_park": 1.0, "train_accuracy_recipient_token_park_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.974308300395257, "step": 493, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_ex": 1.0, "train_accuracy_recipient_token_ex_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.9762845849802372, "grad_norm": 2.1557879594347273, "learning_rate": 7.3954712552953835e-09, "loss": 0.234, "step": 494 }, { "epoch": 0.9762845849802372, "step": 494, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.9762845849802372, "step": 494, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.9782608695652174, "grad_norm": 2.088360252773201, "learning_rate": 6.214739817448634e-09, "loss": 0.2639, "step": 495 }, { "epoch": 0.9782608695652174, "step": 495, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 0.8888888888888888, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.9782608695652174, "step": 495, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_filter": 1.0, "train_accuracy_recipient_token_filter_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.9802371541501976, "grad_norm": 1.6208106966240963, "learning_rate": 5.136518124159162e-09, "loss": 0.1421, "step": 496 }, { "epoch": 0.9802371541501976, "step": 496, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_motor": 1.0, "train_accuracy_recipient_token_motor_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.9802371541501976, "step": 496, "train_accuracy_recipient_token": 0.6842105263157895, "train_accuracy_recipient_token_all": 0.6666666666666666, "train_accuracy_recipient_token_all_total": 15, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.9822134387351779, "grad_norm": 2.040838086529817, "learning_rate": 4.160850496897906e-09, "loss": 0.164, "step": 497 }, { "epoch": 0.9822134387351779, "step": 497, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_building": 1.0, "train_accuracy_recipient_token_building_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.9822134387351779, "step": 497, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 2, "train_accuracy_recipient_token_table": 1.0, "train_accuracy_recipient_token_table_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.9841897233201581, "grad_norm": 1.8868735474227656, "learning_rate": 3.287777041539042e-09, "loss": 0.1909, "step": 498 }, { "epoch": 0.9841897233201581, "step": 498, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_generate": 0.0, "train_accuracy_recipient_token_generate_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_sum": 1.0, "train_accuracy_recipient_token_sum_total": 1, "train_accuracy_recipient_token_translate": 0.5, "train_accuracy_recipient_token_translate_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.9841897233201581, "step": 498, "train_accuracy_recipient_token": 0.9285714285714286, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_evaluate": 0.0, "train_accuracy_recipient_token_evaluate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_total_number_recipient_token": 14 }, { "epoch": 0.9861660079051383, "grad_norm": 1.9933889585193483, "learning_rate": 2.5173336467135266e-09, "loss": 0.2492, "step": 499 }, { "epoch": 0.9861660079051383, "step": 499, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_property": 1.0, "train_accuracy_recipient_token_property_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.9861660079051383, "step": 499, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_ex": 1.0, "train_accuracy_recipient_token_ex_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 18 }, { "epoch": 0.9881422924901185, "grad_norm": 1.780716845359166, "learning_rate": 1.8495519823308329e-09, "loss": 0.1323, "step": 500 }, { "epoch": 0.9881422924901185, "step": 500, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 2, "train_accuracy_recipient_token_calculate": 0.5, "train_accuracy_recipient_token_calculate_total": 2, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.9881422924901185, "step": 500, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.9901185770750988, "grad_norm": 2.296947318897557, "learning_rate": 1.284459498280266e-09, "loss": 0.134, "step": 501 }, { "epoch": 0.9901185770750988, "step": 501, "train_accuracy_recipient_token": 0.7777777777777778, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_schedule": 0.0, "train_accuracy_recipient_token_schedule_total": 1, "train_accuracy_recipient_token_track": 0.5, "train_accuracy_recipient_token_track_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.9901185770750988, "step": 501, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.9920948616600791, "grad_norm": 1.8997002986472211, "learning_rate": 8.220794233004814e-10, "loss": 0.2113, "step": 502 }, { "epoch": 0.9920948616600791, "step": 502, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.9920948616600791, "step": 502, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_write": 1.0, "train_accuracy_recipient_token_write_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.9940711462450593, "grad_norm": 1.812032510476914, "learning_rate": 4.624307640249681e-10, "loss": 0.2103, "step": 503 }, { "epoch": 0.9940711462450593, "step": 503, "train_accuracy_recipient_token": 0.6111111111111112, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_opt": 0.0, "train_accuracy_recipient_token_opt_total": 1, "train_accuracy_recipient_token_plan": 0.0, "train_accuracy_recipient_token_plan_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.9940711462450593, "step": 503, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 15, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.9960474308300395, "grad_norm": 1.9968965605808304, "learning_rate": 2.0552830420184077e-10, "loss": 0.2541, "step": 504 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 0.9285714285714286, "train_accuracy_recipient_token_all": 0.875, "train_accuracy_recipient_token_all_total": 8, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 14 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_ident": 0.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 2, "train_total_number_recipient_token": 16 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_histor": 1.0, "train_accuracy_recipient_token_histor_total": 3, "train_total_number_recipient_token": 18 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 0.7647058823529411, "train_accuracy_recipient_token_all": 0.7272727272727273, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 17 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_interpret": 1.0, "train_accuracy_recipient_token_interpret_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.8666666666666667, "train_accuracy_recipient_token_all_total": 15, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 0.95, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_tour": 0.75, "train_accuracy_recipient_token_tour_total": 4, "train_total_number_recipient_token": 20 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 0.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 19 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 17 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_hang": 1.0, "train_accuracy_recipient_token_hang_total": 1, "train_accuracy_recipient_token_re": 1.0, "train_accuracy_recipient_token_re_total": 1, "train_accuracy_recipient_token_remove": 1.0, "train_accuracy_recipient_token_remove_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_accuracy_recipient_token_invest": 1.0, "train_accuracy_recipient_token_invest_total": 1, "train_total_number_recipient_token": 19 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_add": 1.0, "train_accuracy_recipient_token_add_total": 1, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_construct": 1.0, "train_accuracy_recipient_token_construct_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_select": 1.0, "train_accuracy_recipient_token_select_total": 2, "train_total_number_recipient_token": 16 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 17 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_histor": 0.0, "train_accuracy_recipient_token_histor_total": 1, "train_accuracy_recipient_token_m": 1.0, "train_accuracy_recipient_token_m_total": 1, "train_total_number_recipient_token": 20 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_gr": 1.0, "train_accuracy_recipient_token_gr_total": 3, "train_accuracy_recipient_token_ident": 0.5, "train_accuracy_recipient_token_ident_total": 2, "train_total_number_recipient_token": 20 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 0.9285714285714286, "train_accuracy_recipient_token_all": 0.8888888888888888, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_draw": 1.0, "train_accuracy_recipient_token_draw_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 14 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.75, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_review": 0.5, "train_accuracy_recipient_token_review_total": 2, "train_total_number_recipient_token": 18 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_create": 0.5, "train_accuracy_recipient_token_create_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.8181818181818182, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_get": 0.875, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 19 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.8888888888888888, "train_accuracy_recipient_token_all_total": 9, "train_accuracy_recipient_token_custom": 1.0, "train_accuracy_recipient_token_custom_total": 4, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 19 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_eng": 1.0, "train_accuracy_recipient_token_eng_total": 2, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 1, "train_accuracy_recipient_token_lock": 1.0, "train_accuracy_recipient_token_lock_total": 1, "train_accuracy_recipient_token_tem": 1.0, "train_accuracy_recipient_token_tem_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 0.9375, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 11, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 16 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_car": 1.0, "train_accuracy_recipient_token_car_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 19 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_accuracy_recipient_token_invest": 1.0, "train_accuracy_recipient_token_invest_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 19 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_bird": 1.0, "train_accuracy_recipient_token_bird_total": 2, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_histor": 1.0, "train_accuracy_recipient_token_histor_total": 1, "train_accuracy_recipient_token_symbol": 1.0, "train_accuracy_recipient_token_symbol_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_organ": 1.0, "train_accuracy_recipient_token_organ_total": 1, "train_accuracy_recipient_token_rent": 1.0, "train_accuracy_recipient_token_rent_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_ident": 1.0, "train_accuracy_recipient_token_ident_total": 1, "train_accuracy_recipient_token_predict": 1.0, "train_accuracy_recipient_token_predict_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_evaluate": 0.0, "train_accuracy_recipient_token_evaluate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_total_number_recipient_token": 19 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_locate": 0.0, "train_accuracy_recipient_token_locate_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_park": 1.0, "train_accuracy_recipient_token_park_total": 2, "train_total_number_recipient_token": 17 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 3, "train_accuracy_recipient_token_artist": 0.0, "train_accuracy_recipient_token_artist_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_analy": 1.0, "train_accuracy_recipient_token_analy_total": 2, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_object": 1.0, "train_accuracy_recipient_token_object_total": 3, "train_accuracy_recipient_token_transfer": 1.0, "train_accuracy_recipient_token_transfer_total": 1, "train_total_number_recipient_token": 18 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_fruit": 1.0, "train_accuracy_recipient_token_fruit_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 17 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.8, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_accuracy_recipient_token_organ": 1.0, "train_accuracy_recipient_token_organ_total": 1, "train_total_number_recipient_token": 16 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 1, "train_accuracy_recipient_token_interpret": 1.0, "train_accuracy_recipient_token_interpret_total": 1, "train_accuracy_recipient_token_stage": 1.0, "train_accuracy_recipient_token_stage_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 0.8823529411764706, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.5, "train_accuracy_recipient_token_get_total": 2, "train_accuracy_recipient_token_rent": 1.0, "train_accuracy_recipient_token_rent_total": 1, "train_total_number_recipient_token": 17 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_total_number_recipient_token": 18 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 18 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 18 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 5, "train_accuracy_recipient_token_f": 0.0, "train_accuracy_recipient_token_f_total": 1, "train_accuracy_recipient_token_pest": 1.0, "train_accuracy_recipient_token_pest_total": 1, "train_accuracy_recipient_token_plant": 1.0, "train_accuracy_recipient_token_plant_total": 1, "train_accuracy_recipient_token_sun": 1.0, "train_accuracy_recipient_token_sun_total": 1, "train_total_number_recipient_token": 9 }, { "epoch": 0.9960474308300395, "eval_loss": 0.18871143460273743, "eval_runtime": 412.3849, "eval_samples_per_second": 0.245, "eval_steps_per_second": 0.124, "step": 504 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 0.8421052631578947, "train_accuracy_recipient_token_all": 0.8333333333333334, "train_accuracy_recipient_token_all_total": 12, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 2, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 3, "train_total_number_recipient_token": 19 }, { "epoch": 0.9960474308300395, "step": 504, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_ex": 1.0, "train_accuracy_recipient_token_ex_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_total_number_recipient_token": 19 }, { "epoch": 0.9980237154150198, "grad_norm": 1.6014381828385609, "learning_rate": 5.138260408488194e-11, "loss": 0.1502, "step": 505 }, { "epoch": 0.9980237154150198, "step": 505, "train_accuracy_recipient_token": 1.0, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 10, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 20 }, { "epoch": 0.9980237154150198, "step": 505, "train_accuracy_recipient_token": 0.7142857142857143, "train_accuracy_recipient_token_all": 0.6, "train_accuracy_recipient_token_all_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 2, "train_total_number_recipient_token": 7 }, { "epoch": 1.0, "grad_norm": 1.3267207098037004, "learning_rate": 0.0, "loss": 0.1335, "step": 506 }, { "epoch": 1.0, "step": 506, "total_flos": 0.0, "train_loss": 0.24046664490647937, "train_runtime": 13405.0682, "train_samples_per_second": 0.151, "train_steps_per_second": 0.038 } ], "logging_steps": 1.0, "max_steps": 506, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100.0, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": false, "should_training_stop": false }, "attributes": {} } }, "total_flos": 0.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }