{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 84, "global_step": 253, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0, "step": 0, "train_accuracy_recipient_token": 0.7272727272727273, "train_accuracy_recipient_token_all": 0.725, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_estimate": 0.75, "train_accuracy_recipient_token_estimate_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 8, "train_accuracy_recipient_token_generate": 0.5, "train_accuracy_recipient_token_generate_total": 2, "train_accuracy_recipient_token_get": 0.6363636363636364, "train_accuracy_recipient_token_get_total": 11, "train_total_number_recipient_token": 77 }, { "epoch": 0.003952569169960474, "grad_norm": 2.183184044468419, "learning_rate": 1.25e-07, "loss": 0.2945, "step": 1 }, { "epoch": 0.003952569169960474, "step": 1, "train_accuracy_recipient_token": 0.7236842105263158, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_find": 0.8571428571428571, "train_accuracy_recipient_token_find_total": 7, "train_accuracy_recipient_token_get": 0.7777777777777778, "train_accuracy_recipient_token_get_total": 18, "train_accuracy_recipient_token_interpret": 0.6666666666666666, "train_accuracy_recipient_token_interpret_total": 3, "train_accuracy_recipient_token_search": 0.3333333333333333, "train_accuracy_recipient_token_search_total": 3, "train_total_number_recipient_token": 76 }, { "epoch": 0.007905138339920948, "grad_norm": 2.4188473388642096, "learning_rate": 2.5e-07, "loss": 0.2003, "step": 2 }, { "epoch": 0.007905138339920948, "step": 2, "train_accuracy_recipient_token": 0.6578947368421053, "train_accuracy_recipient_token_all": 0.65, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 2, "train_accuracy_recipient_token_find": 0.8, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 0.7, "train_accuracy_recipient_token_get_total": 20, "train_accuracy_recipient_token_identify": 0.0, "train_accuracy_recipient_token_identify_total": 2, "train_total_number_recipient_token": 76 }, { "epoch": 0.011857707509881422, "grad_norm": 2.8778683824485056, "learning_rate": 3.75e-07, "loss": 0.2402, "step": 3 }, { "epoch": 0.011857707509881422, "step": 3, "train_accuracy_recipient_token": 0.5645161290322581, "train_accuracy_recipient_token_all": 0.5526315789473685, "train_accuracy_recipient_token_all_total": 38, "train_accuracy_recipient_token_analyze": 0.5, "train_accuracy_recipient_token_analyze_total": 4, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_fruit": 1.0, "train_accuracy_recipient_token_fruit_total": 3, "train_accuracy_recipient_token_get": 0.6, "train_accuracy_recipient_token_get_total": 5, "train_total_number_recipient_token": 62 }, { "epoch": 0.015810276679841896, "grad_norm": 2.4405568835024782, "learning_rate": 5e-07, "loss": 0.3288, "step": 4 }, { "epoch": 0.015810276679841896, "step": 4, "train_accuracy_recipient_token": 0.5972222222222222, "train_accuracy_recipient_token_all": 0.5952380952380952, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_g": 1.0, "train_accuracy_recipient_token_g_total": 4, "train_accuracy_recipient_token_get": 0.625, "train_accuracy_recipient_token_get_total": 16, "train_accuracy_recipient_token_identify": 0.0, "train_accuracy_recipient_token_identify_total": 2, "train_accuracy_recipient_token_validate": 0.0, "train_accuracy_recipient_token_validate_total": 2, "train_total_number_recipient_token": 72 }, { "epoch": 0.019762845849802372, "grad_norm": 2.4414211913605297, "learning_rate": 6.249999999999999e-07, "loss": 0.2889, "step": 5 }, { "epoch": 0.019762845849802372, "step": 5, "train_accuracy_recipient_token": 0.5797101449275363, "train_accuracy_recipient_token_all": 0.5952380952380952, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_evaluate": 0.0, "train_accuracy_recipient_token_evaluate_total": 1, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.5652173913043478, "train_accuracy_recipient_token_get_total": 23, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 69 }, { "epoch": 0.023715415019762844, "grad_norm": 2.2326884273282266, "learning_rate": 7.5e-07, "loss": 0.3532, "step": 6 }, { "epoch": 0.023715415019762844, "step": 6, "train_accuracy_recipient_token": 0.6891891891891891, "train_accuracy_recipient_token_all": 0.6585365853658537, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_count": 0.5, "train_accuracy_recipient_token_count_total": 2, "train_accuracy_recipient_token_find": 0.8, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 15, "train_accuracy_recipient_token_sum": 0.5, "train_accuracy_recipient_token_sum_total": 2, "train_total_number_recipient_token": 74 }, { "epoch": 0.02766798418972332, "grad_norm": 2.8493225497141546, "learning_rate": 8.75e-07, "loss": 0.2854, "step": 7 }, { "epoch": 0.02766798418972332, "step": 7, "train_accuracy_recipient_token": 0.6712328767123288, "train_accuracy_recipient_token_all": 0.65, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_compare": 0.75, "train_accuracy_recipient_token_compare_total": 4, "train_accuracy_recipient_token_find": 0.75, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 0.7692307692307693, "train_accuracy_recipient_token_get_total": 13, "train_accuracy_recipient_token_property": 1.0, "train_accuracy_recipient_token_property_total": 3, "train_total_number_recipient_token": 73 }, { "epoch": 0.03162055335968379, "grad_norm": 6.394934509276048, "learning_rate": 1e-06, "loss": 0.25, "step": 8 }, { "epoch": 0.03162055335968379, "step": 8, "train_accuracy_recipient_token": 0.6933333333333334, "train_accuracy_recipient_token_add": 0.8, "train_accuracy_recipient_token_add_total": 5, "train_accuracy_recipient_token_all": 0.7, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 0.7692307692307693, "train_accuracy_recipient_token_get_total": 13, "train_accuracy_recipient_token_h": 1.0, "train_accuracy_recipient_token_h_total": 3, "train_total_number_recipient_token": 75 }, { "epoch": 0.03557312252964427, "grad_norm": 2.135865984530802, "learning_rate": 9.999588943391595e-07, "loss": 0.2792, "step": 9 }, { "epoch": 0.03557312252964427, "step": 9, "train_accuracy_recipient_token": 0.6349206349206349, "train_accuracy_recipient_token_all": 0.6571428571428571, "train_accuracy_recipient_token_all_total": 35, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 2, "train_accuracy_recipient_token_explain": 0.0, "train_accuracy_recipient_token_explain_total": 1, "train_accuracy_recipient_token_find": 0.75, "train_accuracy_recipient_token_find_total": 8, "train_accuracy_recipient_token_get": 0.6666666666666666, "train_accuracy_recipient_token_get_total": 12, "train_total_number_recipient_token": 63 }, { "epoch": 0.039525691699604744, "grad_norm": 12.26339176524548, "learning_rate": 9.9983558411534e-07, "loss": 0.4077, "step": 10 }, { "epoch": 0.039525691699604744, "step": 10, "train_accuracy_recipient_token": 0.6527777777777778, "train_accuracy_recipient_token_all": 0.65, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.7619047619047619, "train_accuracy_recipient_token_get_total": 21, "train_accuracy_recipient_token_verify": 1.0, "train_accuracy_recipient_token_verify_total": 1, "train_total_number_recipient_token": 72 }, { "epoch": 0.043478260869565216, "grad_norm": 2.5853558488363464, "learning_rate": 9.996300896035338e-07, "loss": 0.2641, "step": 11 }, { "epoch": 0.043478260869565216, "step": 11, "train_accuracy_recipient_token": 0.7027027027027027, "train_accuracy_recipient_token_all": 0.65, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_filter": 0.0, "train_accuracy_recipient_token_filter_total": 1, "train_accuracy_recipient_token_find": 0.75, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_generate": 0.8, "train_accuracy_recipient_token_generate_total": 5, "train_accuracy_recipient_token_get": 0.9166666666666666, "train_accuracy_recipient_token_get_total": 12, "train_total_number_recipient_token": 74 }, { "epoch": 0.04743083003952569, "grad_norm": 2.347785831880312, "learning_rate": 9.993424445916922e-07, "loss": 0.206, "step": 12 }, { "epoch": 0.04743083003952569, "step": 12, "train_accuracy_recipient_token": 0.6621621621621622, "train_accuracy_recipient_token_all": 0.6585365853658537, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_find": 0.7777777777777778, "train_accuracy_recipient_token_find_total": 9, "train_accuracy_recipient_token_get": 0.4444444444444444, "train_accuracy_recipient_token_get_total": 9, "train_accuracy_recipient_token_identify": 0.6666666666666666, "train_accuracy_recipient_token_identify_total": 3, "train_total_number_recipient_token": 74 }, { "epoch": 0.05138339920948617, "grad_norm": 1.7892550568881602, "learning_rate": 9.989726963751682e-07, "loss": 0.2269, "step": 13 }, { "epoch": 0.05138339920948617, "step": 13, "train_accuracy_recipient_token": 0.8289473684210527, "train_accuracy_recipient_token_all": 0.7560975609756098, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_compare": 0.5, "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 8, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 18, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 76 }, { "epoch": 0.05533596837944664, "grad_norm": 3.0341081810350192, "learning_rate": 9.985209057489408e-07, "loss": 0.2435, "step": 14 }, { "epoch": 0.05533596837944664, "step": 14, "train_accuracy_recipient_token": 0.8157894736842105, "train_accuracy_recipient_token_all": 0.7380952380952381, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.9523809523809523, "train_accuracy_recipient_token_get_total": 21, "train_accuracy_recipient_token_traffic": 1.0, "train_accuracy_recipient_token_traffic_total": 2, "train_total_number_recipient_token": 76 }, { "epoch": 0.05928853754940711, "grad_norm": 2.5505412107098557, "learning_rate": 9.979871469976195e-07, "loss": 0.2511, "step": 15 }, { "epoch": 0.05928853754940711, "step": 15, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.8780487804878049, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_find": 0.7142857142857143, "train_accuracy_recipient_token_find_total": 7, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 14, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 2, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 2, "train_total_number_recipient_token": 72 }, { "epoch": 0.06324110671936758, "grad_norm": 2.106176367131996, "learning_rate": 9.973715078832286e-07, "loss": 0.2257, "step": 16 }, { "epoch": 0.06324110671936758, "step": 16, "train_accuracy_recipient_token": 0.8405797101449275, "train_accuracy_recipient_token_all": 0.8292682926829268, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.8125, "train_accuracy_recipient_token_get_total": 16, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_total_number_recipient_token": 69 }, { "epoch": 0.06719367588932806, "grad_norm": 3.0814596789592685, "learning_rate": 9.966740896307791e-07, "loss": 0.3841, "step": 17 }, { "epoch": 0.06719367588932806, "step": 17, "train_accuracy_recipient_token": 0.8970588235294118, "train_accuracy_recipient_token_all": 0.8780487804878049, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 7, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 12, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 1, "train_total_number_recipient_token": 68 }, { "epoch": 0.07114624505928854, "grad_norm": 1.8302683207263606, "learning_rate": 9.95895006911623e-07, "loss": 0.1711, "step": 18 }, { "epoch": 0.07114624505928854, "step": 18, "train_accuracy_recipient_token": 0.84, "train_accuracy_recipient_token_all": 0.8571428571428571, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 0.9090909090909091, "train_accuracy_recipient_token_find_total": 11, "train_accuracy_recipient_token_get": 0.9230769230769231, "train_accuracy_recipient_token_get_total": 13, "train_accuracy_recipient_token_identify": 0.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 75 }, { "epoch": 0.07509881422924901, "grad_norm": 1.5911628817962824, "learning_rate": 9.950343878246009e-07, "loss": 0.1658, "step": 19 }, { "epoch": 0.07509881422924901, "step": 19, "train_accuracy_recipient_token": 0.8028169014084507, "train_accuracy_recipient_token_all": 0.875, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_animal": 0.6666666666666666, "train_accuracy_recipient_token_animal_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.8235294117647058, "train_accuracy_recipient_token_get_total": 17, "train_accuracy_recipient_token_plan": 0.0, "train_accuracy_recipient_token_plan_total": 2, "train_total_number_recipient_token": 71 }, { "epoch": 0.07905138339920949, "grad_norm": 1.3004528467355676, "learning_rate": 9.940923738749777e-07, "loss": 0.1674, "step": 20 }, { "epoch": 0.07905138339920949, "step": 20, "train_accuracy_recipient_token": 0.868421052631579, "train_accuracy_recipient_token_all": 0.9534883720930233, "train_accuracy_recipient_token_all_total": 43, "train_accuracy_recipient_token_draw": 1.0, "train_accuracy_recipient_token_draw_total": 5, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.7727272727272727, "train_accuracy_recipient_token_get_total": 22, "train_accuracy_recipient_token_identify": 0.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 76 }, { "epoch": 0.08300395256916997, "grad_norm": 2.1255694355666295, "learning_rate": 9.930691199511773e-07, "loss": 0.2614, "step": 21 }, { "epoch": 0.08300395256916997, "step": 21, "train_accuracy_recipient_token": 0.8333333333333334, "train_accuracy_recipient_token_all": 0.8636363636363636, "train_accuracy_recipient_token_all_total": 44, "train_accuracy_recipient_token_find": 0.7142857142857143, "train_accuracy_recipient_token_find_total": 7, "train_accuracy_recipient_token_get": 0.9375, "train_accuracy_recipient_token_get_total": 16, "train_accuracy_recipient_token_identify": 0.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_schedule": 0.5, "train_accuracy_recipient_token_schedule_total": 2, "train_total_number_recipient_token": 72 }, { "epoch": 0.08695652173913043, "grad_norm": 2.785438041823739, "learning_rate": 9.919647942993147e-07, "loss": 0.2138, "step": 22 }, { "epoch": 0.08695652173913043, "step": 22, "train_accuracy_recipient_token": 0.7941176470588235, "train_accuracy_recipient_token_all": 0.8095238095238095, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_event": 0.6666666666666666, "train_accuracy_recipient_token_event_total": 3, "train_accuracy_recipient_token_find": 0.8333333333333334, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 0.7857142857142857, "train_accuracy_recipient_token_get_total": 14, "train_accuracy_recipient_token_im": 1.0, "train_accuracy_recipient_token_im_total": 1, "train_total_number_recipient_token": 68 }, { "epoch": 0.09090909090909091, "grad_norm": 1.9455454123909355, "learning_rate": 9.907795784955326e-07, "loss": 0.308, "step": 23 }, { "epoch": 0.09090909090909091, "step": 23, "train_accuracy_recipient_token": 0.8571428571428571, "train_accuracy_recipient_token_all": 0.926829268292683, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_convert": 0.8, "train_accuracy_recipient_token_convert_total": 5, "train_accuracy_recipient_token_filter": 0.6, "train_accuracy_recipient_token_filter_total": 5, "train_accuracy_recipient_token_find": 0.8333333333333334, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 0.9, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 77 }, { "epoch": 0.09486166007905138, "grad_norm": 1.9748631586883263, "learning_rate": 9.895136674161464e-07, "loss": 0.2114, "step": 24 }, { "epoch": 0.09486166007905138, "step": 24, "train_accuracy_recipient_token": 0.9178082191780822, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.9545454545454546, "train_accuracy_recipient_token_get_total": 22, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 2, "train_total_number_recipient_token": 73 }, { "epoch": 0.09881422924901186, "grad_norm": 1.795549185533576, "learning_rate": 9.88167269205602e-07, "loss": 0.2272, "step": 25 }, { "epoch": 0.09881422924901186, "step": 25, "train_accuracy_recipient_token": 0.8904109589041096, "train_accuracy_recipient_token_all": 0.8809523809523809, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_assert": 1.0, "train_accuracy_recipient_token_assert_total": 5, "train_accuracy_recipient_token_describe": 1.0, "train_accuracy_recipient_token_describe_total": 1, "train_accuracy_recipient_token_get": 0.8695652173913043, "train_accuracy_recipient_token_get_total": 23, "train_accuracy_recipient_token_mount": 1.0, "train_accuracy_recipient_token_mount_total": 2, "train_total_number_recipient_token": 73 }, { "epoch": 0.10276679841897234, "grad_norm": 1.9443408423782935, "learning_rate": 9.867406052422523e-07, "loss": 0.2074, "step": 26 }, { "epoch": 0.10276679841897234, "step": 26, "train_accuracy_recipient_token": 0.9078947368421053, "train_accuracy_recipient_token_all": 0.925, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.9523809523809523, "train_accuracy_recipient_token_get_total": 21, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 2, "train_total_number_recipient_token": 76 }, { "epoch": 0.1067193675889328, "grad_norm": 5.598767294766525, "learning_rate": 9.852339101019572e-07, "loss": 0.23, "step": 27 }, { "epoch": 0.1067193675889328, "step": 27, "train_accuracy_recipient_token": 0.9452054794520548, "train_accuracy_recipient_token_all": 0.926829268292683, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_calculate": 0.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 25, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_total_number_recipient_token": 73 }, { "epoch": 0.11067193675889328, "grad_norm": 1.973121270993962, "learning_rate": 9.836474315195147e-07, "loss": 0.2007, "step": 28 }, { "epoch": 0.11067193675889328, "step": 28, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9512195121951219, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 19, "train_accuracy_recipient_token_plant": 1.0, "train_accuracy_recipient_token_plant_total": 3, "train_accuracy_recipient_token_urban": 0.6666666666666666, "train_accuracy_recipient_token_urban_total": 3, "train_total_number_recipient_token": 76 }, { "epoch": 0.11462450592885376, "grad_norm": 2.5420962945090495, "learning_rate": 9.819814303479267e-07, "loss": 0.2173, "step": 29 }, { "epoch": 0.11462450592885376, "step": 29, "train_accuracy_recipient_token": 0.9452054794520548, "train_accuracy_recipient_token_all": 0.926829268292683, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 7, "train_accuracy_recipient_token_get": 0.9, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 73 }, { "epoch": 0.11857707509881422, "grad_norm": 3.129518065720253, "learning_rate": 9.802361805155097e-07, "loss": 0.2486, "step": 30 }, { "epoch": 0.11857707509881422, "step": 30, "train_accuracy_recipient_token": 0.8846153846153846, "train_accuracy_recipient_token_all": 0.8888888888888888, "train_accuracy_recipient_token_all_total": 45, "train_accuracy_recipient_token_get": 0.92, "train_accuracy_recipient_token_get_total": 25, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 3, "train_accuracy_recipient_token_setup": 0.0, "train_accuracy_recipient_token_setup_total": 1, "train_accuracy_recipient_token_translate": 0.5, "train_accuracy_recipient_token_translate_total": 2, "train_total_number_recipient_token": 78 }, { "epoch": 0.1225296442687747, "grad_norm": 2.267039732057759, "learning_rate": 9.784119689808542e-07, "loss": 0.2099, "step": 31 }, { "epoch": 0.1225296442687747, "step": 31, "train_accuracy_recipient_token": 0.9054054054054054, "train_accuracy_recipient_token_all": 0.875, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 20, "train_accuracy_recipient_token_recommend": 0.75, "train_accuracy_recipient_token_recommend_total": 4, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 2, "train_total_number_recipient_token": 74 }, { "epoch": 0.12648221343873517, "grad_norm": 2.9755127024237615, "learning_rate": 9.765090956856435e-07, "loss": 0.2367, "step": 32 }, { "epoch": 0.12648221343873517, "step": 32, "train_accuracy_recipient_token": 0.8955223880597015, "train_accuracy_recipient_token_all": 0.8604651162790697, "train_accuracy_recipient_token_all_total": 43, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 2, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_get": 0.9523809523809523, "train_accuracy_recipient_token_get_total": 21, "train_total_number_recipient_token": 67 }, { "epoch": 0.13043478260869565, "grad_norm": 2.1208792730825605, "learning_rate": 9.745278735053343e-07, "loss": 0.29, "step": 33 }, { "epoch": 0.13043478260869565, "step": 33, "train_accuracy_recipient_token": 0.9342105263157895, "train_accuracy_recipient_token_all": 0.9024390243902439, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_artifact": 1.0, "train_accuracy_recipient_token_artifact_total": 3, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 5, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 7, "train_accuracy_recipient_token_get": 0.9, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 76 }, { "epoch": 0.13438735177865613, "grad_norm": 2.090119816024171, "learning_rate": 9.724686281977146e-07, "loss": 0.2518, "step": 34 }, { "epoch": 0.13438735177865613, "step": 34, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.975, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_ex": 1.0, "train_accuracy_recipient_token_ex_total": 3, "train_accuracy_recipient_token_filter": 1.0, "train_accuracy_recipient_token_filter_total": 4, "train_accuracy_recipient_token_find": 0.8, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 0.9090909090909091, "train_accuracy_recipient_token_get_total": 11, "train_total_number_recipient_token": 76 }, { "epoch": 0.1383399209486166, "grad_norm": 2.0882804371111168, "learning_rate": 9.703316983493412e-07, "loss": 0.1936, "step": 35 }, { "epoch": 0.1383399209486166, "step": 35, "train_accuracy_recipient_token": 0.8918918918918919, "train_accuracy_recipient_token_all": 0.8604651162790697, "train_accuracy_recipient_token_all_total": 43, "train_accuracy_recipient_token_city": 1.0, "train_accuracy_recipient_token_city_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 17, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_total_number_recipient_token": 74 }, { "epoch": 0.1422924901185771, "grad_norm": 2.333154966712209, "learning_rate": 9.681174353198686e-07, "loss": 0.2421, "step": 36 }, { "epoch": 0.1422924901185771, "step": 36, "train_accuracy_recipient_token": 0.8947368421052632, "train_accuracy_recipient_token_all": 0.8536585365853658, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 8, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 14, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_total_number_recipient_token": 76 }, { "epoch": 0.14624505928853754, "grad_norm": 2.16958321607675, "learning_rate": 9.658262031842769e-07, "loss": 0.2089, "step": 37 }, { "epoch": 0.14624505928853754, "step": 37, "train_accuracy_recipient_token": 0.8769230769230769, "train_accuracy_recipient_token_all": 0.8157894736842105, "train_accuracy_recipient_token_all_total": 38, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 2, "train_accuracy_recipient_token_find": 0.75, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_g": 1.0, "train_accuracy_recipient_token_g_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 65 }, { "epoch": 0.15019762845849802, "grad_norm": 2.3081856768741553, "learning_rate": 9.634583786730108e-07, "loss": 0.1416, "step": 38 }, { "epoch": 0.15019762845849802, "step": 38, "train_accuracy_recipient_token": 0.8714285714285714, "train_accuracy_recipient_token_all": 0.8048780487804879, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 7, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 4, "train_total_number_recipient_token": 70 }, { "epoch": 0.1541501976284585, "grad_norm": 2.8475202781752302, "learning_rate": 9.610143511100354e-07, "loss": 0.2618, "step": 39 }, { "epoch": 0.1541501976284585, "step": 39, "train_accuracy_recipient_token": 0.9324324324324325, "train_accuracy_recipient_token_all": 0.9024390243902439, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 10, "train_accuracy_recipient_token_h": 1.0, "train_accuracy_recipient_token_h_total": 3, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 2, "train_total_number_recipient_token": 74 }, { "epoch": 0.15810276679841898, "grad_norm": 1.528558740512255, "learning_rate": 9.584945223488226e-07, "loss": 0.1757, "step": 40 }, { "epoch": 0.15810276679841898, "step": 40, "train_accuracy_recipient_token": 0.9295774647887324, "train_accuracy_recipient_token_all": 0.925, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_flag": 1.0, "train_accuracy_recipient_token_flag_total": 3, "train_accuracy_recipient_token_generate": 0.6666666666666666, "train_accuracy_recipient_token_generate_total": 3, "train_accuracy_recipient_token_get": 0.9545454545454546, "train_accuracy_recipient_token_get_total": 22, "train_total_number_recipient_token": 71 }, { "epoch": 0.16205533596837945, "grad_norm": 1.5573159240398418, "learning_rate": 9.558993067062784e-07, "loss": 0.2089, "step": 41 }, { "epoch": 0.16205533596837945, "step": 41, "train_accuracy_recipient_token": 0.9295774647887324, "train_accuracy_recipient_token_all": 0.925, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_build": 0.5, "train_accuracy_recipient_token_build_total": 2, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 2, "train_accuracy_recipient_token_get": 0.9523809523809523, "train_accuracy_recipient_token_get_total": 21, "train_accuracy_recipient_token_log": 1.0, "train_accuracy_recipient_token_log_total": 1, "train_total_number_recipient_token": 71 }, { "epoch": 0.16600790513833993, "grad_norm": 2.8547828361907115, "learning_rate": 9.53229130894619e-07, "loss": 0.2411, "step": 42 }, { "epoch": 0.16600790513833993, "step": 42, "train_accuracy_recipient_token": 0.9466666666666667, "train_accuracy_recipient_token_all": 0.9512195121951219, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 0.9411764705882353, "train_accuracy_recipient_token_get_total": 17, "train_accuracy_recipient_token_number": 1.0, "train_accuracy_recipient_token_number_total": 4, "train_accuracy_recipient_token_remove": 1.0, "train_accuracy_recipient_token_remove_total": 1, "train_total_number_recipient_token": 75 }, { "epoch": 0.16996047430830039, "grad_norm": 2.247423392105134, "learning_rate": 9.504844339512094e-07, "loss": 0.2018, "step": 43 }, { "epoch": 0.16996047430830039, "step": 43, "train_accuracy_recipient_token": 0.9333333333333333, "train_accuracy_recipient_token_all": 0.9285714285714286, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_ex": 1.0, "train_accuracy_recipient_token_ex_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_format": 1.0, "train_accuracy_recipient_token_format_total": 1, "train_accuracy_recipient_token_get": 0.9523809523809523, "train_accuracy_recipient_token_get_total": 21, "train_total_number_recipient_token": 75 }, { "epoch": 0.17391304347826086, "grad_norm": 1.2617503064818263, "learning_rate": 9.476656671663766e-07, "loss": 0.1665, "step": 44 }, { "epoch": 0.17391304347826086, "step": 44, "train_accuracy_recipient_token": 0.9305555555555556, "train_accuracy_recipient_token_all": 0.925, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_arch": 0.0, "train_accuracy_recipient_token_arch_total": 1, "train_accuracy_recipient_token_building": 1.0, "train_accuracy_recipient_token_building_total": 1, "train_accuracy_recipient_token_find": 0.8333333333333334, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 18, "train_total_number_recipient_token": 72 }, { "epoch": 0.17786561264822134, "grad_norm": 1.4958649701923172, "learning_rate": 9.447732940092059e-07, "loss": 0.1582, "step": 45 }, { "epoch": 0.17786561264822134, "step": 45, "train_accuracy_recipient_token": 0.9178082191780822, "train_accuracy_recipient_token_all": 0.9024390243902439, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_animal": 1.0, "train_accuracy_recipient_token_animal_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_accuracy_recipient_token_trans": 1.0, "train_accuracy_recipient_token_trans_total": 3, "train_total_number_recipient_token": 73 }, { "epoch": 0.18181818181818182, "grad_norm": 1.2971539944131008, "learning_rate": 9.418077900513376e-07, "loss": 0.1704, "step": 46 }, { "epoch": 0.18181818181818182, "step": 46, "train_accuracy_recipient_token": 0.9166666666666666, "train_accuracy_recipient_token_all": 0.9047619047619048, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_find": 0.8333333333333334, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 16, "train_accuracy_recipient_token_par": 1.0, "train_accuracy_recipient_token_par_total": 2, "train_accuracy_recipient_token_plant": 0.75, "train_accuracy_recipient_token_plant_total": 4, "train_total_number_recipient_token": 72 }, { "epoch": 0.1857707509881423, "grad_norm": 1.8711399399186879, "learning_rate": 9.387696428887715e-07, "loss": 0.2677, "step": 47 }, { "epoch": 0.1857707509881423, "step": 47, "train_accuracy_recipient_token": 0.9324324324324325, "train_accuracy_recipient_token_all": 0.9523809523809523, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_compare": 0.5, "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 7, "train_accuracy_recipient_token_flower": 0.6666666666666666, "train_accuracy_recipient_token_flower_total": 3, "train_accuracy_recipient_token_get": 0.9166666666666666, "train_accuracy_recipient_token_get_total": 12, "train_total_number_recipient_token": 74 }, { "epoch": 0.18972332015810275, "grad_norm": 4.988758419816404, "learning_rate": 9.356593520616946e-07, "loss": 0.2285, "step": 48 }, { "epoch": 0.18972332015810275, "step": 48, "train_accuracy_recipient_token": 0.9452054794520548, "train_accuracy_recipient_token_all": 0.9024390243902439, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 22, "train_accuracy_recipient_token_gr": 1.0, "train_accuracy_recipient_token_gr_total": 5, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 73 }, { "epoch": 0.19367588932806323, "grad_norm": 2.3197243712561746, "learning_rate": 9.324774289723467e-07, "loss": 0.2168, "step": 49 }, { "epoch": 0.19367588932806323, "step": 49, "train_accuracy_recipient_token": 0.8571428571428571, "train_accuracy_recipient_token_all": 0.8444444444444444, "train_accuracy_recipient_token_all_total": 45, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_compose": 1.0, "train_accuracy_recipient_token_compose_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_accuracy_recipient_token_multiply": 0.0, "train_accuracy_recipient_token_multiply_total": 2, "train_total_number_recipient_token": 70 }, { "epoch": 0.1976284584980237, "grad_norm": 2.6244311851742905, "learning_rate": 9.29224396800933e-07, "loss": 0.234, "step": 50 }, { "epoch": 0.1976284584980237, "step": 50, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.95, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 0.9565217391304348, "train_accuracy_recipient_token_get_total": 23, "train_total_number_recipient_token": 76 }, { "epoch": 0.2015810276679842, "grad_norm": 2.779011972479135, "learning_rate": 9.259007904196021e-07, "loss": 0.1382, "step": 51 }, { "epoch": 0.2015810276679842, "step": 51, "train_accuracy_recipient_token": 0.9305555555555556, "train_accuracy_recipient_token_all": 0.925, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_armor": 1.0, "train_accuracy_recipient_token_armor_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 0.9230769230769231, "train_accuracy_recipient_token_get_total": 13, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_total_number_recipient_token": 72 }, { "epoch": 0.20553359683794467, "grad_norm": 2.177256989435761, "learning_rate": 9.225071563045006e-07, "loss": 0.2344, "step": 52 }, { "epoch": 0.20553359683794467, "step": 52, "train_accuracy_recipient_token": 0.9333333333333333, "train_accuracy_recipient_token_all": 0.9318181818181818, "train_accuracy_recipient_token_all_total": 44, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 2, "train_accuracy_recipient_token_customize": 0.5, "train_accuracy_recipient_token_customize_total": 2, "train_accuracy_recipient_token_get": 0.9375, "train_accuracy_recipient_token_get_total": 16, "train_accuracy_recipient_token_mon": 1.0, "train_accuracy_recipient_token_mon_total": 4, "train_total_number_recipient_token": 75 }, { "epoch": 0.20948616600790515, "grad_norm": 2.8595931173212144, "learning_rate": 9.190440524459202e-07, "loss": 0.1729, "step": 53 }, { "epoch": 0.20948616600790515, "step": 53, "train_accuracy_recipient_token": 0.9285714285714286, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_flower": 0.75, "train_accuracy_recipient_token_flower_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 18, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_total_number_recipient_token": 70 }, { "epoch": 0.2134387351778656, "grad_norm": 2.545206096525683, "learning_rate": 9.155120482565519e-07, "loss": 0.1748, "step": 54 }, { "epoch": 0.2134387351778656, "step": 54, "train_accuracy_recipient_token": 0.9452054794520548, "train_accuracy_recipient_token_all": 0.9024390243902439, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_draw": 1.0, "train_accuracy_recipient_token_draw_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 23, "train_accuracy_recipient_token_store": 1.0, "train_accuracy_recipient_token_store_total": 2, "train_total_number_recipient_token": 73 }, { "epoch": 0.21739130434782608, "grad_norm": 1.7819898307030615, "learning_rate": 9.119117244778607e-07, "loss": 0.1887, "step": 55 }, { "epoch": 0.21739130434782608, "step": 55, "train_accuracy_recipient_token": 0.9014084507042254, "train_accuracy_recipient_token_all": 0.8780487804878049, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 0.9, "train_accuracy_recipient_token_get_total": 10, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 3, "train_accuracy_recipient_token_stat": 0.75, "train_accuracy_recipient_token_stat_total": 4, "train_total_number_recipient_token": 71 }, { "epoch": 0.22134387351778656, "grad_norm": 2.3270808188875343, "learning_rate": 9.082436730845993e-07, "loss": 0.2456, "step": 56 }, { "epoch": 0.22134387351778656, "step": 56, "train_accuracy_recipient_token": 0.8787878787878788, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_f": 0.5, "train_accuracy_recipient_token_f_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.8888888888888888, "train_accuracy_recipient_token_get_total": 9, "train_accuracy_recipient_token_gr": 0.6666666666666666, "train_accuracy_recipient_token_gr_total": 3, "train_total_number_recipient_token": 66 }, { "epoch": 0.22529644268774704, "grad_norm": 1.6199619248707686, "learning_rate": 9.045084971874737e-07, "loss": 0.2076, "step": 57 }, { "epoch": 0.22529644268774704, "step": 57, "train_accuracy_recipient_token": 0.8493150684931506, "train_accuracy_recipient_token_all": 0.85, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_calculate": 0.5, "train_accuracy_recipient_token_calculate_total": 2, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 0.8, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 0.8636363636363636, "train_accuracy_recipient_token_get_total": 22, "train_total_number_recipient_token": 73 }, { "epoch": 0.22924901185770752, "grad_norm": 4.728708432665532, "learning_rate": 9.007068109339783e-07, "loss": 0.2359, "step": 58 }, { "epoch": 0.22924901185770752, "step": 58, "train_accuracy_recipient_token": 0.9324324324324325, "train_accuracy_recipient_token_all": 0.9523809523809523, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_flower": 1.0, "train_accuracy_recipient_token_flower_total": 4, "train_accuracy_recipient_token_get": 0.9166666666666666, "train_accuracy_recipient_token_get_total": 24, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 74 }, { "epoch": 0.233201581027668, "grad_norm": 2.643761932851136, "learning_rate": 8.968392394074163e-07, "loss": 0.2311, "step": 59 }, { "epoch": 0.233201581027668, "step": 59, "train_accuracy_recipient_token": 0.9873417721518988, "train_accuracy_recipient_token_all": 0.975609756097561, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_arc": 1.0, "train_accuracy_recipient_token_arc_total": 4, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 18, "train_total_number_recipient_token": 79 }, { "epoch": 0.23715415019762845, "grad_norm": 1.6208595934050558, "learning_rate": 8.929064185241212e-07, "loss": 0.2563, "step": 60 }, { "epoch": 0.23715415019762845, "step": 60, "train_accuracy_recipient_token": 0.8985507246376812, "train_accuracy_recipient_token_all": 0.875, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 2, "train_accuracy_recipient_token_find": 0.8888888888888888, "train_accuracy_recipient_token_find_total": 9, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_sc": 1.0, "train_accuracy_recipient_token_sc_total": 2, "train_total_number_recipient_token": 69 }, { "epoch": 0.24110671936758893, "grad_norm": 1.9005836945460637, "learning_rate": 8.889089949288986e-07, "loss": 0.1524, "step": 61 }, { "epoch": 0.24110671936758893, "step": 61, "train_accuracy_recipient_token": 0.958904109589041, "train_accuracy_recipient_token_all": 0.975, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 2, "train_accuracy_recipient_token_find": 0.8888888888888888, "train_accuracy_recipient_token_find_total": 9, "train_accuracy_recipient_token_get": 0.9285714285714286, "train_accuracy_recipient_token_get_total": 14, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 2, "train_total_number_recipient_token": 73 }, { "epoch": 0.2450592885375494, "grad_norm": 3.087443078632211, "learning_rate": 8.84847625888703e-07, "loss": 0.1871, "step": 62 }, { "epoch": 0.2450592885375494, "step": 62, "train_accuracy_recipient_token": 0.9583333333333334, "train_accuracy_recipient_token_all": 0.9743589743589743, "train_accuracy_recipient_token_all_total": 39, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_find": 0.7142857142857143, "train_accuracy_recipient_token_find_total": 7, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 17, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 2, "train_total_number_recipient_token": 72 }, { "epoch": 0.2490118577075099, "grad_norm": 1.4965895983954287, "learning_rate": 8.807229791845671e-07, "loss": 0.1752, "step": 63 }, { "epoch": 0.2490118577075099, "step": 63, "train_accuracy_recipient_token": 0.9436619718309859, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_compose": 1.0, "train_accuracy_recipient_token_compose_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 18, "train_accuracy_recipient_token_research": 1.0, "train_accuracy_recipient_token_research_total": 2, "train_total_number_recipient_token": 71 }, { "epoch": 0.25296442687747034, "grad_norm": 1.987492972787394, "learning_rate": 8.765357330018055e-07, "loss": 0.1991, "step": 64 }, { "epoch": 0.25296442687747034, "step": 64, "train_accuracy_recipient_token": 0.9459459459459459, "train_accuracy_recipient_token_all": 0.9302325581395349, "train_accuracy_recipient_token_all_total": 43, "train_accuracy_recipient_token_describe": 1.0, "train_accuracy_recipient_token_describe_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_flag": 1.0, "train_accuracy_recipient_token_flag_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 16, "train_total_number_recipient_token": 74 }, { "epoch": 0.25691699604743085, "grad_norm": 3.587213559295727, "learning_rate": 8.722865758185035e-07, "loss": 0.2138, "step": 65 }, { "epoch": 0.25691699604743085, "step": 65, "train_accuracy_recipient_token": 0.9041095890410958, "train_accuracy_recipient_token_all": 0.9047619047619048, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_ass": 0.5, "train_accuracy_recipient_token_ass_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 17, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_total_number_recipient_token": 73 }, { "epoch": 0.2608695652173913, "grad_norm": 1.9190860764608675, "learning_rate": 8.679762062923175e-07, "loss": 0.2557, "step": 66 }, { "epoch": 0.2608695652173913, "step": 66, "train_accuracy_recipient_token": 0.9436619718309859, "train_accuracy_recipient_token_all": 0.95, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_find": 0.8, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 0.9444444444444444, "train_accuracy_recipient_token_get_total": 18, "train_accuracy_recipient_token_track": 1.0, "train_accuracy_recipient_token_track_total": 2, "train_accuracy_recipient_token_write": 1.0, "train_accuracy_recipient_token_write_total": 1, "train_total_number_recipient_token": 71 }, { "epoch": 0.2648221343873518, "grad_norm": 3.3915027238243534, "learning_rate": 8.636053331455986e-07, "loss": 0.2104, "step": 67 }, { "epoch": 0.2648221343873518, "step": 67, "train_accuracy_recipient_token": 0.9315068493150684, "train_accuracy_recipient_token_all": 0.9069767441860465, "train_accuracy_recipient_token_all_total": 43, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 11, "train_accuracy_recipient_token_m": 1.0, "train_accuracy_recipient_token_m_total": 3, "train_accuracy_recipient_token_sport": 1.0, "train_accuracy_recipient_token_sport_total": 3, "train_total_number_recipient_token": 73 }, { "epoch": 0.26877470355731226, "grad_norm": 1.732729844649354, "learning_rate": 8.591746750488637e-07, "loss": 0.1679, "step": 68 }, { "epoch": 0.26877470355731226, "step": 68, "train_accuracy_recipient_token": 0.9411764705882353, "train_accuracy_recipient_token_all": 0.95, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_ele": 1.0, "train_accuracy_recipient_token_ele_total": 2, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 19, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_total_number_recipient_token": 68 }, { "epoch": 0.2727272727272727, "grad_norm": 1.6629835042359014, "learning_rate": 8.546849605026288e-07, "loss": 0.2198, "step": 69 }, { "epoch": 0.2727272727272727, "step": 69, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.9024390243902439, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 11, "train_accuracy_recipient_token_translate": 0.6666666666666666, "train_accuracy_recipient_token_translate_total": 3, "train_total_number_recipient_token": 72 }, { "epoch": 0.2766798418972332, "grad_norm": 2.5515096421849157, "learning_rate": 8.501369277176274e-07, "loss": 0.2523, "step": 70 }, { "epoch": 0.2766798418972332, "step": 70, "train_accuracy_recipient_token": 0.9178082191780822, "train_accuracy_recipient_token_all": 0.95, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.95, "train_accuracy_recipient_token_get_total": 20, "train_total_number_recipient_token": 73 }, { "epoch": 0.28063241106719367, "grad_norm": 2.7317626681477014, "learning_rate": 8.455313244934324e-07, "loss": 0.1696, "step": 71 }, { "epoch": 0.28063241106719367, "step": 71, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.8571428571428571, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_f": 1.0, "train_accuracy_recipient_token_f_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_sh": 1.0, "train_accuracy_recipient_token_sh_total": 3, "train_total_number_recipient_token": 70 }, { "epoch": 0.2845849802371542, "grad_norm": 1.884067734296459, "learning_rate": 8.408689080954997e-07, "loss": 0.2669, "step": 72 }, { "epoch": 0.2845849802371542, "step": 72, "train_accuracy_recipient_token": 0.96, "train_accuracy_recipient_token_all": 0.926829268292683, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 19, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_accuracy_recipient_token_verify": 1.0, "train_accuracy_recipient_token_verify_total": 1, "train_total_number_recipient_token": 75 }, { "epoch": 0.2885375494071146, "grad_norm": 1.392479035649444, "learning_rate": 8.361504451306584e-07, "loss": 0.1601, "step": 73 }, { "epoch": 0.2885375494071146, "step": 73, "train_accuracy_recipient_token": 0.8676470588235294, "train_accuracy_recipient_token_all": 0.8571428571428571, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_analyze": 0.0, "train_accuracy_recipient_token_analyze_total": 2, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 11, "train_total_number_recipient_token": 68 }, { "epoch": 0.2924901185770751, "grad_norm": 2.0714017485775207, "learning_rate": 8.313767114210615e-07, "loss": 0.2302, "step": 74 }, { "epoch": 0.2924901185770751, "step": 74, "train_accuracy_recipient_token": 0.958904109589041, "train_accuracy_recipient_token_all": 0.95, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 4, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_find": 0.8, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 14, "train_total_number_recipient_token": 73 }, { "epoch": 0.2964426877470356, "grad_norm": 2.276636448944012, "learning_rate": 8.265484918766242e-07, "loss": 0.2046, "step": 75 }, { "epoch": 0.2964426877470356, "step": 75, "train_accuracy_recipient_token": 0.9466666666666667, "train_accuracy_recipient_token_all": 0.9743589743589743, "train_accuracy_recipient_token_all_total": 39, "train_accuracy_recipient_token_facility": 1.0, "train_accuracy_recipient_token_facility_total": 3, "train_accuracy_recipient_token_find": 0.75, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 0.9333333333333333, "train_accuracy_recipient_token_get_total": 15, "train_accuracy_recipient_token_park": 1.0, "train_accuracy_recipient_token_park_total": 3, "train_total_number_recipient_token": 75 }, { "epoch": 0.30039525691699603, "grad_norm": 2.031928550650522, "learning_rate": 8.21666580365967e-07, "loss": 0.122, "step": 76 }, { "epoch": 0.30039525691699603, "step": 76, "train_accuracy_recipient_token": 0.8873239436619719, "train_accuracy_recipient_token_all": 0.9047619047619048, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_city": 1.0, "train_accuracy_recipient_token_city_total": 2, "train_accuracy_recipient_token_detect": 1.0, "train_accuracy_recipient_token_detect_total": 3, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 12, "train_total_number_recipient_token": 71 }, { "epoch": 0.30434782608695654, "grad_norm": 1.734445106532831, "learning_rate": 8.16731779585885e-07, "loss": 0.2224, "step": 77 }, { "epoch": 0.30434782608695654, "step": 77, "train_accuracy_recipient_token": 0.918918918918919, "train_accuracy_recipient_token_all": 0.9285714285714286, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_event": 1.0, "train_accuracy_recipient_token_event_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.9375, "train_accuracy_recipient_token_get_total": 16, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_total_number_recipient_token": 74 }, { "epoch": 0.308300395256917, "grad_norm": 2.29431424801668, "learning_rate": 8.117449009293668e-07, "loss": 0.2031, "step": 78 }, { "epoch": 0.308300395256917, "step": 78, "train_accuracy_recipient_token": 0.972972972972973, "train_accuracy_recipient_token_all": 0.975, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 5, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 0.9333333333333333, "train_accuracy_recipient_token_get_total": 15, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 3, "train_total_number_recipient_token": 74 }, { "epoch": 0.31225296442687744, "grad_norm": 1.4545086251819752, "learning_rate": 8.067067643521833e-07, "loss": 0.1887, "step": 79 }, { "epoch": 0.31225296442687744, "step": 79, "train_accuracy_recipient_token": 0.9861111111111112, "train_accuracy_recipient_token_all": 0.975609756097561, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 5, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 18, "train_accuracy_recipient_token_verify": 1.0, "train_accuracy_recipient_token_verify_total": 1, "train_total_number_recipient_token": 72 }, { "epoch": 0.31620553359683795, "grad_norm": 2.7510123556921102, "learning_rate": 8.016181982380681e-07, "loss": 0.2102, "step": 80 }, { "epoch": 0.31620553359683795, "step": 80, "train_accuracy_recipient_token": 0.9154929577464789, "train_accuracy_recipient_token_all": 0.9069767441860465, "train_accuracy_recipient_token_all_total": 43, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.9230769230769231, "train_accuracy_recipient_token_get_total": 13, "train_accuracy_recipient_token_search": 0.6666666666666666, "train_accuracy_recipient_token_search_total": 3, "train_total_number_recipient_token": 71 }, { "epoch": 0.3201581027667984, "grad_norm": 2.241875603545219, "learning_rate": 7.964800392625128e-07, "loss": 0.2761, "step": 81 }, { "epoch": 0.3201581027667984, "step": 81, "train_accuracy_recipient_token": 0.8235294117647058, "train_accuracy_recipient_token_all": 0.8333333333333334, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 0.8888888888888888, "train_accuracy_recipient_token_get_total": 9, "train_accuracy_recipient_token_histor": 1.0, "train_accuracy_recipient_token_histor_total": 2, "train_total_number_recipient_token": 68 }, { "epoch": 0.3241106719367589, "grad_norm": 4.047547980360494, "learning_rate": 7.91293132255198e-07, "loss": 0.2368, "step": 82 }, { "epoch": 0.3241106719367589, "step": 82, "train_accuracy_recipient_token": 0.9142857142857143, "train_accuracy_recipient_token_all": 0.8666666666666667, "train_accuracy_recipient_token_all_total": 45, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 14, "train_accuracy_recipient_token_race": 1.0, "train_accuracy_recipient_token_race_total": 2, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 1, "train_total_number_recipient_token": 70 }, { "epoch": 0.32806324110671936, "grad_norm": 2.8790385864152412, "learning_rate": 7.860583300610847e-07, "loss": 0.22, "step": 83 }, { "epoch": 0.32806324110671936, "step": 83, "train_accuracy_recipient_token": 0.9066666666666666, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 44, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.9375, "train_accuracy_recipient_token_get_total": 16, "train_accuracy_recipient_token_highlight": 0.5, "train_accuracy_recipient_token_highlight_total": 2, "train_accuracy_recipient_token_remove": 1.0, "train_accuracy_recipient_token_remove_total": 2, "train_total_number_recipient_token": 75 }, { "epoch": 0.33201581027667987, "grad_norm": 2.143665687971683, "learning_rate": 7.807764934001874e-07, "loss": 0.1831, "step": 84 }, { "epoch": 0.33201581027667987, "step": 84, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9473684210526315, "train_accuracy_recipient_token_all_total": 38, "train_accuracy_recipient_token_find": 0.8333333333333334, "train_accuracy_recipient_token_find_total": 12, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 21, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_total_number_recipient_token": 72 }, { "epoch": 0.33201581027667987, "step": 84, "train_accuracy_recipient_token": 0.9295774647887324, "train_accuracy_recipient_token_all": 0.9523809523809523, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_find": 0.8, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 15, "train_accuracy_recipient_token_histor": 0.6666666666666666, "train_accuracy_recipient_token_histor_total": 3, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 3, "train_total_number_recipient_token": 71 }, { "epoch": 0.33201581027667987, "step": 84, "train_accuracy_recipient_token": 0.9166666666666666, "train_accuracy_recipient_token_all": 0.8695652173913043, "train_accuracy_recipient_token_all_total": 46, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 8, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_accuracy_recipient_token_interpret": 1.0, "train_accuracy_recipient_token_interpret_total": 2, "train_total_number_recipient_token": 72 }, { "epoch": 0.33201581027667987, "step": 84, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 14, "train_accuracy_recipient_token_tour": 1.0, "train_accuracy_recipient_token_tour_total": 4, "train_total_number_recipient_token": 76 }, { "epoch": 0.33201581027667987, "step": 84, "train_accuracy_recipient_token": 0.9722222222222222, "train_accuracy_recipient_token_all": 0.9512195121951219, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_artifact": 1.0, "train_accuracy_recipient_token_artifact_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 20, "train_accuracy_recipient_token_select": 1.0, "train_accuracy_recipient_token_select_total": 2, "train_total_number_recipient_token": 72 }, { "epoch": 0.33201581027667987, "step": 84, "train_accuracy_recipient_token": 0.9285714285714286, "train_accuracy_recipient_token_all": 0.925, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.9375, "train_accuracy_recipient_token_get_total": 16, "train_accuracy_recipient_token_gr": 1.0, "train_accuracy_recipient_token_gr_total": 3, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_total_number_recipient_token": 70 }, { "epoch": 0.33201581027667987, "step": 84, "train_accuracy_recipient_token": 0.8767123287671232, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 2, "train_accuracy_recipient_token_customize": 0.5, "train_accuracy_recipient_token_customize_total": 4, "train_accuracy_recipient_token_get": 0.8695652173913043, "train_accuracy_recipient_token_get_total": 23, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 73 }, { "epoch": 0.33201581027667987, "step": 84, "train_accuracy_recipient_token": 0.9142857142857143, "train_accuracy_recipient_token_all": 0.8780487804878049, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_eng": 0.5, "train_accuracy_recipient_token_eng_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 11, "train_accuracy_recipient_token_tem": 1.0, "train_accuracy_recipient_token_tem_total": 3, "train_total_number_recipient_token": 70 }, { "epoch": 0.33201581027667987, "step": 84, "train_accuracy_recipient_token": 0.9565217391304348, "train_accuracy_recipient_token_all": 0.95, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_bird": 1.0, "train_accuracy_recipient_token_bird_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_total_number_recipient_token": 69 }, { "epoch": 0.33201581027667987, "step": 84, "train_accuracy_recipient_token": 0.9027777777777778, "train_accuracy_recipient_token_all": 0.875, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 0.9473684210526315, "train_accuracy_recipient_token_get_total": 19, "train_accuracy_recipient_token_park": 1.0, "train_accuracy_recipient_token_park_total": 2, "train_total_number_recipient_token": 72 }, { "epoch": 0.33201581027667987, "step": 84, "train_accuracy_recipient_token": 0.9436619718309859, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_art": 0.6666666666666666, "train_accuracy_recipient_token_art_total": 3, "train_accuracy_recipient_token_fruit": 1.0, "train_accuracy_recipient_token_fruit_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 15, "train_accuracy_recipient_token_object": 0.6666666666666666, "train_accuracy_recipient_token_object_total": 3, "train_total_number_recipient_token": 71 }, { "epoch": 0.33201581027667987, "step": 84, "train_accuracy_recipient_token": 0.927536231884058, "train_accuracy_recipient_token_all": 0.875, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 14, "train_accuracy_recipient_token_organ": 1.0, "train_accuracy_recipient_token_organ_total": 1, "train_accuracy_recipient_token_stage": 1.0, "train_accuracy_recipient_token_stage_total": 4, "train_total_number_recipient_token": 69 }, { "epoch": 0.33201581027667987, "step": 84, "train_accuracy_recipient_token": 0.9555555555555556, "train_accuracy_recipient_token_all": 0.96, "train_accuracy_recipient_token_all_total": 25, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 5, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 10, "train_accuracy_recipient_token_sun": 1.0, "train_accuracy_recipient_token_sun_total": 1, "train_total_number_recipient_token": 45 }, { "epoch": 0.33201581027667987, "eval_accuracy_recipient_token": 0.9312638580931264, "eval_accuracy_recipient_token_all": 0.9278752436647173, "eval_accuracy_recipient_token_all_total": 513, "eval_accuracy_recipient_token_calculate": 1.0, "eval_accuracy_recipient_token_calculate_total": 13, "eval_accuracy_recipient_token_find": 0.9473684210526315, "eval_accuracy_recipient_token_find_total": 57, "eval_accuracy_recipient_token_get": 0.9642857142857143, "eval_accuracy_recipient_token_get_total": 196, "eval_accuracy_recipient_token_identify": 1.0, "eval_accuracy_recipient_token_identify_total": 6, "eval_loss": 0.1849488615989685, "eval_perplexity": 1.0332535241675367, "eval_runtime": 100.5712, "eval_samples_per_second": 1.004, "eval_steps_per_second": 0.129, "eval_total_number_recipient_token": 902, "step": 84 }, { "epoch": 0.33201581027667987, "step": 84, "train_accuracy_recipient_token": 0.9866666666666667, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_dispose": 1.0, "train_accuracy_recipient_token_dispose_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 22, "train_accuracy_recipient_token_replace": 1.0, "train_accuracy_recipient_token_replace_total": 2, "train_total_number_recipient_token": 75 }, { "epoch": 0.3359683794466403, "grad_norm": 3.218408659085263, "learning_rate": 7.754484907260512e-07, "loss": 0.2015, "step": 85 }, { "epoch": 0.3359683794466403, "step": 85, "train_accuracy_recipient_token": 0.9333333333333333, "train_accuracy_recipient_token_all": 0.9512195121951219, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 2, "train_accuracy_recipient_token_bird": 1.0, "train_accuracy_recipient_token_bird_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.9333333333333333, "train_accuracy_recipient_token_get_total": 15, "train_total_number_recipient_token": 75 }, { "epoch": 0.33992094861660077, "grad_norm": 1.4261895418183008, "learning_rate": 7.700751980829601e-07, "loss": 0.1533, "step": 86 }, { "epoch": 0.33992094861660077, "step": 86, "train_accuracy_recipient_token": 0.9466666666666667, "train_accuracy_recipient_token_all": 0.9523809523809523, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_find": 0.8, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 18, "train_accuracy_recipient_token_organ": 1.0, "train_accuracy_recipient_token_organ_total": 1, "train_accuracy_recipient_token_switch": 1.0, "train_accuracy_recipient_token_switch_total": 2, "train_total_number_recipient_token": 75 }, { "epoch": 0.3438735177865613, "grad_norm": 1.5826006582828822, "learning_rate": 7.646574989618937e-07, "loss": 0.1949, "step": 87 }, { "epoch": 0.3438735177865613, "step": 87, "train_accuracy_recipient_token": 0.8840579710144928, "train_accuracy_recipient_token_all": 0.9285714285714286, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 2, "train_accuracy_recipient_token_design": 0.5, "train_accuracy_recipient_token_design_total": 2, "train_accuracy_recipient_token_find": 0.6, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 69 }, { "epoch": 0.34782608695652173, "grad_norm": 1.9679063472871658, "learning_rate": 7.591962841552626e-07, "loss": 0.2892, "step": 88 }, { "epoch": 0.34782608695652173, "step": 88, "train_accuracy_recipient_token": 0.8985507246376812, "train_accuracy_recipient_token_all": 0.9024390243902439, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_get": 0.9411764705882353, "train_accuracy_recipient_token_get_total": 17, "train_accuracy_recipient_token_sc": 1.0, "train_accuracy_recipient_token_sc_total": 2, "train_accuracy_recipient_token_setup": 0.5, "train_accuracy_recipient_token_setup_total": 2, "train_total_number_recipient_token": 69 }, { "epoch": 0.35177865612648224, "grad_norm": 4.390929513600532, "learning_rate": 7.53692451610441e-07, "loss": 0.2698, "step": 89 }, { "epoch": 0.35177865612648224, "step": 89, "train_accuracy_recipient_token": 0.9487179487179487, "train_accuracy_recipient_token_all": 0.975, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_financial": 1.0, "train_accuracy_recipient_token_financial_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 7, "train_accuracy_recipient_token_get": 0.875, "train_accuracy_recipient_token_get_total": 24, "train_total_number_recipient_token": 78 }, { "epoch": 0.3557312252964427, "grad_norm": 1.3709454578510425, "learning_rate": 7.481469062821251e-07, "loss": 0.1301, "step": 90 }, { "epoch": 0.3557312252964427, "step": 90, "train_accuracy_recipient_token": 0.855072463768116, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_calculate": 0.3333333333333333, "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 0.7272727272727273, "train_accuracy_recipient_token_get_total": 11, "train_accuracy_recipient_token_suggest": 1.0, "train_accuracy_recipient_token_suggest_total": 2, "train_total_number_recipient_token": 69 }, { "epoch": 0.35968379446640314, "grad_norm": 1.9772930981295223, "learning_rate": 7.42560559983536e-07, "loss": 0.1788, "step": 91 }, { "epoch": 0.35968379446640314, "step": 91, "train_accuracy_recipient_token": 0.9866666666666667, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 15, "train_accuracy_recipient_token_search": 0.5, "train_accuracy_recipient_token_search_total": 2, "train_total_number_recipient_token": 75 }, { "epoch": 0.36363636363636365, "grad_norm": 1.5008734703027575, "learning_rate": 7.369343312364993e-07, "loss": 0.1581, "step": 92 }, { "epoch": 0.36363636363636365, "step": 92, "train_accuracy_recipient_token": 0.8918918918918919, "train_accuracy_recipient_token_all": 0.8837209302325582, "train_accuracy_recipient_token_all_total": 43, "train_accuracy_recipient_token_build": 1.0, "train_accuracy_recipient_token_build_total": 3, "train_accuracy_recipient_token_find": 0.9, "train_accuracy_recipient_token_find_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_land": 1.0, "train_accuracy_recipient_token_land_total": 4, "train_total_number_recipient_token": 74 }, { "epoch": 0.3675889328063241, "grad_norm": 1.9066144250946504, "learning_rate": 7.312691451204177e-07, "loss": 0.1903, "step": 93 }, { "epoch": 0.3675889328063241, "step": 93, "train_accuracy_recipient_token": 0.9722222222222222, "train_accuracy_recipient_token_all": 0.975, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 17, "train_accuracy_recipient_token_phrase": 1.0, "train_accuracy_recipient_token_phrase_total": 4, "train_total_number_recipient_token": 72 }, { "epoch": 0.3715415019762846, "grad_norm": 1.4922766534026213, "learning_rate": 7.255659331201673e-07, "loss": 0.2043, "step": 94 }, { "epoch": 0.3715415019762846, "step": 94, "train_accuracy_recipient_token": 0.9154929577464789, "train_accuracy_recipient_token_all": 0.8536585365853658, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 21, "train_accuracy_recipient_token_suggest": 1.0, "train_accuracy_recipient_token_suggest_total": 1, "train_total_number_recipient_token": 71 }, { "epoch": 0.37549407114624506, "grad_norm": 1.7741443720907384, "learning_rate": 7.198256329729411e-07, "loss": 0.2124, "step": 95 }, { "epoch": 0.37549407114624506, "step": 95, "train_accuracy_recipient_token": 0.9178082191780822, "train_accuracy_recipient_token_all": 0.9024390243902439, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_estimate": 0.5, "train_accuracy_recipient_token_estimate_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.9523809523809523, "train_accuracy_recipient_token_get_total": 21, "train_total_number_recipient_token": 73 }, { "epoch": 0.3794466403162055, "grad_norm": 2.3773616292420945, "learning_rate": 7.140491885140628e-07, "loss": 0.2016, "step": 96 }, { "epoch": 0.3794466403162055, "step": 96, "train_accuracy_recipient_token": 0.9324324324324325, "train_accuracy_recipient_token_all": 0.925, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_get": 0.9545454545454546, "train_accuracy_recipient_token_get_total": 22, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 3, "train_accuracy_recipient_token_repair": 1.0, "train_accuracy_recipient_token_repair_total": 2, "train_accuracy_recipient_token_restore": 1.0, "train_accuracy_recipient_token_restore_total": 2, "train_total_number_recipient_token": 74 }, { "epoch": 0.383399209486166, "grad_norm": 3.7229183140804243, "learning_rate": 7.082375495217995e-07, "loss": 0.2367, "step": 97 }, { "epoch": 0.383399209486166, "step": 97, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.8292682926829268, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 19, "train_accuracy_recipient_token_log": 1.0, "train_accuracy_recipient_token_log_total": 4, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 1, "train_total_number_recipient_token": 70 }, { "epoch": 0.38735177865612647, "grad_norm": 2.2023881261603173, "learning_rate": 7.023916715611968e-07, "loss": 0.2086, "step": 98 }, { "epoch": 0.38735177865612647, "step": 98, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.95, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_analyze": 0.6666666666666666, "train_accuracy_recipient_token_analyze_total": 3, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 13, "train_total_number_recipient_token": 76 }, { "epoch": 0.391304347826087, "grad_norm": 1.2096315786946172, "learning_rate": 6.965125158269618e-07, "loss": 0.1504, "step": 99 }, { "epoch": 0.391304347826087, "step": 99, "train_accuracy_recipient_token": 0.9459459459459459, "train_accuracy_recipient_token_all": 0.95, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_analyze": 0.6666666666666666, "train_accuracy_recipient_token_analyze_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 17, "train_accuracy_recipient_token_identify": 0.6666666666666666, "train_accuracy_recipient_token_identify_total": 3, "train_accuracy_recipient_token_trans": 1.0, "train_accuracy_recipient_token_trans_total": 3, "train_total_number_recipient_token": 74 }, { "epoch": 0.3952569169960474, "grad_norm": 2.8098193639343925, "learning_rate": 6.906010489854209e-07, "loss": 0.2267, "step": 100 }, { "epoch": 0.3952569169960474, "step": 100, "train_accuracy_recipient_token": 0.974025974025974, "train_accuracy_recipient_token_add": 1.0, "train_accuracy_recipient_token_add_total": 4, "train_accuracy_recipient_token_all": 0.975609756097561, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 0.9375, "train_accuracy_recipient_token_get_total": 16, "train_accuracy_recipient_token_select": 1.0, "train_accuracy_recipient_token_select_total": 2, "train_total_number_recipient_token": 77 }, { "epoch": 0.39920948616600793, "grad_norm": 1.6225621203034952, "learning_rate": 6.846582430155781e-07, "loss": 0.2228, "step": 101 }, { "epoch": 0.39920948616600793, "step": 101, "train_accuracy_recipient_token": 0.96, "train_accuracy_recipient_token_all": 0.975609756097561, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.9285714285714286, "train_accuracy_recipient_token_get_total": 28, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_total_number_recipient_token": 75 }, { "epoch": 0.4031620553359684, "grad_norm": 1.7599291455571704, "learning_rate": 6.786850750493005e-07, "loss": 0.193, "step": 102 }, { "epoch": 0.4031620553359684, "step": 102, "train_accuracy_recipient_token": 0.948051948051948, "train_accuracy_recipient_token_all": 0.9302325581395349, "train_accuracy_recipient_token_all_total": 43, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 15, "train_accuracy_recipient_token_pay": 1.0, "train_accuracy_recipient_token_pay_total": 2, "train_accuracy_recipient_token_tractor": 1.0, "train_accuracy_recipient_token_tractor_total": 4, "train_total_number_recipient_token": 77 }, { "epoch": 0.40711462450592883, "grad_norm": 2.6639734516195612, "learning_rate": 6.726825272106538e-07, "loss": 0.1258, "step": 103 }, { "epoch": 0.40711462450592883, "step": 103, "train_accuracy_recipient_token": 0.9452054794520548, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_bird": 1.0, "train_accuracy_recipient_token_bird_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 13, "train_accuracy_recipient_token_make": 1.0, "train_accuracy_recipient_token_make_total": 4, "train_total_number_recipient_token": 73 }, { "epoch": 0.41106719367588934, "grad_norm": 6.934210744995126, "learning_rate": 6.666515864544208e-07, "loss": 0.2123, "step": 104 }, { "epoch": 0.41106719367588934, "step": 104, "train_accuracy_recipient_token": 0.9166666666666666, "train_accuracy_recipient_token_all": 0.9069767441860465, "train_accuracy_recipient_token_all_total": 43, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 0.8333333333333334, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 12, "train_total_number_recipient_token": 72 }, { "epoch": 0.4150197628458498, "grad_norm": 1.8263248807894037, "learning_rate": 6.605932444038228e-07, "loss": 0.2588, "step": 105 }, { "epoch": 0.4150197628458498, "step": 105, "train_accuracy_recipient_token": 0.918918918918919, "train_accuracy_recipient_token_all": 0.925, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_explain": 1.0, "train_accuracy_recipient_token_explain_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.9047619047619048, "train_accuracy_recipient_token_get_total": 21, "train_accuracy_recipient_token_p": 1.0, "train_accuracy_recipient_token_p_total": 3, "train_total_number_recipient_token": 74 }, { "epoch": 0.4189723320158103, "grad_norm": 1.6061172103912635, "learning_rate": 6.545084971874736e-07, "loss": 0.2185, "step": 106 }, { "epoch": 0.4189723320158103, "step": 106, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.975, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_find": 0.75, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 12, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 4, "train_accuracy_recipient_token_l": 1.0, "train_accuracy_recipient_token_l_total": 3, "train_total_number_recipient_token": 76 }, { "epoch": 0.42292490118577075, "grad_norm": 1.6625455290791247, "learning_rate": 6.483983452755952e-07, "loss": 0.1861, "step": 107 }, { "epoch": 0.42292490118577075, "step": 107, "train_accuracy_recipient_token": 0.8846153846153846, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_event": 0.6666666666666666, "train_accuracy_recipient_token_event_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 0.9090909090909091, "train_accuracy_recipient_token_get_total": 11, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 4, "train_total_number_recipient_token": 78 }, { "epoch": 0.4268774703557312, "grad_norm": 1.8525360380881017, "learning_rate": 6.422637933155162e-07, "loss": 0.1878, "step": 108 }, { "epoch": 0.4268774703557312, "step": 108, "train_accuracy_recipient_token": 0.8888888888888888, "train_accuracy_recipient_token_all": 0.8536585365853658, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_find": 0.9285714285714286, "train_accuracy_recipient_token_find_total": 14, "train_accuracy_recipient_token_get": 0.8888888888888888, "train_accuracy_recipient_token_get_total": 9, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 4, "train_total_number_recipient_token": 72 }, { "epoch": 0.4308300395256917, "grad_norm": 1.4111688417897226, "learning_rate": 6.361058499664855e-07, "loss": 0.1914, "step": 109 }, { "epoch": 0.4308300395256917, "step": 109, "train_accuracy_recipient_token": 0.9315068493150684, "train_accuracy_recipient_token_all": 0.9512195121951219, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_evaluate": 1.0, "train_accuracy_recipient_token_evaluate_total": 1, "train_accuracy_recipient_token_find": 0.8, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 0.875, "train_accuracy_recipient_token_get_total": 16, "train_total_number_recipient_token": 73 }, { "epoch": 0.43478260869565216, "grad_norm": 2.971458404255017, "learning_rate": 6.299255277338264e-07, "loss": 0.1953, "step": 110 }, { "epoch": 0.43478260869565216, "step": 110, "train_accuracy_recipient_token": 0.9166666666666666, "train_accuracy_recipient_token_all": 0.85, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 14, "train_accuracy_recipient_token_traffic": 1.0, "train_accuracy_recipient_token_traffic_total": 2, "train_total_number_recipient_token": 72 }, { "epoch": 0.43873517786561267, "grad_norm": 5.878940640472501, "learning_rate": 6.237238428024571e-07, "loss": 0.2769, "step": 111 }, { "epoch": 0.43873517786561267, "step": 111, "train_accuracy_recipient_token": 0.9178082191780822, "train_accuracy_recipient_token_all": 0.875, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 7, "train_accuracy_recipient_token_get": 0.9444444444444444, "train_accuracy_recipient_token_get_total": 18, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 3, "train_total_number_recipient_token": 73 }, { "epoch": 0.4426877470355731, "grad_norm": 1.311028563981913, "learning_rate": 6.175018148698076e-07, "loss": 0.1527, "step": 112 }, { "epoch": 0.4426877470355731, "step": 112, "train_accuracy_recipient_token": 0.9315068493150684, "train_accuracy_recipient_token_all": 0.925, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_discover": 0.6666666666666666, "train_accuracy_recipient_token_discover_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.9545454545454546, "train_accuracy_recipient_token_get_total": 22, "train_accuracy_recipient_token_suggest": 1.0, "train_accuracy_recipient_token_suggest_total": 2, "train_total_number_recipient_token": 73 }, { "epoch": 0.44664031620553357, "grad_norm": 1.7400988730477869, "learning_rate": 6.112604669781572e-07, "loss": 0.228, "step": 113 }, { "epoch": 0.44664031620553357, "step": 113, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.8333333333333334, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 16, "train_accuracy_recipient_token_provide": 1.0, "train_accuracy_recipient_token_provide_total": 1, "train_accuracy_recipient_token_stamp": 1.0, "train_accuracy_recipient_token_stamp_total": 3, "train_total_number_recipient_token": 70 }, { "epoch": 0.4505928853754941, "grad_norm": 2.3297546622652, "learning_rate": 6.050008253464246e-07, "loss": 0.1672, "step": 114 }, { "epoch": 0.4505928853754941, "step": 114, "train_accuracy_recipient_token": 0.9315068493150684, "train_accuracy_recipient_token_all": 0.9512195121951219, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 12, "train_accuracy_recipient_token_rec": 1.0, "train_accuracy_recipient_token_rec_total": 4, "train_total_number_recipient_token": 73 }, { "epoch": 0.45454545454545453, "grad_norm": 2.4037260219193346, "learning_rate": 5.987239192014335e-07, "loss": 0.1733, "step": 115 }, { "epoch": 0.45454545454545453, "step": 115, "train_accuracy_recipient_token": 0.9605263157894737, "train_accuracy_recipient_token_all": 0.9545454545454546, "train_accuracy_recipient_token_all_total": 44, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_generate": 0.8, "train_accuracy_recipient_token_generate_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_accuracy_recipient_token_multip": 1.0, "train_accuracy_recipient_token_multip_total": 4, "train_total_number_recipient_token": 76 }, { "epoch": 0.45849802371541504, "grad_norm": 6.009854171938679, "learning_rate": 5.924307806086843e-07, "loss": 0.1849, "step": 116 }, { "epoch": 0.45849802371541504, "step": 116, "train_accuracy_recipient_token": 0.8767123287671232, "train_accuracy_recipient_token_all": 0.8536585365853658, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 0.8461538461538461, "train_accuracy_recipient_token_get_total": 13, "train_accuracy_recipient_token_histor": 1.0, "train_accuracy_recipient_token_histor_total": 2, "train_total_number_recipient_token": 73 }, { "epoch": 0.4624505928853755, "grad_norm": 1.9748274232817422, "learning_rate": 5.861224443026595e-07, "loss": 0.1897, "step": 117 }, { "epoch": 0.4624505928853755, "step": 117, "train_accuracy_recipient_token": 0.958904109589041, "train_accuracy_recipient_token_all": 0.95, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 5, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 18, "train_total_number_recipient_token": 73 }, { "epoch": 0.466403162055336, "grad_norm": 1.780701736933667, "learning_rate": 5.797999475166896e-07, "loss": 0.2088, "step": 118 }, { "epoch": 0.466403162055336, "step": 118, "train_accuracy_recipient_token": 0.96, "train_accuracy_recipient_token_all": 0.9743589743589743, "train_accuracy_recipient_token_all_total": 39, "train_accuracy_recipient_token_find": 0.9166666666666666, "train_accuracy_recipient_token_find_total": 12, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 16, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 3, "train_accuracy_recipient_token_shop": 1.0, "train_accuracy_recipient_token_shop_total": 2, "train_total_number_recipient_token": 75 }, { "epoch": 0.47035573122529645, "grad_norm": 1.9760668589250143, "learning_rate": 5.73464329812409e-07, "loss": 0.1508, "step": 119 }, { "epoch": 0.47035573122529645, "step": 119, "train_accuracy_recipient_token": 0.9452054794520548, "train_accuracy_recipient_token_all": 0.925, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_estimate": 0.5, "train_accuracy_recipient_token_estimate_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 21, "train_total_number_recipient_token": 73 }, { "epoch": 0.4743083003952569, "grad_norm": 1.9282116516525682, "learning_rate": 5.671166329088277e-07, "loss": 0.139, "step": 120 }, { "epoch": 0.4743083003952569, "step": 120, "train_accuracy_recipient_token": 0.9027777777777778, "train_accuracy_recipient_token_all": 0.875, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 16, "train_accuracy_recipient_token_schedule": 0.5, "train_accuracy_recipient_token_schedule_total": 2, "train_total_number_recipient_token": 72 }, { "epoch": 0.4782608695652174, "grad_norm": 1.6545732035439291, "learning_rate": 5.607579005110502e-07, "loss": 0.1691, "step": 121 }, { "epoch": 0.4782608695652174, "step": 121, "train_accuracy_recipient_token": 0.9459459459459459, "train_accuracy_recipient_token_all": 0.9512195121951219, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_find": 0.8, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 18, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_total_number_recipient_token": 74 }, { "epoch": 0.48221343873517786, "grad_norm": 2.0522138550844873, "learning_rate": 5.543891781386655e-07, "loss": 0.1419, "step": 122 }, { "epoch": 0.48221343873517786, "step": 122, "train_accuracy_recipient_token": 0.9466666666666667, "train_accuracy_recipient_token_all": 0.95, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 13, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 3, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 5, "train_total_number_recipient_token": 75 }, { "epoch": 0.48616600790513836, "grad_norm": 1.493107058008911, "learning_rate": 5.480115129538409e-07, "loss": 0.1979, "step": 123 }, { "epoch": 0.48616600790513836, "step": 123, "train_accuracy_recipient_token": 0.971830985915493, "train_accuracy_recipient_token_all": 0.95, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 17, "train_accuracy_recipient_token_inspect": 1.0, "train_accuracy_recipient_token_inspect_total": 4, "train_accuracy_recipient_token_review": 1.0, "train_accuracy_recipient_token_review_total": 1, "train_total_number_recipient_token": 71 }, { "epoch": 0.4901185770750988, "grad_norm": 1.44050549323833, "learning_rate": 5.416259535891446e-07, "loss": 0.2266, "step": 124 }, { "epoch": 0.4901185770750988, "step": 124, "train_accuracy_recipient_token": 0.9605263157894737, "train_accuracy_recipient_token_all": 0.926829268292683, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 13, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_accuracy_recipient_token_st": 1.0, "train_accuracy_recipient_token_st_total": 3, "train_total_number_recipient_token": 76 }, { "epoch": 0.49407114624505927, "grad_norm": 21.5232411231967, "learning_rate": 5.352335499751269e-07, "loss": 0.2722, "step": 125 }, { "epoch": 0.49407114624505927, "step": 125, "train_accuracy_recipient_token": 0.9861111111111112, "train_accuracy_recipient_token_all": 0.975, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 7, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 18, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 72 }, { "epoch": 0.4980237154150198, "grad_norm": 1.842848864670893, "learning_rate": 5.288353531676873e-07, "loss": 0.2354, "step": 126 }, { "epoch": 0.4980237154150198, "step": 126, "train_accuracy_recipient_token": 0.9692307692307692, "train_accuracy_recipient_token_all": 0.9473684210526315, "train_accuracy_recipient_token_all_total": 38, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 19, "train_total_number_recipient_token": 65 }, { "epoch": 0.5019762845849802, "grad_norm": 2.4698981017742803, "learning_rate": 5.224324151752575e-07, "loss": 0.2212, "step": 127 }, { "epoch": 0.5019762845849802, "step": 127, "train_accuracy_recipient_token": 0.9041095890410958, "train_accuracy_recipient_token_all": 0.875, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 17, "train_accuracy_recipient_token_invest": 1.0, "train_accuracy_recipient_token_invest_total": 1, "train_accuracy_recipient_token_multiply": 0.75, "train_accuracy_recipient_token_multiply_total": 4, "train_total_number_recipient_token": 73 }, { "epoch": 0.5059288537549407, "grad_norm": 1.5250038897355298, "learning_rate": 5.160257887858277e-07, "loss": 0.1552, "step": 128 }, { "epoch": 0.5059288537549407, "step": 128, "train_accuracy_recipient_token": 0.9733333333333334, "train_accuracy_recipient_token_all": 0.9761904761904762, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 19, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 6, "train_total_number_recipient_token": 75 }, { "epoch": 0.5098814229249012, "grad_norm": 1.9104117318203697, "learning_rate": 5.096165273938435e-07, "loss": 0.2404, "step": 129 }, { "epoch": 0.5098814229249012, "step": 129, "train_accuracy_recipient_token": 0.9285714285714286, "train_accuracy_recipient_token_all": 0.9024390243902439, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 10, "train_accuracy_recipient_token_histor": 0.5, "train_accuracy_recipient_token_histor_total": 2, "train_accuracy_recipient_token_plant": 1.0, "train_accuracy_recipient_token_plant_total": 3, "train_total_number_recipient_token": 70 }, { "epoch": 0.5138339920948617, "grad_norm": 2.2221416286583002, "learning_rate": 5.032056848270056e-07, "loss": 0.1761, "step": 130 }, { "epoch": 0.5138339920948617, "step": 130, "train_accuracy_recipient_token": 0.9027777777777778, "train_accuracy_recipient_token_all": 0.8571428571428571, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_find": 0.8333333333333334, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 16, "train_accuracy_recipient_token_plan": 1.0, "train_accuracy_recipient_token_plan_total": 1, "train_total_number_recipient_token": 72 }, { "epoch": 0.5177865612648221, "grad_norm": 1.979869503575977, "learning_rate": 4.967943151729944e-07, "loss": 0.2083, "step": 131 }, { "epoch": 0.5177865612648221, "step": 131, "train_accuracy_recipient_token": 0.9861111111111112, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_ca": 1.0, "train_accuracy_recipient_token_ca_total": 4, "train_accuracy_recipient_token_decode": 1.0, "train_accuracy_recipient_token_decode_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.9444444444444444, "train_accuracy_recipient_token_get_total": 18, "train_total_number_recipient_token": 72 }, { "epoch": 0.5217391304347826, "grad_norm": 1.7887580979095283, "learning_rate": 4.903834726061564e-07, "loss": 0.3406, "step": 132 }, { "epoch": 0.5217391304347826, "step": 132, "train_accuracy_recipient_token": 0.9142857142857143, "train_accuracy_recipient_token_all": 0.8809523809523809, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 7, "train_accuracy_recipient_token_generate": 0.5, "train_accuracy_recipient_token_generate_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 11, "train_accuracy_recipient_token_phone": 1.0, "train_accuracy_recipient_token_phone_total": 2, "train_total_number_recipient_token": 70 }, { "epoch": 0.525691699604743, "grad_norm": 1.8995822547641221, "learning_rate": 4.839742112141724e-07, "loss": 0.2433, "step": 133 }, { "epoch": 0.525691699604743, "step": 133, "train_accuracy_recipient_token": 0.9154929577464789, "train_accuracy_recipient_token_all": 0.9024390243902439, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_customize": 0.75, "train_accuracy_recipient_token_customize_total": 4, "train_accuracy_recipient_token_d": 1.0, "train_accuracy_recipient_token_d_total": 2, "train_accuracy_recipient_token_get": 0.9333333333333333, "train_accuracy_recipient_token_get_total": 15, "train_total_number_recipient_token": 71 }, { "epoch": 0.5296442687747036, "grad_norm": 2.0596210153342875, "learning_rate": 4.775675848247427e-07, "loss": 0.2377, "step": 134 }, { "epoch": 0.5296442687747036, "step": 134, "train_accuracy_recipient_token": 0.9393939393939394, "train_accuracy_recipient_token_all": 0.9210526315789473, "train_accuracy_recipient_token_all_total": 38, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.9444444444444444, "train_accuracy_recipient_token_get_total": 18, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_total_number_recipient_token": 66 }, { "epoch": 0.5335968379446641, "grad_norm": 5.705329197338222, "learning_rate": 4.7116464683231285e-07, "loss": 0.5221, "step": 135 }, { "epoch": 0.5335968379446641, "step": 135, "train_accuracy_recipient_token": 0.9210526315789473, "train_accuracy_recipient_token_all": 0.9047619047619048, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_explain": 1.0, "train_accuracy_recipient_token_explain_total": 3, "train_accuracy_recipient_token_find": 0.75, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_fish": 1.0, "train_accuracy_recipient_token_fish_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 22, "train_total_number_recipient_token": 76 }, { "epoch": 0.5375494071146245, "grad_norm": 1.8164950441376346, "learning_rate": 4.6476645002487295e-07, "loss": 0.2125, "step": 136 }, { "epoch": 0.5375494071146245, "step": 136, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.9069767441860465, "train_accuracy_recipient_token_all_total": 43, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 2, "train_accuracy_recipient_token_get": 0.9444444444444444, "train_accuracy_recipient_token_get_total": 18, "train_accuracy_recipient_token_monitor": 1.0, "train_accuracy_recipient_token_monitor_total": 1, "train_accuracy_recipient_token_recogn": 0.0, "train_accuracy_recipient_token_recogn_total": 1, "train_total_number_recipient_token": 70 }, { "epoch": 0.541501976284585, "grad_norm": 1.645661606662186, "learning_rate": 4.5837404641085535e-07, "loss": 0.2501, "step": 137 }, { "epoch": 0.541501976284585, "step": 137, "train_accuracy_recipient_token": 0.9428571428571428, "train_accuracy_recipient_token_all": 0.9302325581395349, "train_accuracy_recipient_token_all_total": 43, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 8, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 70 }, { "epoch": 0.5454545454545454, "grad_norm": 1.545495236722973, "learning_rate": 4.519884870461591e-07, "loss": 0.1921, "step": 138 }, { "epoch": 0.5454545454545454, "step": 138, "train_accuracy_recipient_token": 0.9583333333333334, "train_accuracy_recipient_token_all": 0.9285714285714286, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 14, "train_accuracy_recipient_token_stage": 1.0, "train_accuracy_recipient_token_stage_total": 4, "train_total_number_recipient_token": 72 }, { "epoch": 0.549407114624506, "grad_norm": 1.445672974868626, "learning_rate": 4.4561082186133456e-07, "loss": 0.1907, "step": 139 }, { "epoch": 0.549407114624506, "step": 139, "train_accuracy_recipient_token": 0.9154929577464789, "train_accuracy_recipient_token_all": 0.95, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_find": 0.8, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 17, "train_accuracy_recipient_token_identify": 0.5, "train_accuracy_recipient_token_identify_total": 2, "train_total_number_recipient_token": 71 }, { "epoch": 0.5533596837944664, "grad_norm": 1.7215144726723506, "learning_rate": 4.392420994889498e-07, "loss": 0.2208, "step": 140 }, { "epoch": 0.5533596837944664, "step": 140, "train_accuracy_recipient_token": 0.9583333333333334, "train_accuracy_recipient_token_all": 0.9523809523809523, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_describe": 1.0, "train_accuracy_recipient_token_describe_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.9565217391304348, "train_accuracy_recipient_token_get_total": 23, "train_total_number_recipient_token": 72 }, { "epoch": 0.5573122529644269, "grad_norm": 1.4753468671374175, "learning_rate": 4.328833670911724e-07, "loss": 0.1615, "step": 141 }, { "epoch": 0.5573122529644269, "step": 141, "train_accuracy_recipient_token": 0.9473684210526315, "train_accuracy_recipient_token_all": 0.9302325581395349, "train_accuracy_recipient_token_all_total": 43, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_gesture": 1.0, "train_accuracy_recipient_token_gesture_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 15, "train_accuracy_recipient_token_product": 0.75, "train_accuracy_recipient_token_product_total": 4, "train_total_number_recipient_token": 76 }, { "epoch": 0.5612648221343873, "grad_norm": 1.8504808058857731, "learning_rate": 4.2653567018759103e-07, "loss": 0.2084, "step": 142 }, { "epoch": 0.5612648221343873, "step": 142, "train_accuracy_recipient_token": 0.9054054054054054, "train_accuracy_recipient_token_all": 0.9285714285714286, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.8888888888888888, "train_accuracy_recipient_token_get_total": 18, "train_accuracy_recipient_token_remove": 1.0, "train_accuracy_recipient_token_remove_total": 1, "train_total_number_recipient_token": 74 }, { "epoch": 0.5652173913043478, "grad_norm": 2.931367519601243, "learning_rate": 4.202000524833105e-07, "loss": 0.2336, "step": 143 }, { "epoch": 0.5652173913043478, "step": 143, "train_accuracy_recipient_token": 0.9577464788732394, "train_accuracy_recipient_token_all": 0.925, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_draw": 1.0, "train_accuracy_recipient_token_draw_total": 4, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 71 }, { "epoch": 0.5691699604743083, "grad_norm": 1.7028237172592313, "learning_rate": 4.1387755569734054e-07, "loss": 0.2115, "step": 144 }, { "epoch": 0.5691699604743083, "step": 144, "train_accuracy_recipient_token": 0.9428571428571428, "train_accuracy_recipient_token_all": 0.9285714285714286, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 5, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_g": 1.0, "train_accuracy_recipient_token_g_total": 3, "train_accuracy_recipient_token_get": 0.9090909090909091, "train_accuracy_recipient_token_get_total": 11, "train_total_number_recipient_token": 70 }, { "epoch": 0.5731225296442688, "grad_norm": 2.0478509803811167, "learning_rate": 4.0756921939131563e-07, "loss": 0.2019, "step": 145 }, { "epoch": 0.5731225296442688, "step": 145, "train_accuracy_recipient_token": 0.9701492537313433, "train_accuracy_recipient_token_all": 0.975, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_check": 0.6666666666666666, "train_accuracy_recipient_token_check_total": 3, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 12, "train_total_number_recipient_token": 67 }, { "epoch": 0.5770750988142292, "grad_norm": 1.7126152835105626, "learning_rate": 4.0127608079856644e-07, "loss": 0.2038, "step": 146 }, { "epoch": 0.5770750988142292, "step": 146, "train_accuracy_recipient_token": 0.9305555555555556, "train_accuracy_recipient_token_all": 0.9285714285714286, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 0.8, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 18, "train_accuracy_recipient_token_stat": 1.0, "train_accuracy_recipient_token_stat_total": 2, "train_total_number_recipient_token": 72 }, { "epoch": 0.5810276679841897, "grad_norm": 1.3379397195314802, "learning_rate": 3.949991746535753e-07, "loss": 0.1422, "step": 147 }, { "epoch": 0.5810276679841897, "step": 147, "train_accuracy_recipient_token": 0.9285714285714286, "train_accuracy_recipient_token_all": 0.9024390243902439, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 1, "train_accuracy_recipient_token_construction": 1.0, "train_accuracy_recipient_token_construction_total": 2, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 5, "train_accuracy_recipient_token_get": 0.95, "train_accuracy_recipient_token_get_total": 20, "train_total_number_recipient_token": 70 }, { "epoch": 0.5849802371541502, "grad_norm": 1.8321894920998039, "learning_rate": 3.8873953302184283e-07, "loss": 0.1446, "step": 148 }, { "epoch": 0.5849802371541502, "step": 148, "train_accuracy_recipient_token": 0.9324324324324325, "train_accuracy_recipient_token_all": 0.9512195121951219, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 3, "train_accuracy_recipient_token_find": 0.8571428571428571, "train_accuracy_recipient_token_find_total": 7, "train_accuracy_recipient_token_get": 0.9375, "train_accuracy_recipient_token_get_total": 16, "train_total_number_recipient_token": 74 }, { "epoch": 0.5889328063241107, "grad_norm": 2.1305085757946207, "learning_rate": 3.824981851301924e-07, "loss": 0.1342, "step": 149 }, { "epoch": 0.5889328063241107, "step": 149, "train_accuracy_recipient_token": 0.9736842105263158, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_compile": 1.0, "train_accuracy_recipient_token_compile_total": 1, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 23, "train_total_number_recipient_token": 76 }, { "epoch": 0.5928853754940712, "grad_norm": 1.2296405729963737, "learning_rate": 3.762761571975429e-07, "loss": 0.1259, "step": 150 }, { "epoch": 0.5928853754940712, "step": 150, "train_accuracy_recipient_token": 0.8873239436619719, "train_accuracy_recipient_token_air": 1.0, "train_accuracy_recipient_token_air_total": 1, "train_accuracy_recipient_token_all": 0.875, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.8, "train_accuracy_recipient_token_get_total": 15, "train_total_number_recipient_token": 71 }, { "epoch": 0.5968379446640316, "grad_norm": 1.545422961044898, "learning_rate": 3.700744722661736e-07, "loss": 0.2172, "step": 151 }, { "epoch": 0.5968379446640316, "step": 151, "train_accuracy_recipient_token": 0.958904109589041, "train_accuracy_recipient_token_all": 0.975, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_fire": 1.0, "train_accuracy_recipient_token_fire_total": 4, "train_accuracy_recipient_token_get": 0.9285714285714286, "train_accuracy_recipient_token_get_total": 14, "train_accuracy_recipient_token_plant": 1.0, "train_accuracy_recipient_token_plant_total": 3, "train_total_number_recipient_token": 73 }, { "epoch": 0.6007905138339921, "grad_norm": 1.3564960203861138, "learning_rate": 3.638941500335144e-07, "loss": 0.1649, "step": 152 }, { "epoch": 0.6007905138339921, "step": 152, "train_accuracy_recipient_token": 0.918918918918919, "train_accuracy_recipient_token_all": 0.8888888888888888, "train_accuracy_recipient_token_all_total": 45, "train_accuracy_recipient_token_bridge": 1.0, "train_accuracy_recipient_token_bridge_total": 3, "train_accuracy_recipient_token_d": 1.0, "train_accuracy_recipient_token_d_total": 3, "train_accuracy_recipient_token_find": 0.75, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 16, "train_total_number_recipient_token": 74 }, { "epoch": 0.6047430830039525, "grad_norm": 1.7435124580719308, "learning_rate": 3.577362066844838e-07, "loss": 0.2873, "step": 153 }, { "epoch": 0.6047430830039525, "step": 153, "train_accuracy_recipient_token": 0.9295774647887324, "train_accuracy_recipient_token_all": 0.975609756097561, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 0.9166666666666666, "train_accuracy_recipient_token_get_total": 12, "train_accuracy_recipient_token_review": 0.5, "train_accuracy_recipient_token_review_total": 2, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 3, "train_total_number_recipient_token": 71 }, { "epoch": 0.6086956521739131, "grad_norm": 1.3569175167442251, "learning_rate": 3.5160165472440467e-07, "loss": 0.1383, "step": 154 }, { "epoch": 0.6086956521739131, "step": 154, "train_accuracy_recipient_token": 0.9605263157894737, "train_accuracy_recipient_token_all": 0.9777777777777777, "train_accuracy_recipient_token_all_total": 45, "train_accuracy_recipient_token_convert": 1.0, "train_accuracy_recipient_token_convert_total": 2, "train_accuracy_recipient_token_find": 0.8, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 19, "train_accuracy_recipient_token_parse": 0.5, "train_accuracy_recipient_token_parse_total": 2, "train_total_number_recipient_token": 76 }, { "epoch": 0.6126482213438735, "grad_norm": 2.049376727932652, "learning_rate": 3.454915028125263e-07, "loss": 0.1745, "step": 155 }, { "epoch": 0.6126482213438735, "step": 155, "train_accuracy_recipient_token": 0.9428571428571428, "train_accuracy_recipient_token_all": 0.9545454545454546, "train_accuracy_recipient_token_all_total": 44, "train_accuracy_recipient_token_average": 1.0, "train_accuracy_recipient_token_average_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 0.9, "train_accuracy_recipient_token_get_total": 10, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 2, "train_total_number_recipient_token": 70 }, { "epoch": 0.616600790513834, "grad_norm": 2.0096000106386316, "learning_rate": 3.3940675559617723e-07, "loss": 0.1776, "step": 156 }, { "epoch": 0.616600790513834, "step": 156, "train_accuracy_recipient_token": 0.9041095890410958, "train_accuracy_recipient_token_all": 0.8888888888888888, "train_accuracy_recipient_token_all_total": 45, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 2, "train_accuracy_recipient_token_get": 0.9444444444444444, "train_accuracy_recipient_token_get_total": 18, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_accuracy_recipient_token_p": 1.0, "train_accuracy_recipient_token_p_total": 3, "train_total_number_recipient_token": 73 }, { "epoch": 0.6205533596837944, "grad_norm": 2.4791175345887044, "learning_rate": 3.333484135455792e-07, "loss": 0.1972, "step": 157 }, { "epoch": 0.6205533596837944, "step": 157, "train_accuracy_recipient_token": 0.972972972972973, "train_accuracy_recipient_token_all": 0.9512195121951219, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 24, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 2, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 3, "train_total_number_recipient_token": 74 }, { "epoch": 0.6245059288537549, "grad_norm": 1.4862240167115932, "learning_rate": 3.2731747278934623e-07, "loss": 0.1859, "step": 158 }, { "epoch": 0.6245059288537549, "step": 158, "train_accuracy_recipient_token": 0.9295774647887324, "train_accuracy_recipient_token_all": 0.9024390243902439, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 23, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_accuracy_recipient_token_interpret": 1.0, "train_accuracy_recipient_token_interpret_total": 1, "train_total_number_recipient_token": 71 }, { "epoch": 0.6284584980237155, "grad_norm": 1.9778836023162545, "learning_rate": 3.2131492495069965e-07, "loss": 0.2368, "step": 159 }, { "epoch": 0.6284584980237155, "step": 159, "train_accuracy_recipient_token": 0.9452054794520548, "train_accuracy_recipient_token_all": 0.975609756097561, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_find": 0.9090909090909091, "train_accuracy_recipient_token_find_total": 11, "train_accuracy_recipient_token_get": 0.9333333333333333, "train_accuracy_recipient_token_get_total": 15, "train_accuracy_recipient_token_sum": 1.0, "train_accuracy_recipient_token_sum_total": 2, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 2, "train_total_number_recipient_token": 73 }, { "epoch": 0.6324110671936759, "grad_norm": 1.277403570518443, "learning_rate": 3.153417569844219e-07, "loss": 0.1704, "step": 160 }, { "epoch": 0.6324110671936759, "step": 160, "train_accuracy_recipient_token": 0.9154929577464789, "train_accuracy_recipient_token_all": 0.9024390243902439, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_boat": 1.0, "train_accuracy_recipient_token_boat_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 14, "train_accuracy_recipient_token_suggest": 1.0, "train_accuracy_recipient_token_suggest_total": 2, "train_total_number_recipient_token": 71 }, { "epoch": 0.6363636363636364, "grad_norm": 1.7595528358576455, "learning_rate": 3.0939895101457914e-07, "loss": 0.1485, "step": 161 }, { "epoch": 0.6363636363636364, "step": 161, "train_accuracy_recipient_token": 0.9054054054054054, "train_accuracy_recipient_token_all": 0.9512195121951219, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_find": 0.6, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_hare": 0.6666666666666666, "train_accuracy_recipient_token_hare_total": 3, "train_accuracy_recipient_token_ind": 1.0, "train_accuracy_recipient_token_ind_total": 4, "train_total_number_recipient_token": 74 }, { "epoch": 0.6403162055335968, "grad_norm": 2.3477414539381876, "learning_rate": 3.034874841730382e-07, "loss": 0.1873, "step": 162 }, { "epoch": 0.6403162055335968, "step": 162, "train_accuracy_recipient_token": 0.9358974358974359, "train_accuracy_recipient_token_all": 0.9069767441860465, "train_accuracy_recipient_token_all_total": 43, "train_accuracy_recipient_token_architecture": 1.0, "train_accuracy_recipient_token_architecture_total": 1, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.9655172413793104, "train_accuracy_recipient_token_get_total": 29, "train_total_number_recipient_token": 78 }, { "epoch": 0.6442687747035574, "grad_norm": 1.3483023159120098, "learning_rate": 2.976083284388031e-07, "loss": 0.2091, "step": 163 }, { "epoch": 0.6442687747035574, "step": 163, "train_accuracy_recipient_token": 0.9295774647887324, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_find": 0.8571428571428571, "train_accuracy_recipient_token_find_total": 7, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 16, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 1, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 4, "train_total_number_recipient_token": 71 }, { "epoch": 0.6482213438735178, "grad_norm": 2.269124662711222, "learning_rate": 2.917624504782006e-07, "loss": 0.1085, "step": 164 }, { "epoch": 0.6482213438735178, "step": 164, "train_accuracy_recipient_token": 0.9594594594594594, "train_accuracy_recipient_token_all": 0.975, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 4, "train_accuracy_recipient_token_explain": 1.0, "train_accuracy_recipient_token_explain_total": 5, "train_accuracy_recipient_token_find": 0.8333333333333334, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 0.9285714285714286, "train_accuracy_recipient_token_get_total": 14, "train_total_number_recipient_token": 74 }, { "epoch": 0.6521739130434783, "grad_norm": 1.3990047306065339, "learning_rate": 2.8595081148593737e-07, "loss": 0.1484, "step": 165 }, { "epoch": 0.6521739130434783, "step": 165, "train_accuracy_recipient_token": 0.9577464788732394, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 8, "train_accuracy_recipient_token_get": 0.8947368421052632, "train_accuracy_recipient_token_get_total": 19, "train_accuracy_recipient_token_inspect": 0.0, "train_accuracy_recipient_token_inspect_total": 1, "train_total_number_recipient_token": 71 }, { "epoch": 0.6561264822134387, "grad_norm": 1.7322399985781272, "learning_rate": 2.80174367027059e-07, "loss": 0.1917, "step": 166 }, { "epoch": 0.6561264822134387, "step": 166, "train_accuracy_recipient_token": 0.9722222222222222, "train_accuracy_recipient_token_all": 0.975609756097561, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.9545454545454546, "train_accuracy_recipient_token_get_total": 22, "train_accuracy_recipient_token_load": 1.0, "train_accuracy_recipient_token_load_total": 1, "train_total_number_recipient_token": 72 }, { "epoch": 0.6600790513833992, "grad_norm": 1.9490777157692185, "learning_rate": 2.7443406687983264e-07, "loss": 0.1517, "step": 167 }, { "epoch": 0.6600790513833992, "step": 167, "train_accuracy_recipient_token": 0.9285714285714286, "train_accuracy_recipient_token_all": 0.9302325581395349, "train_accuracy_recipient_token_all_total": 43, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 11, "train_accuracy_recipient_token_park": 1.0, "train_accuracy_recipient_token_park_total": 2, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 2, "train_total_number_recipient_token": 70 }, { "epoch": 0.6640316205533597, "grad_norm": 2.281852433597984, "learning_rate": 2.687308548795825e-07, "loss": 0.2231, "step": 168 }, { "epoch": 0.6640316205533597, "step": 168, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9210526315789473, "train_accuracy_recipient_token_all_total": 38, "train_accuracy_recipient_token_find": 0.9166666666666666, "train_accuracy_recipient_token_find_total": 12, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 21, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_total_number_recipient_token": 72 }, { "epoch": 0.6640316205533597, "step": 168, "train_accuracy_recipient_token": 0.9436619718309859, "train_accuracy_recipient_token_all": 0.9285714285714286, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_find": 0.8, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 15, "train_accuracy_recipient_token_histor": 1.0, "train_accuracy_recipient_token_histor_total": 3, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 3, "train_total_number_recipient_token": 71 }, { "epoch": 0.6640316205533597, "step": 168, "train_accuracy_recipient_token": 0.9166666666666666, "train_accuracy_recipient_token_all": 0.8695652173913043, "train_accuracy_recipient_token_all_total": 46, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 8, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_accuracy_recipient_token_interpret": 1.0, "train_accuracy_recipient_token_interpret_total": 2, "train_total_number_recipient_token": 72 }, { "epoch": 0.6640316205533597, "step": 168, "train_accuracy_recipient_token": 0.9736842105263158, "train_accuracy_recipient_token_all": 0.975, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 14, "train_accuracy_recipient_token_tour": 1.0, "train_accuracy_recipient_token_tour_total": 4, "train_total_number_recipient_token": 76 }, { "epoch": 0.6640316205533597, "step": 168, "train_accuracy_recipient_token": 0.9722222222222222, "train_accuracy_recipient_token_all": 0.9512195121951219, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_artifact": 1.0, "train_accuracy_recipient_token_artifact_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 20, "train_accuracy_recipient_token_select": 1.0, "train_accuracy_recipient_token_select_total": 2, "train_total_number_recipient_token": 72 }, { "epoch": 0.6640316205533597, "step": 168, "train_accuracy_recipient_token": 0.9571428571428572, "train_accuracy_recipient_token_all": 0.95, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 16, "train_accuracy_recipient_token_gr": 1.0, "train_accuracy_recipient_token_gr_total": 3, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_total_number_recipient_token": 70 }, { "epoch": 0.6640316205533597, "step": 168, "train_accuracy_recipient_token": 0.8904109589041096, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 2, "train_accuracy_recipient_token_customize": 0.75, "train_accuracy_recipient_token_customize_total": 4, "train_accuracy_recipient_token_get": 0.8695652173913043, "train_accuracy_recipient_token_get_total": 23, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 73 }, { "epoch": 0.6640316205533597, "step": 168, "train_accuracy_recipient_token": 0.9285714285714286, "train_accuracy_recipient_token_all": 0.8780487804878049, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_eng": 1.0, "train_accuracy_recipient_token_eng_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 11, "train_accuracy_recipient_token_tem": 1.0, "train_accuracy_recipient_token_tem_total": 3, "train_total_number_recipient_token": 70 }, { "epoch": 0.6640316205533597, "step": 168, "train_accuracy_recipient_token": 0.9420289855072463, "train_accuracy_recipient_token_all": 0.925, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_bird": 1.0, "train_accuracy_recipient_token_bird_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_total_number_recipient_token": 69 }, { "epoch": 0.6640316205533597, "step": 168, "train_accuracy_recipient_token": 0.9583333333333334, "train_accuracy_recipient_token_all": 0.925, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 19, "train_accuracy_recipient_token_park": 1.0, "train_accuracy_recipient_token_park_total": 2, "train_total_number_recipient_token": 72 }, { "epoch": 0.6640316205533597, "step": 168, "train_accuracy_recipient_token": 0.9436619718309859, "train_accuracy_recipient_token_all": 0.975, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 3, "train_accuracy_recipient_token_fruit": 1.0, "train_accuracy_recipient_token_fruit_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 15, "train_accuracy_recipient_token_object": 0.6666666666666666, "train_accuracy_recipient_token_object_total": 3, "train_total_number_recipient_token": 71 }, { "epoch": 0.6640316205533597, "step": 168, "train_accuracy_recipient_token": 0.9130434782608695, "train_accuracy_recipient_token_all": 0.875, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 0.9285714285714286, "train_accuracy_recipient_token_get_total": 14, "train_accuracy_recipient_token_organ": 1.0, "train_accuracy_recipient_token_organ_total": 1, "train_accuracy_recipient_token_stage": 1.0, "train_accuracy_recipient_token_stage_total": 4, "train_total_number_recipient_token": 69 }, { "epoch": 0.6640316205533597, "step": 168, "train_accuracy_recipient_token": 0.9555555555555556, "train_accuracy_recipient_token_all": 0.96, "train_accuracy_recipient_token_all_total": 25, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 5, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 10, "train_accuracy_recipient_token_sun": 1.0, "train_accuracy_recipient_token_sun_total": 1, "train_total_number_recipient_token": 45 }, { "epoch": 0.6640316205533597, "eval_accuracy_recipient_token": 0.9412416851441242, "eval_accuracy_recipient_token_all": 0.9239766081871345, "eval_accuracy_recipient_token_all_total": 513, "eval_accuracy_recipient_token_calculate": 1.0, "eval_accuracy_recipient_token_calculate_total": 13, "eval_accuracy_recipient_token_find": 0.9649122807017544, "eval_accuracy_recipient_token_find_total": 57, "eval_accuracy_recipient_token_get": 0.9795918367346939, "eval_accuracy_recipient_token_get_total": 196, "eval_accuracy_recipient_token_identify": 1.0, "eval_accuracy_recipient_token_identify_total": 6, "eval_loss": 0.17430317401885986, "eval_perplexity": 1.0313331954401281, "eval_runtime": 97.9718, "eval_samples_per_second": 1.031, "eval_steps_per_second": 0.133, "eval_total_number_recipient_token": 902, "step": 168 }, { "epoch": 0.6640316205533597, "step": 168, "train_accuracy_recipient_token": 0.92, "train_accuracy_recipient_token_all": 0.9523809523809523, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_g": 1.0, "train_accuracy_recipient_token_g_total": 2, "train_accuracy_recipient_token_get": 0.8333333333333334, "train_accuracy_recipient_token_get_total": 12, "train_total_number_recipient_token": 75 }, { "epoch": 0.6679841897233202, "grad_norm": 1.3241581185562914, "learning_rate": 2.6306566876350067e-07, "loss": 0.1672, "step": 169 }, { "epoch": 0.6679841897233202, "step": 169, "train_accuracy_recipient_token": 0.9324324324324325, "train_accuracy_recipient_token_all": 0.9512195121951219, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_christ": 1.0, "train_accuracy_recipient_token_christ_total": 1, "train_accuracy_recipient_token_explain": 0.0, "train_accuracy_recipient_token_explain_total": 1, "train_accuracy_recipient_token_find": 0.9090909090909091, "train_accuracy_recipient_token_find_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 18, "train_total_number_recipient_token": 74 }, { "epoch": 0.6719367588932806, "grad_norm": 1.530195223252461, "learning_rate": 2.5743944001646387e-07, "loss": 0.2033, "step": 170 }, { "epoch": 0.6719367588932806, "step": 170, "train_accuracy_recipient_token": 0.9558823529411765, "train_accuracy_recipient_token_all": 0.9302325581395349, "train_accuracy_recipient_token_all_total": 43, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 4, "train_accuracy_recipient_token_draw": 1.0, "train_accuracy_recipient_token_draw_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 11, "train_total_number_recipient_token": 68 }, { "epoch": 0.6758893280632411, "grad_norm": 1.7520275631743933, "learning_rate": 2.518530937178751e-07, "loss": 0.2536, "step": 171 }, { "epoch": 0.6758893280632411, "step": 171, "train_accuracy_recipient_token": 0.835820895522388, "train_accuracy_recipient_token_all": 0.8636363636363636, "train_accuracy_recipient_token_all_total": 44, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_generate": 0.8, "train_accuracy_recipient_token_generate_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 67 }, { "epoch": 0.6798418972332015, "grad_norm": 1.9500671864208103, "learning_rate": 2.4630754838955896e-07, "loss": 0.322, "step": 172 }, { "epoch": 0.6798418972332015, "step": 172, "train_accuracy_recipient_token": 0.9315068493150684, "train_accuracy_recipient_token_all": 0.9285714285714286, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.9259259259259259, "train_accuracy_recipient_token_get_total": 27, "train_total_number_recipient_token": 73 }, { "epoch": 0.6837944664031621, "grad_norm": 1.4313198422378177, "learning_rate": 2.4080371584473745e-07, "loss": 0.1518, "step": 173 }, { "epoch": 0.6837944664031621, "step": 173, "train_accuracy_recipient_token": 0.9571428571428572, "train_accuracy_recipient_token_all": 0.9347826086956522, "train_accuracy_recipient_token_all_total": 46, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 14, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 3, "train_accuracy_recipient_token_wine": 1.0, "train_accuracy_recipient_token_wine_total": 2, "train_total_number_recipient_token": 70 }, { "epoch": 0.6877470355731226, "grad_norm": 1.7822737846911771, "learning_rate": 2.3534250103810627e-07, "loss": 0.257, "step": 174 }, { "epoch": 0.6877470355731226, "step": 174, "train_accuracy_recipient_token": 0.9253731343283582, "train_accuracy_recipient_token_all": 0.9047619047619048, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_convert": 0.75, "train_accuracy_recipient_token_convert_total": 4, "train_accuracy_recipient_token_draw": 1.0, "train_accuracy_recipient_token_draw_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_total_number_recipient_token": 67 }, { "epoch": 0.691699604743083, "grad_norm": 2.6402860344919272, "learning_rate": 2.2992480191704e-07, "loss": 0.1903, "step": 175 }, { "epoch": 0.691699604743083, "step": 175, "train_accuracy_recipient_token": 0.9305555555555556, "train_accuracy_recipient_token_all": 0.9130434782608695, "train_accuracy_recipient_token_all_total": 46, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_find": 0.9090909090909091, "train_accuracy_recipient_token_find_total": 11, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 11, "train_accuracy_recipient_token_optimize": 1.0, "train_accuracy_recipient_token_optimize_total": 2, "train_total_number_recipient_token": 72 }, { "epoch": 0.6956521739130435, "grad_norm": 1.817064457251319, "learning_rate": 2.2455150927394878e-07, "loss": 0.1544, "step": 176 }, { "epoch": 0.6956521739130435, "step": 176, "train_accuracy_recipient_token": 0.9324324324324325, "train_accuracy_recipient_token_all": 0.9024390243902439, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_detect": 1.0, "train_accuracy_recipient_token_detect_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 7, "train_accuracy_recipient_token_get": 0.9333333333333333, "train_accuracy_recipient_token_get_total": 15, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_total_number_recipient_token": 74 }, { "epoch": 0.6996047430830039, "grad_norm": 1.248905704249522, "learning_rate": 2.192235065998126e-07, "loss": 0.1609, "step": 177 }, { "epoch": 0.6996047430830039, "step": 177, "train_accuracy_recipient_token": 0.9253731343283582, "train_accuracy_recipient_token_all": 0.9024390243902439, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.9285714285714286, "train_accuracy_recipient_token_get_total": 14, "train_accuracy_recipient_token_optimize": 1.0, "train_accuracy_recipient_token_optimize_total": 1, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 1, "train_total_number_recipient_token": 67 }, { "epoch": 0.7035573122529645, "grad_norm": 2.574896993706397, "learning_rate": 2.1394166993891526e-07, "loss": 0.206, "step": 178 }, { "epoch": 0.7035573122529645, "step": 178, "train_accuracy_recipient_token": 0.9014084507042254, "train_accuracy_recipient_token_all": 0.9285714285714286, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_generate": 1.0, "train_accuracy_recipient_token_generate_total": 5, "train_accuracy_recipient_token_get": 0.7142857142857143, "train_accuracy_recipient_token_get_total": 14, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_total_number_recipient_token": 71 }, { "epoch": 0.7075098814229249, "grad_norm": 1.4313755596673818, "learning_rate": 2.0870686774480196e-07, "loss": 0.2645, "step": 179 }, { "epoch": 0.7075098814229249, "step": 179, "train_accuracy_recipient_token": 0.9852941176470589, "train_accuracy_recipient_token_all": 0.975, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 19, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 4, "train_accuracy_recipient_token_merge": 1.0, "train_accuracy_recipient_token_merge_total": 1, "train_total_number_recipient_token": 68 }, { "epoch": 0.7114624505928854, "grad_norm": 1.6532367430564563, "learning_rate": 2.0351996073748713e-07, "loss": 0.219, "step": 180 }, { "epoch": 0.7114624505928854, "step": 180, "train_accuracy_recipient_token": 0.9466666666666667, "train_accuracy_recipient_token_all": 0.975609756097561, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_find": 0.875, "train_accuracy_recipient_token_find_total": 8, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 14, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 2, "train_accuracy_recipient_token_vehicle": 1.0, "train_accuracy_recipient_token_vehicle_total": 3, "train_total_number_recipient_token": 75 }, { "epoch": 0.7154150197628458, "grad_norm": 1.35129359221742, "learning_rate": 1.9838180176193176e-07, "loss": 0.1513, "step": 181 }, { "epoch": 0.7154150197628458, "step": 181, "train_accuracy_recipient_token": 0.9452054794520548, "train_accuracy_recipient_token_all": 0.975, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_building": 1.0, "train_accuracy_recipient_token_building_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.9375, "train_accuracy_recipient_token_get_total": 16, "train_accuracy_recipient_token_object": 0.6666666666666666, "train_accuracy_recipient_token_object_total": 3, "train_total_number_recipient_token": 73 }, { "epoch": 0.7193675889328063, "grad_norm": 2.051560956688538, "learning_rate": 1.932932356478168e-07, "loss": 0.2662, "step": 182 }, { "epoch": 0.7193675889328063, "step": 182, "train_accuracy_recipient_token": 0.9861111111111112, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_compile": 1.0, "train_accuracy_recipient_token_compile_total": 1, "train_accuracy_recipient_token_exp": 1.0, "train_accuracy_recipient_token_exp_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 0.9473684210526315, "train_accuracy_recipient_token_get_total": 19, "train_total_number_recipient_token": 72 }, { "epoch": 0.7233201581027668, "grad_norm": 1.8134216070193625, "learning_rate": 1.8825509907063326e-07, "loss": 0.1656, "step": 183 }, { "epoch": 0.7233201581027668, "step": 183, "train_accuracy_recipient_token": 0.9305555555555556, "train_accuracy_recipient_token_all": 0.8780487804878049, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 3, "train_accuracy_recipient_token_design": 1.0, "train_accuracy_recipient_token_design_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 8, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 4, "train_total_number_recipient_token": 72 }, { "epoch": 0.7272727272727273, "grad_norm": 1.288526990226647, "learning_rate": 1.832682204141152e-07, "loss": 0.1944, "step": 184 }, { "epoch": 0.7272727272727273, "step": 184, "train_accuracy_recipient_token": 0.9428571428571428, "train_accuracy_recipient_token_all": 0.9318181818181818, "train_accuracy_recipient_token_all_total": 44, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 2, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 15, "train_total_number_recipient_token": 70 }, { "epoch": 0.7312252964426877, "grad_norm": 1.4837016134238015, "learning_rate": 1.783334196340331e-07, "loss": 0.1801, "step": 185 }, { "epoch": 0.7312252964426877, "step": 185, "train_accuracy_recipient_token": 0.961038961038961, "train_accuracy_recipient_token_all": 0.9555555555555556, "train_accuracy_recipient_token_all_total": 45, "train_accuracy_recipient_token_buy": 1.0, "train_accuracy_recipient_token_buy_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 13, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 11, "train_accuracy_recipient_token_m": 1.0, "train_accuracy_recipient_token_m_total": 4, "train_total_number_recipient_token": 77 }, { "epoch": 0.7351778656126482, "grad_norm": 1.3239039944418263, "learning_rate": 1.7345150812337562e-07, "loss": 0.1368, "step": 186 }, { "epoch": 0.7351778656126482, "step": 186, "train_accuracy_recipient_token": 0.9178082191780822, "train_accuracy_recipient_token_all": 0.9565217391304348, "train_accuracy_recipient_token_all_total": 46, "train_accuracy_recipient_token_construction": 1.0, "train_accuracy_recipient_token_construction_total": 2, "train_accuracy_recipient_token_find": 0.8571428571428571, "train_accuracy_recipient_token_find_total": 7, "train_accuracy_recipient_token_get": 0.8571428571428571, "train_accuracy_recipient_token_get_total": 14, "train_accuracy_recipient_token_translate": 0.0, "train_accuracy_recipient_token_translate_total": 1, "train_total_number_recipient_token": 73 }, { "epoch": 0.7391304347826086, "grad_norm": 1.6330843612455617, "learning_rate": 1.6862328857893855e-07, "loss": 0.2325, "step": 187 }, { "epoch": 0.7391304347826086, "step": 187, "train_accuracy_recipient_token": 0.890625, "train_accuracy_recipient_token_all": 0.9714285714285714, "train_accuracy_recipient_token_all_total": 35, "train_accuracy_recipient_token_animal": 1.0, "train_accuracy_recipient_token_animal_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.7727272727272727, "train_accuracy_recipient_token_get_total": 22, "train_accuracy_recipient_token_locate": 0.0, "train_accuracy_recipient_token_locate_total": 1, "train_total_number_recipient_token": 64 }, { "epoch": 0.7430830039525692, "grad_norm": 6.489452901624477, "learning_rate": 1.6384955486934154e-07, "loss": 0.3315, "step": 188 }, { "epoch": 0.7430830039525692, "step": 188, "train_accuracy_recipient_token": 0.971830985915493, "train_accuracy_recipient_token_all": 0.975609756097561, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 1, "train_accuracy_recipient_token_compose": 1.0, "train_accuracy_recipient_token_compose_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.9583333333333334, "train_accuracy_recipient_token_get_total": 24, "train_total_number_recipient_token": 71 }, { "epoch": 0.7470355731225297, "grad_norm": 1.9110123601740296, "learning_rate": 1.591310919045003e-07, "loss": 0.2183, "step": 189 }, { "epoch": 0.7470355731225297, "step": 189, "train_accuracy_recipient_token": 0.92, "train_accuracy_recipient_token_all": 0.9047619047619048, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_evaluate": 0.75, "train_accuracy_recipient_token_evaluate_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 9, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 13, "train_accuracy_recipient_token_suggest": 1.0, "train_accuracy_recipient_token_suggest_total": 2, "train_total_number_recipient_token": 75 }, { "epoch": 0.7509881422924901, "grad_norm": 1.4520620744956039, "learning_rate": 1.5446867550656767e-07, "loss": 0.1594, "step": 190 }, { "epoch": 0.7509881422924901, "step": 190, "train_accuracy_recipient_token": 0.8904109589041096, "train_accuracy_recipient_token_all": 0.9069767441860465, "train_accuracy_recipient_token_all_total": 43, "train_accuracy_recipient_token_event": 0.75, "train_accuracy_recipient_token_event_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 4, "train_accuracy_recipient_token_stamp": 0.75, "train_accuracy_recipient_token_stamp_total": 4, "train_total_number_recipient_token": 73 }, { "epoch": 0.7549407114624506, "grad_norm": 4.169150616614093, "learning_rate": 1.4986307228237267e-07, "loss": 0.2188, "step": 191 }, { "epoch": 0.7549407114624506, "step": 191, "train_accuracy_recipient_token": 0.9861111111111112, "train_accuracy_recipient_token_all": 0.975609756097561, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 7, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 11, "train_total_number_recipient_token": 72 }, { "epoch": 0.758893280632411, "grad_norm": 1.9971252284625571, "learning_rate": 1.4531503949737106e-07, "loss": 0.1583, "step": 192 }, { "epoch": 0.758893280632411, "step": 192, "train_accuracy_recipient_token": 0.9733333333333334, "train_accuracy_recipient_token_all": 0.975, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 7, "train_accuracy_recipient_token_get": 0.9523809523809523, "train_accuracy_recipient_token_get_total": 21, "train_accuracy_recipient_token_rent": 1.0, "train_accuracy_recipient_token_rent_total": 2, "train_total_number_recipient_token": 75 }, { "epoch": 0.7628458498023716, "grad_norm": 1.8316567369607857, "learning_rate": 1.4082532495113624e-07, "loss": 0.1519, "step": 193 }, { "epoch": 0.7628458498023716, "step": 193, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9512195121951219, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 2, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_find": 0.9, "train_accuracy_recipient_token_find_total": 10, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 10, "train_total_number_recipient_token": 72 }, { "epoch": 0.766798418972332, "grad_norm": 1.847634442384187, "learning_rate": 1.3639466685440132e-07, "loss": 0.1487, "step": 194 }, { "epoch": 0.766798418972332, "step": 194, "train_accuracy_recipient_token": 0.9333333333333333, "train_accuracy_recipient_token_all": 0.926829268292683, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_event": 0.6666666666666666, "train_accuracy_recipient_token_event_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 19, "train_accuracy_recipient_token_suggest": 1.0, "train_accuracy_recipient_token_suggest_total": 2, "train_total_number_recipient_token": 75 }, { "epoch": 0.7707509881422925, "grad_norm": 1.2853544598349524, "learning_rate": 1.320237937076825e-07, "loss": 0.1125, "step": 195 }, { "epoch": 0.7707509881422925, "step": 195, "train_accuracy_recipient_token": 0.9726027397260274, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 39, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 0.8888888888888888, "train_accuracy_recipient_token_find_total": 9, "train_accuracy_recipient_token_get": 0.9473684210526315, "train_accuracy_recipient_token_get_total": 19, "train_accuracy_recipient_token_restore": 1.0, "train_accuracy_recipient_token_restore_total": 1, "train_total_number_recipient_token": 73 }, { "epoch": 0.7747035573122529, "grad_norm": 1.3929449724445488, "learning_rate": 1.2771342418149656e-07, "loss": 0.1971, "step": 196 }, { "epoch": 0.7747035573122529, "step": 196, "train_accuracy_recipient_token": 0.9315068493150684, "train_accuracy_recipient_token_all": 0.926829268292683, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_f": 0.75, "train_accuracy_recipient_token_f_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 14, "train_accuracy_recipient_token_recommend": 1.0, "train_accuracy_recipient_token_recommend_total": 2, "train_total_number_recipient_token": 73 }, { "epoch": 0.7786561264822134, "grad_norm": 1.3946518618786823, "learning_rate": 1.2346426699819456e-07, "loss": 0.1365, "step": 197 }, { "epoch": 0.7786561264822134, "step": 197, "train_accuracy_recipient_token": 0.948051948051948, "train_accuracy_recipient_token_all": 0.9090909090909091, "train_accuracy_recipient_token_all_total": 44, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 19, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 2, "train_total_number_recipient_token": 77 }, { "epoch": 0.782608695652174, "grad_norm": 1.9314611523202747, "learning_rate": 1.1927702081543278e-07, "loss": 0.2085, "step": 198 }, { "epoch": 0.782608695652174, "step": 198, "train_accuracy_recipient_token": 0.92, "train_accuracy_recipient_token_all": 0.95, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.8823529411764706, "train_accuracy_recipient_token_get_total": 17, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_accuracy_recipient_token_stamp": 0.75, "train_accuracy_recipient_token_stamp_total": 4, "train_total_number_recipient_token": 75 }, { "epoch": 0.7865612648221344, "grad_norm": 2.6148618471256224, "learning_rate": 1.1515237411129697e-07, "loss": 0.2887, "step": 199 }, { "epoch": 0.7865612648221344, "step": 199, "train_accuracy_recipient_token": 0.9733333333333334, "train_accuracy_recipient_token_all": 0.975609756097561, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 0.9473684210526315, "train_accuracy_recipient_token_get_total": 19, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_accuracy_recipient_token_initialize": 1.0, "train_accuracy_recipient_token_initialize_total": 1, "train_total_number_recipient_token": 75 }, { "epoch": 0.7905138339920948, "grad_norm": 1.7909900659299498, "learning_rate": 1.1109100507110131e-07, "loss": 0.1908, "step": 200 }, { "epoch": 0.7905138339920948, "step": 200, "train_accuracy_recipient_token": 0.9436619718309859, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_building": 1.0, "train_accuracy_recipient_token_building_total": 4, "train_accuracy_recipient_token_configure": 1.0, "train_accuracy_recipient_token_configure_total": 5, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 12, "train_total_number_recipient_token": 71 }, { "epoch": 0.7944664031620553, "grad_norm": 1.7701515306384108, "learning_rate": 1.0709358147587883e-07, "loss": 0.1555, "step": 201 }, { "epoch": 0.7944664031620553, "step": 201, "train_accuracy_recipient_token": 0.9493670886075949, "train_accuracy_recipient_token_all": 0.9318181818181818, "train_accuracy_recipient_token_all_total": 44, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 0.96, "train_accuracy_recipient_token_get_total": 25, "train_accuracy_recipient_token_locate": 1.0, "train_accuracy_recipient_token_locate_total": 1, "train_total_number_recipient_token": 79 }, { "epoch": 0.7984189723320159, "grad_norm": 1.9341012616201887, "learning_rate": 1.0316076059258389e-07, "loss": 0.2009, "step": 202 }, { "epoch": 0.7984189723320159, "step": 202, "train_accuracy_recipient_token": 0.9324324324324325, "train_accuracy_recipient_token_all": 0.9302325581395349, "train_accuracy_recipient_token_all_total": 43, "train_accuracy_recipient_token_find": 0.8333333333333334, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_accuracy_recipient_token_land": 0.75, "train_accuracy_recipient_token_land_total": 4, "train_accuracy_recipient_token_recipe": 1.0, "train_accuracy_recipient_token_recipe_total": 4, "train_total_number_recipient_token": 74 }, { "epoch": 0.8023715415019763, "grad_norm": 1.5295063133595792, "learning_rate": 9.929318906602174e-08, "loss": 0.1584, "step": 203 }, { "epoch": 0.8023715415019763, "step": 203, "train_accuracy_recipient_token": 0.9459459459459459, "train_accuracy_recipient_token_all": 0.975, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.9473684210526315, "train_accuracy_recipient_token_get_total": 19, "train_accuracy_recipient_token_go": 1.0, "train_accuracy_recipient_token_go_total": 3, "train_accuracy_recipient_token_identify": 0.6666666666666666, "train_accuracy_recipient_token_identify_total": 3, "train_total_number_recipient_token": 74 }, { "epoch": 0.8063241106719368, "grad_norm": 2.1045397688936602, "learning_rate": 9.549150281252632e-08, "loss": 0.2011, "step": 204 }, { "epoch": 0.8063241106719368, "step": 204, "train_accuracy_recipient_token": 0.9, "train_accuracy_recipient_token_all": 0.8536585365853658, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_convert": 1.0, "train_accuracy_recipient_token_convert_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 0.9444444444444444, "train_accuracy_recipient_token_get_total": 18, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_total_number_recipient_token": 70 }, { "epoch": 0.8102766798418972, "grad_norm": 2.118710128724103, "learning_rate": 9.175632691540064e-08, "loss": 0.2123, "step": 205 }, { "epoch": 0.8102766798418972, "step": 205, "train_accuracy_recipient_token": 0.9428571428571428, "train_accuracy_recipient_token_all": 0.9024390243902439, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 16, "train_accuracy_recipient_token_h": 1.0, "train_accuracy_recipient_token_h_total": 4, "train_total_number_recipient_token": 70 }, { "epoch": 0.8142292490118577, "grad_norm": 1.9451652992044783, "learning_rate": 8.808827552213916e-08, "loss": 0.1354, "step": 206 }, { "epoch": 0.8142292490118577, "step": 206, "train_accuracy_recipient_token": 0.9466666666666667, "train_accuracy_recipient_token_all": 0.9523809523809523, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_crop": 1.0, "train_accuracy_recipient_token_crop_total": 1, "train_accuracy_recipient_token_f": 1.0, "train_accuracy_recipient_token_f_total": 1, "train_accuracy_recipient_token_get": 0.9259259259259259, "train_accuracy_recipient_token_get_total": 27, "train_accuracy_recipient_token_har": 1.0, "train_accuracy_recipient_token_har_total": 1, "train_total_number_recipient_token": 75 }, { "epoch": 0.8181818181818182, "grad_norm": 2.636108327778601, "learning_rate": 8.448795174344803e-08, "loss": 0.1748, "step": 207 }, { "epoch": 0.8181818181818182, "step": 207, "train_accuracy_recipient_token": 0.9305555555555556, "train_accuracy_recipient_token_all": 0.9069767441860465, "train_accuracy_recipient_token_all_total": 43, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 1, "train_accuracy_recipient_token_find": 0.8333333333333334, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 15, "train_accuracy_recipient_token_rail": 1.0, "train_accuracy_recipient_token_rail_total": 4, "train_total_number_recipient_token": 72 }, { "epoch": 0.8221343873517787, "grad_norm": 2.0390257850352396, "learning_rate": 8.09559475540797e-08, "loss": 0.1921, "step": 208 }, { "epoch": 0.8221343873517787, "step": 208, "train_accuracy_recipient_token": 0.9452054794520548, "train_accuracy_recipient_token_all": 0.925, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_find": 0.5, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 23, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 4, "train_accuracy_recipient_token_motor": 1.0, "train_accuracy_recipient_token_motor_total": 3, "train_total_number_recipient_token": 73 }, { "epoch": 0.8260869565217391, "grad_norm": 1.7670861334022596, "learning_rate": 7.749284369549952e-08, "loss": 0.1319, "step": 209 }, { "epoch": 0.8260869565217391, "step": 209, "train_accuracy_recipient_token": 0.935064935064935, "train_accuracy_recipient_token_all": 0.95, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_be": 0.75, "train_accuracy_recipient_token_be_total": 4, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 3, "train_accuracy_recipient_token_find": 0.9333333333333333, "train_accuracy_recipient_token_find_total": 15, "train_accuracy_recipient_token_gr": 1.0, "train_accuracy_recipient_token_gr_total": 2, "train_total_number_recipient_token": 77 }, { "epoch": 0.8300395256916996, "grad_norm": 1.1644905378645993, "learning_rate": 7.409920958039794e-08, "loss": 0.1307, "step": 210 }, { "epoch": 0.8300395256916996, "step": 210, "train_accuracy_recipient_token": 0.9130434782608695, "train_accuracy_recipient_token_all": 0.8780487804878049, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_calculate": 0.6666666666666666, "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_fish": 1.0, "train_accuracy_recipient_token_fish_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 7, "train_total_number_recipient_token": 69 }, { "epoch": 0.83399209486166, "grad_norm": 2.743333625917044, "learning_rate": 7.077560319906694e-08, "loss": 0.2152, "step": 211 }, { "epoch": 0.83399209486166, "step": 211, "train_accuracy_recipient_token": 0.9868421052631579, "train_accuracy_recipient_token_all": 0.9761904761904762, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 20, "train_accuracy_recipient_token_pill": 1.0, "train_accuracy_recipient_token_pill_total": 2, "train_accuracy_recipient_token_wild": 1.0, "train_accuracy_recipient_token_wild_total": 2, "train_total_number_recipient_token": 76 }, { "epoch": 0.8379446640316206, "grad_norm": 1.5221311005160505, "learning_rate": 6.752257102765324e-08, "loss": 0.2088, "step": 212 }, { "epoch": 0.8379446640316206, "step": 212, "train_accuracy_recipient_token": 0.9577464788732394, "train_accuracy_recipient_token_all": 0.975609756097561, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_calculate": 0.5, "train_accuracy_recipient_token_calculate_total": 2, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.9285714285714286, "train_accuracy_recipient_token_get_total": 14, "train_total_number_recipient_token": 71 }, { "epoch": 0.841897233201581, "grad_norm": 1.6788114880444847, "learning_rate": 6.43406479383053e-08, "loss": 0.2107, "step": 213 }, { "epoch": 0.841897233201581, "step": 213, "train_accuracy_recipient_token": 0.9324324324324325, "train_accuracy_recipient_token_all": 0.9512195121951219, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_dec": 0.0, "train_accuracy_recipient_token_dec_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.9230769230769231, "train_accuracy_recipient_token_get_total": 26, "train_total_number_recipient_token": 74 }, { "epoch": 0.8458498023715415, "grad_norm": 3.878765368712377, "learning_rate": 6.123035711122859e-08, "loss": 0.099, "step": 214 }, { "epoch": 0.8458498023715415, "step": 214, "train_accuracy_recipient_token": 0.9466666666666667, "train_accuracy_recipient_token_all": 0.9318181818181818, "train_accuracy_recipient_token_all_total": 44, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_detect": 1.0, "train_accuracy_recipient_token_detect_total": 1, "train_accuracy_recipient_token_find": 0.8, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 17, "train_total_number_recipient_token": 75 }, { "epoch": 0.849802371541502, "grad_norm": 2.0913476399248614, "learning_rate": 5.819220994866236e-08, "loss": 0.1635, "step": 215 }, { "epoch": 0.849802371541502, "step": 215, "train_accuracy_recipient_token": 0.9722222222222222, "train_accuracy_recipient_token_all": 0.95, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 26, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 2, "train_total_number_recipient_token": 72 }, { "epoch": 0.8537549407114624, "grad_norm": 1.486281846638071, "learning_rate": 5.5226705990794156e-08, "loss": 0.1866, "step": 216 }, { "epoch": 0.8537549407114624, "step": 216, "train_accuracy_recipient_token": 0.9210526315789473, "train_accuracy_recipient_token_all": 0.8863636363636364, "train_accuracy_recipient_token_all_total": 44, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 2, "train_accuracy_recipient_token_find": 0.8, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 10, "train_accuracy_recipient_token_land": 1.0, "train_accuracy_recipient_token_land_total": 3, "train_total_number_recipient_token": 76 }, { "epoch": 0.857707509881423, "grad_norm": 1.3477409446750337, "learning_rate": 5.2334332833623487e-08, "loss": 0.1529, "step": 217 }, { "epoch": 0.857707509881423, "step": 217, "train_accuracy_recipient_token": 0.9459459459459459, "train_accuracy_recipient_token_all": 0.9523809523809523, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 0.9230769230769231, "train_accuracy_recipient_token_get_total": 13, "train_accuracy_recipient_token_mach": 1.0, "train_accuracy_recipient_token_mach_total": 3, "train_accuracy_recipient_token_train": 1.0, "train_accuracy_recipient_token_train_total": 4, "train_total_number_recipient_token": 74 }, { "epoch": 0.8616600790513834, "grad_norm": 1.4209581709952837, "learning_rate": 4.951556604879048e-08, "loss": 0.147, "step": 218 }, { "epoch": 0.8616600790513834, "step": 218, "train_accuracy_recipient_token": 0.9459459459459459, "train_accuracy_recipient_token_all": 0.9024390243902439, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_check": 1.0, "train_accuracy_recipient_token_check_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 19, "train_accuracy_recipient_token_pig": 1.0, "train_accuracy_recipient_token_pig_total": 3, "train_total_number_recipient_token": 74 }, { "epoch": 0.8656126482213439, "grad_norm": 2.353150905350018, "learning_rate": 4.6770869105380914e-08, "loss": 0.3008, "step": 219 }, { "epoch": 0.8656126482213439, "step": 219, "train_accuracy_recipient_token": 0.9210526315789473, "train_accuracy_recipient_token_all": 0.9487179487179487, "train_accuracy_recipient_token_all_total": 39, "train_accuracy_recipient_token_calculate": 0.75, "train_accuracy_recipient_token_calculate_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 8, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 10, "train_accuracy_recipient_token_multiply": 0.75, "train_accuracy_recipient_token_multiply_total": 4, "train_total_number_recipient_token": 76 }, { "epoch": 0.8695652173913043, "grad_norm": 6.643712517533024, "learning_rate": 4.4100693293721516e-08, "loss": 0.1543, "step": 220 }, { "epoch": 0.8695652173913043, "step": 220, "train_accuracy_recipient_token": 0.9871794871794872, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_draw": 1.0, "train_accuracy_recipient_token_draw_total": 5, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.9545454545454546, "train_accuracy_recipient_token_get_total": 22, "train_accuracy_recipient_token_text": 1.0, "train_accuracy_recipient_token_text_total": 6, "train_total_number_recipient_token": 78 }, { "epoch": 0.8735177865612648, "grad_norm": 1.8814271761070478, "learning_rate": 4.150547765117746e-08, "loss": 0.1942, "step": 221 }, { "epoch": 0.8735177865612648, "step": 221, "train_accuracy_recipient_token": 0.9285714285714286, "train_accuracy_recipient_token_all": 0.9024390243902439, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 2, "train_accuracy_recipient_token_extract": 1.0, "train_accuracy_recipient_token_extract_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 14, "train_total_number_recipient_token": 70 }, { "epoch": 0.8774703557312253, "grad_norm": 1.7844915508947763, "learning_rate": 3.898564888996475e-08, "loss": 0.2197, "step": 222 }, { "epoch": 0.8774703557312253, "step": 222, "train_accuracy_recipient_token": 0.92, "train_accuracy_recipient_token_all": 0.9069767441860465, "train_accuracy_recipient_token_all_total": 43, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.9, "train_accuracy_recipient_token_get_total": 20, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 2, "train_total_number_recipient_token": 75 }, { "epoch": 0.8814229249011858, "grad_norm": 3.299872748493493, "learning_rate": 3.654162132698918e-08, "loss": 0.1194, "step": 223 }, { "epoch": 0.8814229249011858, "step": 223, "train_accuracy_recipient_token": 0.9102564102564102, "train_accuracy_recipient_token_all": 0.9555555555555556, "train_accuracy_recipient_token_all_total": 45, "train_accuracy_recipient_token_building": 0.5, "train_accuracy_recipient_token_building_total": 2, "train_accuracy_recipient_token_find": 0.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 0.92, "train_accuracy_recipient_token_get_total": 25, "train_accuracy_recipient_token_stat": 1.0, "train_accuracy_recipient_token_stat_total": 2, "train_total_number_recipient_token": 78 }, { "epoch": 0.8853754940711462, "grad_norm": 1.6955366557778766, "learning_rate": 3.417379681572296e-08, "loss": 0.2528, "step": 224 }, { "epoch": 0.8853754940711462, "step": 224, "train_accuracy_recipient_token": 0.96, "train_accuracy_recipient_token_all": 0.975609756097561, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_draw": 0.8, "train_accuracy_recipient_token_draw_total": 5, "train_accuracy_recipient_token_exp": 1.0, "train_accuracy_recipient_token_exp_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 16, "train_accuracy_recipient_token_plant": 1.0, "train_accuracy_recipient_token_plant_total": 3, "train_total_number_recipient_token": 75 }, { "epoch": 0.8893280632411067, "grad_norm": 1.7655077870039768, "learning_rate": 3.188256468013139e-08, "loss": 0.2176, "step": 225 }, { "epoch": 0.8893280632411067, "step": 225, "train_accuracy_recipient_token": 0.9583333333333334, "train_accuracy_recipient_token_all": 0.9285714285714286, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_bird": 1.0, "train_accuracy_recipient_token_bird_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 22, "train_accuracy_recipient_token_search": 1.0, "train_accuracy_recipient_token_search_total": 1, "train_total_number_recipient_token": 72 }, { "epoch": 0.8932806324110671, "grad_norm": 1.9719588151581182, "learning_rate": 2.9668301650658756e-08, "loss": 0.2298, "step": 226 }, { "epoch": 0.8932806324110671, "step": 226, "train_accuracy_recipient_token": 0.9333333333333333, "train_accuracy_recipient_token_all": 0.9285714285714286, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 0.9545454545454546, "train_accuracy_recipient_token_get_total": 22, "train_accuracy_recipient_token_list": 1.0, "train_accuracy_recipient_token_list_total": 1, "train_accuracy_recipient_token_people": 1.0, "train_accuracy_recipient_token_people_total": 2, "train_total_number_recipient_token": 75 }, { "epoch": 0.8972332015810277, "grad_norm": 11.545044626155494, "learning_rate": 2.753137180228543e-08, "loss": 0.2913, "step": 227 }, { "epoch": 0.8972332015810277, "step": 227, "train_accuracy_recipient_token": 0.9605263157894737, "train_accuracy_recipient_token_all": 0.9512195121951219, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_building": 1.0, "train_accuracy_recipient_token_building_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 20, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_total_number_recipient_token": 76 }, { "epoch": 0.9011857707509882, "grad_norm": 1.8421617553050926, "learning_rate": 2.547212649466568e-08, "loss": 0.2079, "step": 228 }, { "epoch": 0.9011857707509882, "step": 228, "train_accuracy_recipient_token": 0.9142857142857143, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_check": 0.3333333333333333, "train_accuracy_recipient_token_check_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 20, "train_accuracy_recipient_token_recogn": 1.0, "train_accuracy_recipient_token_recogn_total": 1, "train_total_number_recipient_token": 70 }, { "epoch": 0.9051383399209486, "grad_norm": 1.68373471973172, "learning_rate": 2.3490904314356407e-08, "loss": 0.2109, "step": 229 }, { "epoch": 0.9051383399209486, "step": 229, "train_accuracy_recipient_token": 0.9452054794520548, "train_accuracy_recipient_token_all": 0.925, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_event": 1.0, "train_accuracy_recipient_token_event_total": 2, "train_accuracy_recipient_token_find": 0.75, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 18, "train_accuracy_recipient_token_stat": 1.0, "train_accuracy_recipient_token_stat_total": 3, "train_total_number_recipient_token": 73 }, { "epoch": 0.9090909090909091, "grad_norm": 1.4731306126600507, "learning_rate": 2.1588031019145636e-08, "loss": 0.1555, "step": 230 }, { "epoch": 0.9090909090909091, "step": 230, "train_accuracy_recipient_token": 0.9367088607594937, "train_accuracy_recipient_token_all": 0.926829268292683, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 0.9473684210526315, "train_accuracy_recipient_token_get_total": 19, "train_accuracy_recipient_token_gr": 1.0, "train_accuracy_recipient_token_gr_total": 3, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 4, "train_total_number_recipient_token": 79 }, { "epoch": 0.9130434782608695, "grad_norm": 2.2815977902694637, "learning_rate": 1.976381948449035e-08, "loss": 0.2256, "step": 231 }, { "epoch": 0.9130434782608695, "step": 231, "train_accuracy_recipient_token": 0.96, "train_accuracy_recipient_token_all": 0.975609756097561, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.9523809523809523, "train_accuracy_recipient_token_get_total": 21, "train_accuracy_recipient_token_house": 1.0, "train_accuracy_recipient_token_house_total": 3, "train_accuracy_recipient_token_market": 1.0, "train_accuracy_recipient_token_market_total": 3, "train_total_number_recipient_token": 75 }, { "epoch": 0.9169960474308301, "grad_norm": 1.3438083724717036, "learning_rate": 1.8018569652073378e-08, "loss": 0.1013, "step": 232 }, { "epoch": 0.9169960474308301, "step": 232, "train_accuracy_recipient_token": 0.9852941176470589, "train_accuracy_recipient_token_all": 0.975609756097561, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_explain": 1.0, "train_accuracy_recipient_token_explain_total": 3, "train_accuracy_recipient_token_extract": 1.0, "train_accuracy_recipient_token_extract_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 18, "train_total_number_recipient_token": 68 }, { "epoch": 0.9209486166007905, "grad_norm": 2.383894577722877, "learning_rate": 1.6352568480485275e-08, "loss": 0.2999, "step": 233 }, { "epoch": 0.9209486166007905, "step": 233, "train_accuracy_recipient_token": 0.9285714285714286, "train_accuracy_recipient_token_all": 0.926829268292683, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_estimate": 1.0, "train_accuracy_recipient_token_estimate_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 0.9473684210526315, "train_accuracy_recipient_token_get_total": 19, "train_accuracy_recipient_token_property": 1.0, "train_accuracy_recipient_token_property_total": 1, "train_total_number_recipient_token": 70 }, { "epoch": 0.924901185770751, "grad_norm": 2.1686977501107547, "learning_rate": 1.4766089898042677e-08, "loss": 0.1885, "step": 234 }, { "epoch": 0.924901185770751, "step": 234, "train_accuracy_recipient_token": 0.9466666666666667, "train_accuracy_recipient_token_all": 0.925, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_ap": 1.0, "train_accuracy_recipient_token_ap_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 25, "train_accuracy_recipient_token_sand": 0.75, "train_accuracy_recipient_token_sand_total": 4, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 75 }, { "epoch": 0.9288537549407114, "grad_norm": 1.598510439225642, "learning_rate": 1.3259394757747677e-08, "loss": 0.1762, "step": 235 }, { "epoch": 0.9288537549407114, "step": 235, "train_accuracy_recipient_token": 0.875, "train_accuracy_recipient_token_all": 0.8780487804878049, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_compare": 0.6666666666666666, "train_accuracy_recipient_token_compare_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 7, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_total_number_recipient_token": 72 }, { "epoch": 0.932806324110672, "grad_norm": 1.397622750202262, "learning_rate": 1.183273079439795e-08, "loss": 0.2255, "step": 236 }, { "epoch": 0.932806324110672, "step": 236, "train_accuracy_recipient_token": 0.9466666666666667, "train_accuracy_recipient_token_all": 0.975, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_building": 0.6666666666666666, "train_accuracy_recipient_token_building_total": 3, "train_accuracy_recipient_token_fabric": 1.0, "train_accuracy_recipient_token_fabric_total": 4, "train_accuracy_recipient_token_get": 0.9333333333333333, "train_accuracy_recipient_token_get_total": 15, "train_accuracy_recipient_token_gr": 0.6666666666666666, "train_accuracy_recipient_token_gr_total": 3, "train_total_number_recipient_token": 75 }, { "epoch": 0.9367588932806324, "grad_norm": 5.074140188359309, "learning_rate": 1.0486332583853564e-08, "loss": 0.1866, "step": 237 }, { "epoch": 0.9367588932806324, "step": 237, "train_accuracy_recipient_token": 0.918918918918919, "train_accuracy_recipient_token_all": 0.9069767441860465, "train_accuracy_recipient_token_all_total": 43, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_format": 1.0, "train_accuracy_recipient_token_format_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 17, "train_accuracy_recipient_token_sum": 1.0, "train_accuracy_recipient_token_sum_total": 1, "train_total_number_recipient_token": 74 }, { "epoch": 0.9407114624505929, "grad_norm": 1.4226611971572756, "learning_rate": 9.22042150446728e-09, "loss": 0.1539, "step": 238 }, { "epoch": 0.9407114624505929, "step": 238, "train_accuracy_recipient_token": 0.9315068493150684, "train_accuracy_recipient_token_all": 0.9761904761904762, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_find": 0.8333333333333334, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 0.9375, "train_accuracy_recipient_token_get_total": 16, "train_accuracy_recipient_token_plane": 1.0, "train_accuracy_recipient_token_plane_total": 3, "train_accuracy_recipient_token_zi": 0.0, "train_accuracy_recipient_token_zi_total": 2, "train_total_number_recipient_token": 73 }, { "epoch": 0.9446640316205533, "grad_norm": 1.6213155939366684, "learning_rate": 8.035205700685165e-09, "loss": 0.1698, "step": 239 }, { "epoch": 0.9446640316205533, "step": 239, "train_accuracy_recipient_token": 0.9342105263157895, "train_accuracy_recipient_token_all": 0.9285714285714286, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_find": 0.8333333333333334, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 0.9411764705882353, "train_accuracy_recipient_token_get_total": 17, "train_accuracy_recipient_token_plant": 1.0, "train_accuracy_recipient_token_plant_total": 3, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 4, "train_total_number_recipient_token": 76 }, { "epoch": 0.9486166007905138, "grad_norm": 1.3790819094027, "learning_rate": 6.930880048822529e-09, "loss": 0.2001, "step": 240 }, { "epoch": 0.9486166007905138, "step": 240, "train_accuracy_recipient_token": 0.9166666666666666, "train_accuracy_recipient_token_all": 0.8936170212765957, "train_accuracy_recipient_token_all_total": 47, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 16, "train_accuracy_recipient_token_predict": 1.0, "train_accuracy_recipient_token_predict_total": 1, "train_accuracy_recipient_token_stamp": 1.0, "train_accuracy_recipient_token_stamp_total": 2, "train_total_number_recipient_token": 72 }, { "epoch": 0.9525691699604744, "grad_norm": 1.8142970154520788, "learning_rate": 5.907626125022158e-09, "loss": 0.2401, "step": 241 }, { "epoch": 0.9525691699604744, "step": 241, "train_accuracy_recipient_token": 0.9857142857142858, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 0.9411764705882353, "train_accuracy_recipient_token_get_total": 17, "train_accuracy_recipient_token_ne": 1.0, "train_accuracy_recipient_token_ne_total": 3, "train_total_number_recipient_token": 70 }, { "epoch": 0.9565217391304348, "grad_norm": 2.027889935194688, "learning_rate": 4.965612175399092e-09, "loss": 0.2322, "step": 242 }, { "epoch": 0.9565217391304348, "step": 242, "train_accuracy_recipient_token": 0.9863013698630136, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_c": 1.0, "train_accuracy_recipient_token_c_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 0.9411764705882353, "train_accuracy_recipient_token_get_total": 17, "train_accuracy_recipient_token_pool": 1.0, "train_accuracy_recipient_token_pool_total": 4, "train_total_number_recipient_token": 73 }, { "epoch": 0.9604743083003953, "grad_norm": 1.7115391757637188, "learning_rate": 4.104993088376974e-09, "loss": 0.1392, "step": 243 }, { "epoch": 0.9604743083003953, "step": 243, "train_accuracy_recipient_token": 0.9615384615384616, "train_accuracy_recipient_token_all": 0.9512195121951219, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_count": 1.0, "train_accuracy_recipient_token_count_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 7, "train_accuracy_recipient_token_get": 0.9523809523809523, "train_accuracy_recipient_token_get_total": 21, "train_accuracy_recipient_token_learn": 1.0, "train_accuracy_recipient_token_learn_total": 1, "train_total_number_recipient_token": 78 }, { "epoch": 0.9644268774703557, "grad_norm": 1.8892062199572783, "learning_rate": 3.3259103692209745e-09, "loss": 0.133, "step": 244 }, { "epoch": 0.9644268774703557, "step": 244, "train_accuracy_recipient_token": 0.9714285714285714, "train_accuracy_recipient_token_all": 0.9761904761904762, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_find": 0.8888888888888888, "train_accuracy_recipient_token_find_total": 9, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 6, "train_accuracy_recipient_token_recogn": 1.0, "train_accuracy_recipient_token_recogn_total": 1, "train_total_number_recipient_token": 70 }, { "epoch": 0.9683794466403162, "grad_norm": 1.9799653566585742, "learning_rate": 2.628492116771297e-09, "loss": 0.2771, "step": 245 }, { "epoch": 0.9683794466403162, "step": 245, "train_accuracy_recipient_token": 0.9583333333333334, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_compare": 1.0, "train_accuracy_recipient_token_compare_total": 1, "train_accuracy_recipient_token_download": 0.0, "train_accuracy_recipient_token_download_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 19, "train_total_number_recipient_token": 72 }, { "epoch": 0.9723320158102767, "grad_norm": 1.3093227803963279, "learning_rate": 2.0128530023804656e-09, "loss": 0.1729, "step": 246 }, { "epoch": 0.9723320158102767, "step": 246, "train_accuracy_recipient_token": 0.9333333333333333, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 1, "train_accuracy_recipient_token_find": 0.6666666666666666, "train_accuracy_recipient_token_find_total": 9, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 16, "train_accuracy_recipient_token_park": 0.75, "train_accuracy_recipient_token_park_total": 4, "train_total_number_recipient_token": 75 }, { "epoch": 0.9762845849802372, "grad_norm": 1.608532435029974, "learning_rate": 1.4790942510590766e-09, "loss": 0.193, "step": 247 }, { "epoch": 0.9762845849802372, "step": 247, "train_accuracy_recipient_token": 0.971830985915493, "train_accuracy_recipient_token_all": 0.975, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 1, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 27, "train_accuracy_recipient_token_validate": 1.0, "train_accuracy_recipient_token_validate_total": 1, "train_total_number_recipient_token": 71 }, { "epoch": 0.9802371541501976, "grad_norm": 1.9569259889356765, "learning_rate": 1.0273036248318324e-09, "loss": 0.1858, "step": 248 }, { "epoch": 0.9802371541501976, "step": 248, "train_accuracy_recipient_token": 0.9041095890410958, "train_accuracy_recipient_token_all": 0.9333333333333333, "train_accuracy_recipient_token_all_total": 45, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 0.8181818181818182, "train_accuracy_recipient_token_get_total": 11, "train_accuracy_recipient_token_motor": 1.0, "train_accuracy_recipient_token_motor_total": 3, "train_accuracy_recipient_token_schedule": 0.5, "train_accuracy_recipient_token_schedule_total": 2, "train_total_number_recipient_token": 73 }, { "epoch": 0.9841897233201581, "grad_norm": 8.659445713267978, "learning_rate": 6.575554083078083e-10, "loss": 0.1525, "step": 249 }, { "epoch": 0.9841897233201581, "step": 249, "train_accuracy_recipient_token": 0.9420289855072463, "train_accuracy_recipient_token_all": 0.9512195121951219, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_generate": 0.5, "train_accuracy_recipient_token_generate_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 14, "train_accuracy_recipient_token_property": 1.0, "train_accuracy_recipient_token_property_total": 4, "train_accuracy_recipient_token_translate": 0.5, "train_accuracy_recipient_token_translate_total": 2, "train_total_number_recipient_token": 69 }, { "epoch": 0.9881422924901185, "grad_norm": 1.488518723492788, "learning_rate": 3.699103964661665e-10, "loss": 0.1536, "step": 250 }, { "epoch": 0.9881422924901185, "step": 250, "train_accuracy_recipient_token": 0.881578947368421, "train_accuracy_recipient_token_all": 0.9024390243902439, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 3, "train_accuracy_recipient_token_artifact": 1.0, "train_accuracy_recipient_token_artifact_total": 4, "train_accuracy_recipient_token_find": 0.75, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 0.9285714285714286, "train_accuracy_recipient_token_get_total": 14, "train_total_number_recipient_token": 76 }, { "epoch": 0.9920948616600791, "grad_norm": 3.2215711951742074, "learning_rate": 1.6441588466009627e-10, "loss": 0.155, "step": 251 }, { "epoch": 0.9920948616600791, "step": 251, "train_accuracy_recipient_token": 0.9315068493150684, "train_accuracy_recipient_token_all": 0.9565217391304348, "train_accuracy_recipient_token_all_total": 46, "train_accuracy_recipient_token_find": 0.8571428571428571, "train_accuracy_recipient_token_find_total": 7, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 13, "train_accuracy_recipient_token_plan": 0.0, "train_accuracy_recipient_token_plan_total": 1, "train_accuracy_recipient_token_write": 1.0, "train_accuracy_recipient_token_write_total": 3, "train_total_number_recipient_token": 73 }, { "epoch": 0.9960474308300395, "grad_norm": 1.978305515531869, "learning_rate": 4.110566084036815e-11, "loss": 0.1976, "step": 252 }, { "epoch": 0.9960474308300395, "step": 252, "train_accuracy_recipient_token": 0.9444444444444444, "train_accuracy_recipient_token_all": 0.9210526315789473, "train_accuracy_recipient_token_all_total": 38, "train_accuracy_recipient_token_find": 0.9166666666666666, "train_accuracy_recipient_token_find_total": 12, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 21, "train_accuracy_recipient_token_order": 1.0, "train_accuracy_recipient_token_order_total": 1, "train_total_number_recipient_token": 72 }, { "epoch": 0.9960474308300395, "step": 252, "train_accuracy_recipient_token": 0.9436619718309859, "train_accuracy_recipient_token_all": 0.9047619047619048, "train_accuracy_recipient_token_all_total": 42, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 15, "train_accuracy_recipient_token_histor": 1.0, "train_accuracy_recipient_token_histor_total": 3, "train_accuracy_recipient_token_translate": 1.0, "train_accuracy_recipient_token_translate_total": 3, "train_total_number_recipient_token": 71 }, { "epoch": 0.9960474308300395, "step": 252, "train_accuracy_recipient_token": 0.9027777777777778, "train_accuracy_recipient_token_all": 0.8478260869565217, "train_accuracy_recipient_token_all_total": 46, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 4, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 8, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_accuracy_recipient_token_interpret": 1.0, "train_accuracy_recipient_token_interpret_total": 2, "train_total_number_recipient_token": 72 }, { "epoch": 0.9960474308300395, "step": 252, "train_accuracy_recipient_token": 0.9868421052631579, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 3, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 6, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 14, "train_accuracy_recipient_token_tour": 1.0, "train_accuracy_recipient_token_tour_total": 4, "train_total_number_recipient_token": 76 }, { "epoch": 0.9960474308300395, "step": 252, "train_accuracy_recipient_token": 0.9722222222222222, "train_accuracy_recipient_token_all": 0.9512195121951219, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_artifact": 1.0, "train_accuracy_recipient_token_artifact_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 2, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 20, "train_accuracy_recipient_token_select": 1.0, "train_accuracy_recipient_token_select_total": 2, "train_total_number_recipient_token": 72 }, { "epoch": 0.9960474308300395, "step": 252, "train_accuracy_recipient_token": 0.9285714285714286, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 3, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 16, "train_accuracy_recipient_token_gr": 1.0, "train_accuracy_recipient_token_gr_total": 3, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_total_number_recipient_token": 70 }, { "epoch": 0.9960474308300395, "step": 252, "train_accuracy_recipient_token": 0.8904109589041096, "train_accuracy_recipient_token_all": 0.9, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_create": 1.0, "train_accuracy_recipient_token_create_total": 2, "train_accuracy_recipient_token_customize": 0.5, "train_accuracy_recipient_token_customize_total": 4, "train_accuracy_recipient_token_get": 0.9130434782608695, "train_accuracy_recipient_token_get_total": 23, "train_accuracy_recipient_token_schedule": 1.0, "train_accuracy_recipient_token_schedule_total": 1, "train_total_number_recipient_token": 73 }, { "epoch": 0.9960474308300395, "step": 252, "train_accuracy_recipient_token": 0.9285714285714286, "train_accuracy_recipient_token_all": 0.8780487804878049, "train_accuracy_recipient_token_all_total": 41, "train_accuracy_recipient_token_eng": 1.0, "train_accuracy_recipient_token_eng_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 11, "train_accuracy_recipient_token_tem": 1.0, "train_accuracy_recipient_token_tem_total": 3, "train_total_number_recipient_token": 70 }, { "epoch": 0.9960474308300395, "step": 252, "train_accuracy_recipient_token": 0.9710144927536232, "train_accuracy_recipient_token_all": 0.975, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_bird": 1.0, "train_accuracy_recipient_token_bird_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 9, "train_accuracy_recipient_token_identify": 1.0, "train_accuracy_recipient_token_identify_total": 2, "train_total_number_recipient_token": 69 }, { "epoch": 0.9960474308300395, "step": 252, "train_accuracy_recipient_token": 0.9583333333333334, "train_accuracy_recipient_token_all": 0.925, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_analyze": 1.0, "train_accuracy_recipient_token_analyze_total": 2, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 19, "train_accuracy_recipient_token_park": 1.0, "train_accuracy_recipient_token_park_total": 2, "train_total_number_recipient_token": 72 }, { "epoch": 0.9960474308300395, "step": 252, "train_accuracy_recipient_token": 0.9436619718309859, "train_accuracy_recipient_token_all": 0.975, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_art": 1.0, "train_accuracy_recipient_token_art_total": 3, "train_accuracy_recipient_token_fruit": 1.0, "train_accuracy_recipient_token_fruit_total": 4, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 15, "train_accuracy_recipient_token_object": 0.6666666666666666, "train_accuracy_recipient_token_object_total": 3, "train_total_number_recipient_token": 71 }, { "epoch": 0.9960474308300395, "step": 252, "train_accuracy_recipient_token": 0.9130434782608695, "train_accuracy_recipient_token_all": 0.875, "train_accuracy_recipient_token_all_total": 40, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 5, "train_accuracy_recipient_token_get": 0.9285714285714286, "train_accuracy_recipient_token_get_total": 14, "train_accuracy_recipient_token_organ": 1.0, "train_accuracy_recipient_token_organ_total": 1, "train_accuracy_recipient_token_stage": 1.0, "train_accuracy_recipient_token_stage_total": 4, "train_total_number_recipient_token": 69 }, { "epoch": 0.9960474308300395, "step": 252, "train_accuracy_recipient_token": 0.9777777777777777, "train_accuracy_recipient_token_all": 1.0, "train_accuracy_recipient_token_all_total": 25, "train_accuracy_recipient_token_calculate": 1.0, "train_accuracy_recipient_token_calculate_total": 5, "train_accuracy_recipient_token_find": 1.0, "train_accuracy_recipient_token_find_total": 1, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 10, "train_accuracy_recipient_token_sun": 1.0, "train_accuracy_recipient_token_sun_total": 1, "train_total_number_recipient_token": 45 }, { "epoch": 0.9960474308300395, "eval_accuracy_recipient_token": 0.9423503325942351, "eval_accuracy_recipient_token_all": 0.9239766081871345, "eval_accuracy_recipient_token_all_total": 513, "eval_accuracy_recipient_token_calculate": 1.0, "eval_accuracy_recipient_token_calculate_total": 13, "eval_accuracy_recipient_token_find": 0.9824561403508771, "eval_accuracy_recipient_token_find_total": 57, "eval_accuracy_recipient_token_get": 0.9846938775510204, "eval_accuracy_recipient_token_get_total": 196, "eval_accuracy_recipient_token_identify": 1.0, "eval_accuracy_recipient_token_identify_total": 6, "eval_loss": 0.17267043888568878, "eval_perplexity": 1.0310433990724794, "eval_runtime": 98.1578, "eval_samples_per_second": 1.029, "eval_steps_per_second": 0.132, "eval_total_number_recipient_token": 902, "step": 252 }, { "epoch": 0.9960474308300395, "step": 252, "train_accuracy_recipient_token": 0.9230769230769231, "train_accuracy_recipient_token_all": 0.8918918918918919, "train_accuracy_recipient_token_all_total": 37, "train_accuracy_recipient_token_book": 1.0, "train_accuracy_recipient_token_book_total": 2, "train_accuracy_recipient_token_explain": 1.0, "train_accuracy_recipient_token_explain_total": 1, "train_accuracy_recipient_token_find": 0.8888888888888888, "train_accuracy_recipient_token_find_total": 9, "train_accuracy_recipient_token_get": 1.0, "train_accuracy_recipient_token_get_total": 14, "train_total_number_recipient_token": 65 }, { "epoch": 1.0, "grad_norm": 1.3026938825316514, "learning_rate": 0.0, "loss": 0.1268, "step": 253 }, { "epoch": 1.0, "step": 253, "total_flos": 0.0, "train_loss": 0.2055490122658933, "train_runtime": 25732.7663, "train_samples_per_second": 0.079, "train_steps_per_second": 0.01 } ], "logging_steps": 1.0, "max_steps": 253, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100.0, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": false, "should_training_stop": false }, "attributes": {} } }, "total_flos": 0.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }