{ "epoch": 15.0, "eval_AerospaceManufacturer_f1": 0.7682119205298014, "eval_AerospaceManufacturer_number": 138, "eval_AerospaceManufacturer_precision": 0.7073170731707317, "eval_AerospaceManufacturer_recall": 0.8405797101449275, "eval_AnatomicalStructure_f1": 0.7006369426751592, "eval_AnatomicalStructure_number": 227, "eval_AnatomicalStructure_precision": 0.6762295081967213, "eval_AnatomicalStructure_recall": 0.7268722466960352, "eval_ArtWork_f1": 0.5801526717557252, "eval_ArtWork_number": 131, "eval_ArtWork_precision": 0.5801526717557252, "eval_ArtWork_recall": 0.5801526717557252, "eval_Artist_f1": 0.7747237177670728, "eval_Artist_number": 1722, "eval_Artist_precision": 0.7565024903154399, "eval_Artist_recall": 0.7938443670150988, "eval_Athlete_f1": 0.7408906882591093, "eval_Athlete_number": 719, "eval_Athlete_precision": 0.7195281782437746, "eval_Athlete_recall": 0.7635605006954103, "eval_CarManufacturer_f1": 0.742857142857143, "eval_CarManufacturer_number": 159, "eval_CarManufacturer_precision": 0.680628272251309, "eval_CarManufacturer_recall": 0.8176100628930818, "eval_Cleric_f1": 0.5868945868945868, "eval_Cleric_number": 201, "eval_Cleric_precision": 0.6866666666666666, "eval_Cleric_recall": 0.5124378109452736, "eval_Clothing_f1": 0.6015037593984963, "eval_Clothing_number": 128, "eval_Clothing_precision": 0.5797101449275363, "eval_Clothing_recall": 0.625, "eval_Disease_f1": 0.6504854368932038, "eval_Disease_number": 198, "eval_Disease_precision": 0.6261682242990654, "eval_Disease_recall": 0.6767676767676768, "eval_Drink_f1": 0.7682119205298013, "eval_Drink_number": 143, "eval_Drink_precision": 0.7295597484276729, "eval_Drink_recall": 0.8111888111888111, "eval_Facility_f1": 0.6799620132953467, "eval_Facility_number": 497, "eval_Facility_precision": 0.6438848920863309, "eval_Facility_recall": 0.7203219315895373, "eval_Food_f1": 0.5968586387434555, "eval_Food_number": 214, "eval_Food_precision": 0.6785714285714286, "eval_Food_recall": 0.5327102803738317, "eval_HumanSettlement_f1": 0.8691834942932396, "eval_HumanSettlement_number": 1689, "eval_HumanSettlement_precision": 0.859375, "eval_HumanSettlement_recall": 0.8792184724689165, "eval_MedicalProcedure_f1": 0.7035830618892507, "eval_MedicalProcedure_number": 142, "eval_MedicalProcedure_precision": 0.6545454545454545, "eval_MedicalProcedure_recall": 0.7605633802816901, "eval_Medication/Vaccine_f1": 0.7409200968523003, "eval_Medication/Vaccine_number": 200, "eval_Medication/Vaccine_precision": 0.7183098591549296, "eval_Medication/Vaccine_recall": 0.765, "eval_MusicalGRP_f1": 0.7399741267787839, "eval_MusicalGRP_number": 372, "eval_MusicalGRP_precision": 0.713216957605985, "eval_MusicalGRP_recall": 0.7688172043010753, "eval_MusicalWork_f1": 0.7275031685678074, "eval_MusicalWork_number": 407, "eval_MusicalWork_precision": 0.7513089005235603, "eval_MusicalWork_recall": 0.7051597051597052, "eval_ORG_f1": 0.6224256292906178, "eval_ORG_number": 667, "eval_ORG_precision": 0.6335403726708074, "eval_ORG_recall": 0.6116941529235382, "eval_OtherLOC_f1": 0.6797066014669927, "eval_OtherLOC_number": 224, "eval_OtherLOC_precision": 0.7513513513513513, "eval_OtherLOC_recall": 0.6205357142857143, "eval_OtherPER_f1": 0.5112474437627812, "eval_OtherPER_number": 859, "eval_OtherPER_precision": 0.45578851412944393, "eval_OtherPER_recall": 0.5820721769499418, "eval_OtherPROD_f1": 0.5797101449275364, "eval_OtherPROD_number": 433, "eval_OtherPROD_precision": 0.6075949367088608, "eval_OtherPROD_recall": 0.5542725173210161, "eval_Politician_f1": 0.5417057169634489, "eval_Politician_number": 603, "eval_Politician_precision": 0.6228448275862069, "eval_Politician_recall": 0.4792703150912106, "eval_PrivateCorp_f1": 0.5806451612903226, "eval_PrivateCorp_number": 129, "eval_PrivateCorp_precision": 0.7159090909090909, "eval_PrivateCorp_recall": 0.4883720930232558, "eval_PublicCorp_f1": 0.6187845303867403, "eval_PublicCorp_number": 243, "eval_PublicCorp_precision": 0.56, "eval_PublicCorp_recall": 0.691358024691358, "eval_Scientist_f1": 0.45212765957446804, "eval_Scientist_number": 189, "eval_Scientist_precision": 0.45454545454545453, "eval_Scientist_recall": 0.4497354497354497, "eval_Software_f1": 0.7576687116564418, "eval_Software_number": 307, "eval_Software_precision": 0.7159420289855073, "eval_Software_recall": 0.8045602605863192, "eval_SportsGRP_f1": 0.8251231527093595, "eval_SportsGRP_number": 385, "eval_SportsGRP_precision": 0.7845433255269321, "eval_SportsGRP_recall": 0.8701298701298701, "eval_SportsManager_f1": 0.5942857142857142, "eval_SportsManager_number": 194, "eval_SportsManager_precision": 0.6666666666666666, "eval_SportsManager_recall": 0.5360824742268041, "eval_Station_f1": 0.7733990147783251, "eval_Station_number": 194, "eval_Station_precision": 0.7405660377358491, "eval_Station_recall": 0.8092783505154639, "eval_Symptom_f1": 0.5925925925925927, "eval_Symptom_number": 129, "eval_Symptom_precision": 0.631578947368421, "eval_Symptom_recall": 0.5581395348837209, "eval_Vehicle_f1": 0.5968819599109132, "eval_Vehicle_number": 206, "eval_Vehicle_precision": 0.551440329218107, "eval_Vehicle_recall": 0.6504854368932039, "eval_VisualWork_f1": 0.7738764044943821, "eval_VisualWork_number": 693, "eval_VisualWork_precision": 0.7537619699042407, "eval_VisualWork_recall": 0.7950937950937951, "eval_WrittenWork_f1": 0.6857654431512982, "eval_WrittenWork_number": 563, "eval_WrittenWork_precision": 0.6913357400722022, "eval_WrittenWork_recall": 0.6802841918294849, "eval_loss": 0.21482913196086884, "eval_overall_accuracy": 0.9355327807299229, "eval_overall_f1": 0.7033048369786461, "eval_overall_precision": 0.6927675707203266, "eval_overall_recall": 0.7141676061630966, "eval_runtime": 44.2333, "eval_samples": 8895, "eval_samples_per_second": 201.093, "eval_steps_per_second": 25.139 }