{ "best_metric": 0.9148148148148149, "best_model_checkpoint": "dinov2-finetuned-har/checkpoint-830", "epoch": 9.91044776119403, "eval_steps": 500, "global_step": 830, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.11940298507462686, "grad_norm": 42.725677490234375, "learning_rate": 6.024096385542169e-06, "loss": 2.9994, "step": 10 }, { "epoch": 0.23880597014925373, "grad_norm": 40.75918197631836, "learning_rate": 1.2048192771084338e-05, "loss": 2.1467, "step": 20 }, { "epoch": 0.3582089552238806, "grad_norm": 36.34815979003906, "learning_rate": 1.8072289156626505e-05, "loss": 1.368, "step": 30 }, { "epoch": 0.47761194029850745, "grad_norm": 41.22068405151367, "learning_rate": 2.4096385542168677e-05, "loss": 1.0901, "step": 40 }, { "epoch": 0.5970149253731343, "grad_norm": 35.41105270385742, "learning_rate": 3.012048192771085e-05, "loss": 0.9148, "step": 50 }, { "epoch": 0.7164179104477612, "grad_norm": 45.00641632080078, "learning_rate": 3.614457831325301e-05, "loss": 0.8958, "step": 60 }, { "epoch": 0.835820895522388, "grad_norm": 32.627593994140625, "learning_rate": 4.2168674698795186e-05, "loss": 0.9745, "step": 70 }, { "epoch": 0.9552238805970149, "grad_norm": 44.3660888671875, "learning_rate": 4.8192771084337354e-05, "loss": 0.9429, "step": 80 }, { "epoch": 0.991044776119403, "eval_accuracy": 0.8328042328042328, "eval_loss": 0.562383234500885, "eval_runtime": 21.3212, "eval_samples_per_second": 88.644, "eval_steps_per_second": 2.814, "step": 83 }, { "epoch": 1.0746268656716418, "grad_norm": 66.64344024658203, "learning_rate": 4.953145917001339e-05, "loss": 0.8966, "step": 90 }, { "epoch": 1.1940298507462686, "grad_norm": 31.583904266357422, "learning_rate": 4.886211512717537e-05, "loss": 0.8292, "step": 100 }, { "epoch": 1.3134328358208955, "grad_norm": 70.04341888427734, "learning_rate": 4.8192771084337354e-05, "loss": 0.8166, "step": 110 }, { "epoch": 1.4328358208955223, "grad_norm": 42.080177307128906, "learning_rate": 4.7523427041499336e-05, "loss": 0.9254, "step": 120 }, { "epoch": 1.5522388059701493, "grad_norm": 123.05072784423828, "learning_rate": 4.685408299866131e-05, "loss": 0.9084, "step": 130 }, { "epoch": 1.671641791044776, "grad_norm": 25.725753784179688, "learning_rate": 4.61847389558233e-05, "loss": 0.8126, "step": 140 }, { "epoch": 1.7910447761194028, "grad_norm": 20.468259811401367, "learning_rate": 4.5515394912985275e-05, "loss": 0.8057, "step": 150 }, { "epoch": 1.9104477611940298, "grad_norm": 24.806772232055664, "learning_rate": 4.484605087014726e-05, "loss": 0.7912, "step": 160 }, { "epoch": 1.9940298507462688, "eval_accuracy": 0.8587301587301587, "eval_loss": 0.47548335790634155, "eval_runtime": 21.1591, "eval_samples_per_second": 89.323, "eval_steps_per_second": 2.836, "step": 167 }, { "epoch": 2.029850746268657, "grad_norm": 26.82819175720215, "learning_rate": 4.417670682730924e-05, "loss": 0.7938, "step": 170 }, { "epoch": 2.1492537313432836, "grad_norm": 30.11927604675293, "learning_rate": 4.350736278447122e-05, "loss": 0.7185, "step": 180 }, { "epoch": 2.2686567164179103, "grad_norm": 32.696815490722656, "learning_rate": 4.2838018741633203e-05, "loss": 0.7642, "step": 190 }, { "epoch": 2.388059701492537, "grad_norm": 34.70363998413086, "learning_rate": 4.2168674698795186e-05, "loss": 0.7867, "step": 200 }, { "epoch": 2.5074626865671643, "grad_norm": 21.691709518432617, "learning_rate": 4.149933065595716e-05, "loss": 0.7376, "step": 210 }, { "epoch": 2.626865671641791, "grad_norm": 37.57550811767578, "learning_rate": 4.082998661311915e-05, "loss": 0.7446, "step": 220 }, { "epoch": 2.746268656716418, "grad_norm": 21.075855255126953, "learning_rate": 4.0160642570281125e-05, "loss": 0.6815, "step": 230 }, { "epoch": 2.8656716417910446, "grad_norm": 29.882211685180664, "learning_rate": 3.949129852744311e-05, "loss": 0.686, "step": 240 }, { "epoch": 2.9850746268656714, "grad_norm": 31.888534545898438, "learning_rate": 3.882195448460509e-05, "loss": 0.7371, "step": 250 }, { "epoch": 2.997014925373134, "eval_accuracy": 0.855026455026455, "eval_loss": 0.4584466516971588, "eval_runtime": 21.3468, "eval_samples_per_second": 88.538, "eval_steps_per_second": 2.811, "step": 251 }, { "epoch": 3.1044776119402986, "grad_norm": 31.14760971069336, "learning_rate": 3.815261044176707e-05, "loss": 0.6986, "step": 260 }, { "epoch": 3.2238805970149254, "grad_norm": 20.720378875732422, "learning_rate": 3.748326639892905e-05, "loss": 0.6277, "step": 270 }, { "epoch": 3.343283582089552, "grad_norm": 22.004247665405273, "learning_rate": 3.6813922356091035e-05, "loss": 0.5936, "step": 280 }, { "epoch": 3.4626865671641793, "grad_norm": 24.986722946166992, "learning_rate": 3.614457831325301e-05, "loss": 0.5917, "step": 290 }, { "epoch": 3.582089552238806, "grad_norm": 27.637083053588867, "learning_rate": 3.5475234270415e-05, "loss": 0.5619, "step": 300 }, { "epoch": 3.701492537313433, "grad_norm": 37.6093864440918, "learning_rate": 3.4805890227576974e-05, "loss": 0.6538, "step": 310 }, { "epoch": 3.8208955223880596, "grad_norm": 23.496280670166016, "learning_rate": 3.413654618473896e-05, "loss": 0.6455, "step": 320 }, { "epoch": 3.9402985074626864, "grad_norm": 23.340776443481445, "learning_rate": 3.346720214190094e-05, "loss": 0.5915, "step": 330 }, { "epoch": 4.0, "eval_accuracy": 0.8761904761904762, "eval_loss": 0.38699546456336975, "eval_runtime": 21.0126, "eval_samples_per_second": 89.946, "eval_steps_per_second": 2.855, "step": 335 }, { "epoch": 4.059701492537314, "grad_norm": 24.867502212524414, "learning_rate": 3.279785809906292e-05, "loss": 0.5275, "step": 340 }, { "epoch": 4.17910447761194, "grad_norm": 22.04759407043457, "learning_rate": 3.21285140562249e-05, "loss": 0.5317, "step": 350 }, { "epoch": 4.298507462686567, "grad_norm": 23.60555076599121, "learning_rate": 3.1459170013386885e-05, "loss": 0.6104, "step": 360 }, { "epoch": 4.417910447761194, "grad_norm": 30.320343017578125, "learning_rate": 3.078982597054887e-05, "loss": 0.5362, "step": 370 }, { "epoch": 4.537313432835821, "grad_norm": 19.35739517211914, "learning_rate": 3.012048192771085e-05, "loss": 0.5487, "step": 380 }, { "epoch": 4.656716417910448, "grad_norm": 27.066503524780273, "learning_rate": 2.9451137884872827e-05, "loss": 0.5296, "step": 390 }, { "epoch": 4.776119402985074, "grad_norm": 35.0575065612793, "learning_rate": 2.878179384203481e-05, "loss": 0.5051, "step": 400 }, { "epoch": 4.895522388059701, "grad_norm": 19.96647071838379, "learning_rate": 2.8112449799196788e-05, "loss": 0.5635, "step": 410 }, { "epoch": 4.991044776119403, "eval_accuracy": 0.8703703703703703, "eval_loss": 0.4037255644798279, "eval_runtime": 21.0465, "eval_samples_per_second": 89.801, "eval_steps_per_second": 2.851, "step": 418 }, { "epoch": 5.014925373134329, "grad_norm": 18.035507202148438, "learning_rate": 2.7443105756358774e-05, "loss": 0.5311, "step": 420 }, { "epoch": 5.134328358208955, "grad_norm": 27.073444366455078, "learning_rate": 2.6773761713520752e-05, "loss": 0.4792, "step": 430 }, { "epoch": 5.253731343283582, "grad_norm": 17.406234741210938, "learning_rate": 2.6104417670682734e-05, "loss": 0.5018, "step": 440 }, { "epoch": 5.373134328358209, "grad_norm": 19.9872989654541, "learning_rate": 2.5435073627844713e-05, "loss": 0.4636, "step": 450 }, { "epoch": 5.492537313432836, "grad_norm": 20.06911849975586, "learning_rate": 2.4765729585006695e-05, "loss": 0.4911, "step": 460 }, { "epoch": 5.611940298507463, "grad_norm": 19.70343780517578, "learning_rate": 2.4096385542168677e-05, "loss": 0.4889, "step": 470 }, { "epoch": 5.731343283582089, "grad_norm": 21.2330379486084, "learning_rate": 2.3427041499330656e-05, "loss": 0.4411, "step": 480 }, { "epoch": 5.850746268656716, "grad_norm": 20.05991554260254, "learning_rate": 2.2757697456492638e-05, "loss": 0.4584, "step": 490 }, { "epoch": 5.970149253731344, "grad_norm": 19.27924919128418, "learning_rate": 2.208835341365462e-05, "loss": 0.498, "step": 500 }, { "epoch": 5.994029850746268, "eval_accuracy": 0.8804232804232804, "eval_loss": 0.3876107931137085, "eval_runtime": 21.1814, "eval_samples_per_second": 89.229, "eval_steps_per_second": 2.833, "step": 502 }, { "epoch": 6.08955223880597, "grad_norm": 23.8977108001709, "learning_rate": 2.1419009370816602e-05, "loss": 0.4212, "step": 510 }, { "epoch": 6.208955223880597, "grad_norm": 25.704334259033203, "learning_rate": 2.074966532797858e-05, "loss": 0.4529, "step": 520 }, { "epoch": 6.3283582089552235, "grad_norm": 24.296876907348633, "learning_rate": 2.0080321285140562e-05, "loss": 0.4707, "step": 530 }, { "epoch": 6.447761194029851, "grad_norm": 17.248048782348633, "learning_rate": 1.9410977242302544e-05, "loss": 0.4178, "step": 540 }, { "epoch": 6.567164179104478, "grad_norm": 18.4821834564209, "learning_rate": 1.8741633199464527e-05, "loss": 0.449, "step": 550 }, { "epoch": 6.686567164179104, "grad_norm": 19.246505737304688, "learning_rate": 1.8072289156626505e-05, "loss": 0.4209, "step": 560 }, { "epoch": 6.8059701492537314, "grad_norm": 20.905685424804688, "learning_rate": 1.7402945113788487e-05, "loss": 0.4344, "step": 570 }, { "epoch": 6.925373134328359, "grad_norm": 16.702898025512695, "learning_rate": 1.673360107095047e-05, "loss": 0.4541, "step": 580 }, { "epoch": 6.997014925373135, "eval_accuracy": 0.8883597883597883, "eval_loss": 0.3611726760864258, "eval_runtime": 21.6673, "eval_samples_per_second": 87.228, "eval_steps_per_second": 2.769, "step": 586 }, { "epoch": 7.044776119402985, "grad_norm": 13.757901191711426, "learning_rate": 1.606425702811245e-05, "loss": 0.3419, "step": 590 }, { "epoch": 7.164179104477612, "grad_norm": 20.324989318847656, "learning_rate": 1.5394912985274433e-05, "loss": 0.3985, "step": 600 }, { "epoch": 7.2835820895522385, "grad_norm": 15.066502571105957, "learning_rate": 1.4725568942436414e-05, "loss": 0.3917, "step": 610 }, { "epoch": 7.402985074626866, "grad_norm": 14.179686546325684, "learning_rate": 1.4056224899598394e-05, "loss": 0.372, "step": 620 }, { "epoch": 7.522388059701493, "grad_norm": 18.657678604125977, "learning_rate": 1.3386880856760376e-05, "loss": 0.3744, "step": 630 }, { "epoch": 7.641791044776119, "grad_norm": 20.829837799072266, "learning_rate": 1.2717536813922356e-05, "loss": 0.3809, "step": 640 }, { "epoch": 7.7611940298507465, "grad_norm": 16.62371063232422, "learning_rate": 1.2048192771084338e-05, "loss": 0.3727, "step": 650 }, { "epoch": 7.880597014925373, "grad_norm": 14.337791442871094, "learning_rate": 1.1378848728246319e-05, "loss": 0.3274, "step": 660 }, { "epoch": 8.0, "grad_norm": 19.433561325073242, "learning_rate": 1.0709504685408301e-05, "loss": 0.3513, "step": 670 }, { "epoch": 8.0, "eval_accuracy": 0.9052910052910053, "eval_loss": 0.32395628094673157, "eval_runtime": 21.5863, "eval_samples_per_second": 87.555, "eval_steps_per_second": 2.78, "step": 670 }, { "epoch": 8.119402985074627, "grad_norm": 21.07122802734375, "learning_rate": 1.0040160642570281e-05, "loss": 0.3392, "step": 680 }, { "epoch": 8.238805970149254, "grad_norm": 21.168922424316406, "learning_rate": 9.370816599732263e-06, "loss": 0.3038, "step": 690 }, { "epoch": 8.35820895522388, "grad_norm": 17.61219024658203, "learning_rate": 8.701472556894244e-06, "loss": 0.3049, "step": 700 }, { "epoch": 8.477611940298507, "grad_norm": 18.213272094726562, "learning_rate": 8.032128514056226e-06, "loss": 0.3423, "step": 710 }, { "epoch": 8.597014925373134, "grad_norm": 16.038349151611328, "learning_rate": 7.362784471218207e-06, "loss": 0.2895, "step": 720 }, { "epoch": 8.716417910447761, "grad_norm": 14.367926597595215, "learning_rate": 6.693440428380188e-06, "loss": 0.3142, "step": 730 }, { "epoch": 8.835820895522389, "grad_norm": 16.960412979125977, "learning_rate": 6.024096385542169e-06, "loss": 0.3338, "step": 740 }, { "epoch": 8.955223880597014, "grad_norm": 13.258671760559082, "learning_rate": 5.3547523427041504e-06, "loss": 0.2963, "step": 750 }, { "epoch": 8.991044776119402, "eval_accuracy": 0.9116402116402117, "eval_loss": 0.3176145553588867, "eval_runtime": 21.2968, "eval_samples_per_second": 88.746, "eval_steps_per_second": 2.817, "step": 753 }, { "epoch": 9.074626865671641, "grad_norm": 17.589509963989258, "learning_rate": 4.685408299866132e-06, "loss": 0.265, "step": 760 }, { "epoch": 9.194029850746269, "grad_norm": 13.759172439575195, "learning_rate": 4.016064257028113e-06, "loss": 0.2718, "step": 770 }, { "epoch": 9.313432835820896, "grad_norm": 13.576536178588867, "learning_rate": 3.346720214190094e-06, "loss": 0.2744, "step": 780 }, { "epoch": 9.432835820895523, "grad_norm": 13.831005096435547, "learning_rate": 2.6773761713520752e-06, "loss": 0.2789, "step": 790 }, { "epoch": 9.552238805970148, "grad_norm": 14.289161682128906, "learning_rate": 2.0080321285140564e-06, "loss": 0.2793, "step": 800 }, { "epoch": 9.671641791044776, "grad_norm": 13.557323455810547, "learning_rate": 1.3386880856760376e-06, "loss": 0.2893, "step": 810 }, { "epoch": 9.791044776119403, "grad_norm": 13.76667308807373, "learning_rate": 6.693440428380188e-07, "loss": 0.2918, "step": 820 }, { "epoch": 9.91044776119403, "grad_norm": 15.072478294372559, "learning_rate": 0.0, "loss": 0.2815, "step": 830 }, { "epoch": 9.91044776119403, "eval_accuracy": 0.9148148148148149, "eval_loss": 0.30782097578048706, "eval_runtime": 22.2064, "eval_samples_per_second": 85.111, "eval_steps_per_second": 2.702, "step": 830 }, { "epoch": 9.91044776119403, "step": 830, "total_flos": 1.0844463059730432e+19, "train_loss": 0.6027212039533868, "train_runtime": 2771.0404, "train_samples_per_second": 38.65, "train_steps_per_second": 0.3 } ], "logging_steps": 10, "max_steps": 830, "num_input_tokens_seen": 0, "num_train_epochs": 10, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.0844463059730432e+19, "train_batch_size": 32, "trial_name": null, "trial_params": null }