|
{ |
|
"best_metric": 0.9164021164021164, |
|
"best_model_checkpoint": "dinov2-base-fa-disabled-finetuned-har/checkpoint-830", |
|
"epoch": 9.91044776119403, |
|
"eval_steps": 500, |
|
"global_step": 830, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.11940298507462686, |
|
"grad_norm": 38.431739807128906, |
|
"learning_rate": 6.024096385542169e-06, |
|
"loss": 3.0066, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.23880597014925373, |
|
"grad_norm": 39.09220504760742, |
|
"learning_rate": 1.2048192771084338e-05, |
|
"loss": 2.0974, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.3582089552238806, |
|
"grad_norm": 36.01054000854492, |
|
"learning_rate": 1.8072289156626505e-05, |
|
"loss": 1.3051, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.47761194029850745, |
|
"grad_norm": 38.050384521484375, |
|
"learning_rate": 2.4096385542168677e-05, |
|
"loss": 1.0243, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.5970149253731343, |
|
"grad_norm": 36.212276458740234, |
|
"learning_rate": 3.012048192771085e-05, |
|
"loss": 0.9214, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.7164179104477612, |
|
"grad_norm": 44.00043487548828, |
|
"learning_rate": 3.614457831325301e-05, |
|
"loss": 0.9179, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.835820895522388, |
|
"grad_norm": 50.121219635009766, |
|
"learning_rate": 4.2168674698795186e-05, |
|
"loss": 0.9083, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.9552238805970149, |
|
"grad_norm": 41.91326141357422, |
|
"learning_rate": 4.8192771084337354e-05, |
|
"loss": 0.8554, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.991044776119403, |
|
"eval_accuracy": 0.8322751322751323, |
|
"eval_loss": 0.5251870155334473, |
|
"eval_runtime": 36.3735, |
|
"eval_samples_per_second": 51.961, |
|
"eval_steps_per_second": 1.65, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 1.0746268656716418, |
|
"grad_norm": 30.97153663635254, |
|
"learning_rate": 4.953145917001339e-05, |
|
"loss": 0.9029, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 1.1940298507462686, |
|
"grad_norm": 42.186763763427734, |
|
"learning_rate": 4.886211512717537e-05, |
|
"loss": 0.8259, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 1.3134328358208955, |
|
"grad_norm": 34.84912109375, |
|
"learning_rate": 4.8192771084337354e-05, |
|
"loss": 0.8652, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 1.4328358208955223, |
|
"grad_norm": 33.463218688964844, |
|
"learning_rate": 4.7523427041499336e-05, |
|
"loss": 0.9613, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 1.5522388059701493, |
|
"grad_norm": 25.84391975402832, |
|
"learning_rate": 4.685408299866131e-05, |
|
"loss": 0.9065, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 1.671641791044776, |
|
"grad_norm": 37.509464263916016, |
|
"learning_rate": 4.61847389558233e-05, |
|
"loss": 0.7881, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 1.7910447761194028, |
|
"grad_norm": 21.088834762573242, |
|
"learning_rate": 4.5515394912985275e-05, |
|
"loss": 0.8766, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 1.9104477611940298, |
|
"grad_norm": 33.41486358642578, |
|
"learning_rate": 4.484605087014726e-05, |
|
"loss": 0.8162, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 1.9940298507462688, |
|
"eval_accuracy": 0.8597883597883598, |
|
"eval_loss": 0.4597439467906952, |
|
"eval_runtime": 21.2204, |
|
"eval_samples_per_second": 89.065, |
|
"eval_steps_per_second": 2.827, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 2.029850746268657, |
|
"grad_norm": 37.96963882446289, |
|
"learning_rate": 4.417670682730924e-05, |
|
"loss": 0.7496, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 2.1492537313432836, |
|
"grad_norm": 28.700788497924805, |
|
"learning_rate": 4.350736278447122e-05, |
|
"loss": 0.678, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 2.2686567164179103, |
|
"grad_norm": 26.9548397064209, |
|
"learning_rate": 4.2838018741633203e-05, |
|
"loss": 0.7308, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 2.388059701492537, |
|
"grad_norm": 29.515504837036133, |
|
"learning_rate": 4.2168674698795186e-05, |
|
"loss": 0.7616, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 2.5074626865671643, |
|
"grad_norm": 25.0078067779541, |
|
"learning_rate": 4.149933065595716e-05, |
|
"loss": 0.7281, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 2.626865671641791, |
|
"grad_norm": 30.829315185546875, |
|
"learning_rate": 4.082998661311915e-05, |
|
"loss": 0.7334, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 2.746268656716418, |
|
"grad_norm": 18.689016342163086, |
|
"learning_rate": 4.0160642570281125e-05, |
|
"loss": 0.7141, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 2.8656716417910446, |
|
"grad_norm": 22.5184383392334, |
|
"learning_rate": 3.949129852744311e-05, |
|
"loss": 0.6661, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 2.9850746268656714, |
|
"grad_norm": 25.77634048461914, |
|
"learning_rate": 3.882195448460509e-05, |
|
"loss": 0.7303, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 2.997014925373134, |
|
"eval_accuracy": 0.8587301587301587, |
|
"eval_loss": 0.44030269980430603, |
|
"eval_runtime": 21.4202, |
|
"eval_samples_per_second": 88.234, |
|
"eval_steps_per_second": 2.801, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 3.1044776119402986, |
|
"grad_norm": 19.202600479125977, |
|
"learning_rate": 3.815261044176707e-05, |
|
"loss": 0.6748, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 3.2238805970149254, |
|
"grad_norm": 23.559417724609375, |
|
"learning_rate": 3.748326639892905e-05, |
|
"loss": 0.6641, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 3.343283582089552, |
|
"grad_norm": 33.11872100830078, |
|
"learning_rate": 3.6813922356091035e-05, |
|
"loss": 0.6263, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 3.4626865671641793, |
|
"grad_norm": 20.86111831665039, |
|
"learning_rate": 3.614457831325301e-05, |
|
"loss": 0.6203, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 3.582089552238806, |
|
"grad_norm": 18.73584747314453, |
|
"learning_rate": 3.5475234270415e-05, |
|
"loss": 0.6101, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 3.701492537313433, |
|
"grad_norm": 30.896474838256836, |
|
"learning_rate": 3.4805890227576974e-05, |
|
"loss": 0.632, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 3.8208955223880596, |
|
"grad_norm": 23.907413482666016, |
|
"learning_rate": 3.413654618473896e-05, |
|
"loss": 0.6742, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 3.9402985074626864, |
|
"grad_norm": 24.39226722717285, |
|
"learning_rate": 3.346720214190094e-05, |
|
"loss": 0.5644, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_accuracy": 0.8746031746031746, |
|
"eval_loss": 0.3922259509563446, |
|
"eval_runtime": 21.3995, |
|
"eval_samples_per_second": 88.32, |
|
"eval_steps_per_second": 2.804, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 4.059701492537314, |
|
"grad_norm": 18.170991897583008, |
|
"learning_rate": 3.279785809906292e-05, |
|
"loss": 0.5415, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 4.17910447761194, |
|
"grad_norm": 57.57135772705078, |
|
"learning_rate": 3.21285140562249e-05, |
|
"loss": 0.508, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 4.298507462686567, |
|
"grad_norm": 29.80289077758789, |
|
"learning_rate": 3.1459170013386885e-05, |
|
"loss": 0.5873, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 4.417910447761194, |
|
"grad_norm": 18.16728973388672, |
|
"learning_rate": 3.078982597054887e-05, |
|
"loss": 0.5387, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 4.537313432835821, |
|
"grad_norm": 21.491649627685547, |
|
"learning_rate": 3.012048192771085e-05, |
|
"loss": 0.5715, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 4.656716417910448, |
|
"grad_norm": 37.463436126708984, |
|
"learning_rate": 2.9451137884872827e-05, |
|
"loss": 0.5179, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 4.776119402985074, |
|
"grad_norm": 26.464433670043945, |
|
"learning_rate": 2.878179384203481e-05, |
|
"loss": 0.5064, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 4.895522388059701, |
|
"grad_norm": 20.331892013549805, |
|
"learning_rate": 2.8112449799196788e-05, |
|
"loss": 0.5672, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 4.991044776119403, |
|
"eval_accuracy": 0.8857142857142857, |
|
"eval_loss": 0.3783741593360901, |
|
"eval_runtime": 21.5183, |
|
"eval_samples_per_second": 87.832, |
|
"eval_steps_per_second": 2.788, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 5.014925373134329, |
|
"grad_norm": 16.20704460144043, |
|
"learning_rate": 2.7443105756358774e-05, |
|
"loss": 0.5492, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 5.134328358208955, |
|
"grad_norm": 24.48471450805664, |
|
"learning_rate": 2.6773761713520752e-05, |
|
"loss": 0.4528, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 5.253731343283582, |
|
"grad_norm": 27.444704055786133, |
|
"learning_rate": 2.6104417670682734e-05, |
|
"loss": 0.4718, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 5.373134328358209, |
|
"grad_norm": 19.921144485473633, |
|
"learning_rate": 2.5435073627844713e-05, |
|
"loss": 0.4575, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 5.492537313432836, |
|
"grad_norm": 24.932510375976562, |
|
"learning_rate": 2.4765729585006695e-05, |
|
"loss": 0.5138, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 5.611940298507463, |
|
"grad_norm": 29.121278762817383, |
|
"learning_rate": 2.4096385542168677e-05, |
|
"loss": 0.4647, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 5.731343283582089, |
|
"grad_norm": 21.379953384399414, |
|
"learning_rate": 2.3427041499330656e-05, |
|
"loss": 0.4648, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 5.850746268656716, |
|
"grad_norm": 23.85083770751953, |
|
"learning_rate": 2.2757697456492638e-05, |
|
"loss": 0.4935, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 5.970149253731344, |
|
"grad_norm": 17.832536697387695, |
|
"learning_rate": 2.208835341365462e-05, |
|
"loss": 0.454, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 5.994029850746268, |
|
"eval_accuracy": 0.8830687830687831, |
|
"eval_loss": 0.38558998703956604, |
|
"eval_runtime": 21.2656, |
|
"eval_samples_per_second": 88.876, |
|
"eval_steps_per_second": 2.821, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 6.08955223880597, |
|
"grad_norm": 33.72905349731445, |
|
"learning_rate": 2.1419009370816602e-05, |
|
"loss": 0.4187, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 6.208955223880597, |
|
"grad_norm": 20.838478088378906, |
|
"learning_rate": 2.074966532797858e-05, |
|
"loss": 0.4227, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 6.3283582089552235, |
|
"grad_norm": 21.023418426513672, |
|
"learning_rate": 2.0080321285140562e-05, |
|
"loss": 0.466, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 6.447761194029851, |
|
"grad_norm": 17.22841453552246, |
|
"learning_rate": 1.9410977242302544e-05, |
|
"loss": 0.4033, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 6.567164179104478, |
|
"grad_norm": 16.87044906616211, |
|
"learning_rate": 1.8741633199464527e-05, |
|
"loss": 0.4196, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 6.686567164179104, |
|
"grad_norm": 21.26498794555664, |
|
"learning_rate": 1.8072289156626505e-05, |
|
"loss": 0.4334, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 6.8059701492537314, |
|
"grad_norm": 18.514463424682617, |
|
"learning_rate": 1.7402945113788487e-05, |
|
"loss": 0.4471, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 6.925373134328359, |
|
"grad_norm": 21.403791427612305, |
|
"learning_rate": 1.673360107095047e-05, |
|
"loss": 0.4379, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 6.997014925373135, |
|
"eval_accuracy": 0.8888888888888888, |
|
"eval_loss": 0.35099852085113525, |
|
"eval_runtime": 21.3551, |
|
"eval_samples_per_second": 88.504, |
|
"eval_steps_per_second": 2.81, |
|
"step": 586 |
|
}, |
|
{ |
|
"epoch": 7.044776119402985, |
|
"grad_norm": 17.73862648010254, |
|
"learning_rate": 1.606425702811245e-05, |
|
"loss": 0.3597, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 7.164179104477612, |
|
"grad_norm": 21.782960891723633, |
|
"learning_rate": 1.5394912985274433e-05, |
|
"loss": 0.3952, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 7.2835820895522385, |
|
"grad_norm": 14.208507537841797, |
|
"learning_rate": 1.4725568942436414e-05, |
|
"loss": 0.3769, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 7.402985074626866, |
|
"grad_norm": 15.11253833770752, |
|
"learning_rate": 1.4056224899598394e-05, |
|
"loss": 0.3558, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 7.522388059701493, |
|
"grad_norm": 17.576494216918945, |
|
"learning_rate": 1.3386880856760376e-05, |
|
"loss": 0.3477, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 7.641791044776119, |
|
"grad_norm": 19.59687042236328, |
|
"learning_rate": 1.2717536813922356e-05, |
|
"loss": 0.3656, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 7.7611940298507465, |
|
"grad_norm": 23.697837829589844, |
|
"learning_rate": 1.2048192771084338e-05, |
|
"loss": 0.3668, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 7.880597014925373, |
|
"grad_norm": 13.360515594482422, |
|
"learning_rate": 1.1378848728246319e-05, |
|
"loss": 0.3576, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"grad_norm": 23.808975219726562, |
|
"learning_rate": 1.0709504685408301e-05, |
|
"loss": 0.3356, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_accuracy": 0.9063492063492063, |
|
"eval_loss": 0.31871286034584045, |
|
"eval_runtime": 21.467, |
|
"eval_samples_per_second": 88.042, |
|
"eval_steps_per_second": 2.795, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 8.119402985074627, |
|
"grad_norm": 14.76814079284668, |
|
"learning_rate": 1.0040160642570281e-05, |
|
"loss": 0.3394, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 8.238805970149254, |
|
"grad_norm": 15.642696380615234, |
|
"learning_rate": 9.370816599732263e-06, |
|
"loss": 0.2949, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 8.35820895522388, |
|
"grad_norm": 16.03794288635254, |
|
"learning_rate": 8.701472556894244e-06, |
|
"loss": 0.3072, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 8.477611940298507, |
|
"grad_norm": 14.671366691589355, |
|
"learning_rate": 8.032128514056226e-06, |
|
"loss": 0.3508, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 8.597014925373134, |
|
"grad_norm": 14.290385246276855, |
|
"learning_rate": 7.362784471218207e-06, |
|
"loss": 0.2763, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 8.716417910447761, |
|
"grad_norm": 17.65837287902832, |
|
"learning_rate": 6.693440428380188e-06, |
|
"loss": 0.3131, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 8.835820895522389, |
|
"grad_norm": 21.202238082885742, |
|
"learning_rate": 6.024096385542169e-06, |
|
"loss": 0.3269, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 8.955223880597014, |
|
"grad_norm": 13.900936126708984, |
|
"learning_rate": 5.3547523427041504e-06, |
|
"loss": 0.2877, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 8.991044776119402, |
|
"eval_accuracy": 0.9116402116402117, |
|
"eval_loss": 0.3208979070186615, |
|
"eval_runtime": 21.2786, |
|
"eval_samples_per_second": 88.822, |
|
"eval_steps_per_second": 2.82, |
|
"step": 753 |
|
}, |
|
{ |
|
"epoch": 9.074626865671641, |
|
"grad_norm": 28.610790252685547, |
|
"learning_rate": 4.685408299866132e-06, |
|
"loss": 0.2574, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 9.194029850746269, |
|
"grad_norm": 15.897197723388672, |
|
"learning_rate": 4.016064257028113e-06, |
|
"loss": 0.269, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 9.313432835820896, |
|
"grad_norm": 16.005224227905273, |
|
"learning_rate": 3.346720214190094e-06, |
|
"loss": 0.2783, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 9.432835820895523, |
|
"grad_norm": 12.884819984436035, |
|
"learning_rate": 2.6773761713520752e-06, |
|
"loss": 0.2813, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 9.552238805970148, |
|
"grad_norm": 15.747310638427734, |
|
"learning_rate": 2.0080321285140564e-06, |
|
"loss": 0.2777, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 9.671641791044776, |
|
"grad_norm": 13.63232421875, |
|
"learning_rate": 1.3386880856760376e-06, |
|
"loss": 0.2908, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 9.791044776119403, |
|
"grad_norm": 13.028614044189453, |
|
"learning_rate": 6.693440428380188e-07, |
|
"loss": 0.2894, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 9.91044776119403, |
|
"grad_norm": 12.740144729614258, |
|
"learning_rate": 0.0, |
|
"loss": 0.2717, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 9.91044776119403, |
|
"eval_accuracy": 0.9164021164021164, |
|
"eval_loss": 0.3027323782444, |
|
"eval_runtime": 22.4884, |
|
"eval_samples_per_second": 84.043, |
|
"eval_steps_per_second": 2.668, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 9.91044776119403, |
|
"step": 830, |
|
"total_flos": 1.0844463059730432e+19, |
|
"train_loss": 0.5978287294686558, |
|
"train_runtime": 2802.0121, |
|
"train_samples_per_second": 38.223, |
|
"train_steps_per_second": 0.296 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 830, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 10, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.0844463059730432e+19, |
|
"train_batch_size": 32, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|