|
{ |
|
"best_metric": 0.7291744947433472, |
|
"best_model_checkpoint": "models_gitignored/bert-base-uncased-finetuned-sentence-classification/checkpoint-25252", |
|
"epoch": 3.0, |
|
"global_step": 75756, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.9920798352605737e-05, |
|
"loss": 1.2823, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.984159670521147e-05, |
|
"loss": 1.0188, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.9762395057817204e-05, |
|
"loss": 0.918, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.968319341042294e-05, |
|
"loss": 0.8996, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.9603991763028674e-05, |
|
"loss": 0.8202, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.9524790115634406e-05, |
|
"loss": 0.8602, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.944558846824014e-05, |
|
"loss": 0.8396, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.9366386820845877e-05, |
|
"loss": 0.8208, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.9287185173451612e-05, |
|
"loss": 0.8678, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.9207983526057344e-05, |
|
"loss": 0.7868, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.912878187866308e-05, |
|
"loss": 0.8267, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.904958023126881e-05, |
|
"loss": 0.7753, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.8970378583874546e-05, |
|
"loss": 0.7755, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.889117693648028e-05, |
|
"loss": 0.7856, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.8811975289086016e-05, |
|
"loss": 0.8204, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.8732773641691748e-05, |
|
"loss": 0.8075, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.8653571994297483e-05, |
|
"loss": 0.7782, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.8574370346903215e-05, |
|
"loss": 0.7993, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.849516869950895e-05, |
|
"loss": 0.785, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.8415967052114686e-05, |
|
"loss": 0.7602, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.833676540472042e-05, |
|
"loss": 0.7536, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.8257563757326153e-05, |
|
"loss": 0.763, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.8178362109931888e-05, |
|
"loss": 0.7935, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.8099160462537623e-05, |
|
"loss": 0.7566, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.8019958815143358e-05, |
|
"loss": 0.7797, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.794075716774909e-05, |
|
"loss": 0.7548, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 1.7861555520354825e-05, |
|
"loss": 0.7658, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 1.7782353872960557e-05, |
|
"loss": 0.7945, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 1.7703152225566292e-05, |
|
"loss": 0.779, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 1.7623950578172028e-05, |
|
"loss": 0.7521, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 1.7544748930777763e-05, |
|
"loss": 0.745, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 1.7465547283383495e-05, |
|
"loss": 0.7792, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 1.738634563598923e-05, |
|
"loss": 0.7402, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 1.730714398859496e-05, |
|
"loss": 0.7552, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 1.7227942341200697e-05, |
|
"loss": 0.7849, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.7148740693806432e-05, |
|
"loss": 0.7717, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.7069539046412167e-05, |
|
"loss": 0.7652, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.69903373990179e-05, |
|
"loss": 0.7521, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.6911135751623634e-05, |
|
"loss": 0.7684, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.683193410422937e-05, |
|
"loss": 0.7492, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.6752732456835105e-05, |
|
"loss": 0.7718, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.667353080944084e-05, |
|
"loss": 0.7287, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.6594329162046572e-05, |
|
"loss": 0.7685, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 1.6515127514652307e-05, |
|
"loss": 0.7597, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 1.643592586725804e-05, |
|
"loss": 0.7373, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.6356724219863774e-05, |
|
"loss": 0.7615, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.627752257246951e-05, |
|
"loss": 0.7356, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.6198320925075244e-05, |
|
"loss": 0.7738, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 1.6119119277680976e-05, |
|
"loss": 0.7424, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.603991763028671e-05, |
|
"loss": 0.728, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_accuracy": 0.7828767360509818, |
|
"eval_f1": 0.7791440516416176, |
|
"eval_kappa": 0.7055247162851201, |
|
"eval_loss": 0.7291744947433472, |
|
"eval_precision": 0.7811546650503057, |
|
"eval_recall": 0.7828767360509818, |
|
"eval_runtime": 304.9713, |
|
"eval_samples_per_second": 94.674, |
|
"eval_steps_per_second": 23.671, |
|
"step": 25252 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 1.5960715982892447e-05, |
|
"loss": 0.659, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 1.5881514335498182e-05, |
|
"loss": 0.5986, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 1.5802312688103914e-05, |
|
"loss": 0.6172, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 1.572311104070965e-05, |
|
"loss": 0.6544, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.564390939331538e-05, |
|
"loss": 0.6441, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.5564707745921116e-05, |
|
"loss": 0.604, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 1.548550609852685e-05, |
|
"loss": 0.5808, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 1.5406304451132586e-05, |
|
"loss": 0.6685, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 1.5327102803738318e-05, |
|
"loss": 0.6581, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 1.5247901156344053e-05, |
|
"loss": 0.6344, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 1.5168699508949787e-05, |
|
"loss": 0.6548, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 1.5089497861555522e-05, |
|
"loss": 0.6726, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 1.5010296214161256e-05, |
|
"loss": 0.6694, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 1.4931094566766991e-05, |
|
"loss": 0.6791, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 1.4851892919372723e-05, |
|
"loss": 0.6322, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 1.4772691271978458e-05, |
|
"loss": 0.648, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 1.4693489624584191e-05, |
|
"loss": 0.6643, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 1.4614287977189927e-05, |
|
"loss": 0.6545, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 1.453508632979566e-05, |
|
"loss": 0.5887, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 1.4455884682401395e-05, |
|
"loss": 0.6611, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 1.4376683035007129e-05, |
|
"loss": 0.6729, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 1.4297481387612864e-05, |
|
"loss": 0.631, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 1.4218279740218596e-05, |
|
"loss": 0.6706, |
|
"step": 36500 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 1.4139078092824333e-05, |
|
"loss": 0.6433, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 1.4059876445430065e-05, |
|
"loss": 0.5823, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 1.39806747980358e-05, |
|
"loss": 0.6519, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 1.3901473150641533e-05, |
|
"loss": 0.644, |
|
"step": 38500 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 1.3822271503247269e-05, |
|
"loss": 0.6669, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 1.3743069855853002e-05, |
|
"loss": 0.6936, |
|
"step": 39500 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 1.3663868208458737e-05, |
|
"loss": 0.6298, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 1.3584666561064473e-05, |
|
"loss": 0.7215, |
|
"step": 40500 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 1.3505464913670206e-05, |
|
"loss": 0.6555, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 1.3426263266275941e-05, |
|
"loss": 0.6789, |
|
"step": 41500 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 1.3347061618881673e-05, |
|
"loss": 0.6386, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 1.3267859971487408e-05, |
|
"loss": 0.593, |
|
"step": 42500 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 1.3188658324093142e-05, |
|
"loss": 0.6504, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 1.3109456676698877e-05, |
|
"loss": 0.6565, |
|
"step": 43500 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 1.303025502930461e-05, |
|
"loss": 0.6541, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 1.2951053381910346e-05, |
|
"loss": 0.6761, |
|
"step": 44500 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 1.287185173451608e-05, |
|
"loss": 0.658, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 1.2792650087121815e-05, |
|
"loss": 0.6487, |
|
"step": 45500 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 1.2713448439727546e-05, |
|
"loss": 0.6753, |
|
"step": 46000 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 1.2634246792333282e-05, |
|
"loss": 0.64, |
|
"step": 46500 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 1.2555045144939015e-05, |
|
"loss": 0.6226, |
|
"step": 47000 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 1.247584349754475e-05, |
|
"loss": 0.7033, |
|
"step": 47500 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 1.2396641850150484e-05, |
|
"loss": 0.6832, |
|
"step": 48000 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 1.2317440202756219e-05, |
|
"loss": 0.6237, |
|
"step": 48500 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 1.2238238555361953e-05, |
|
"loss": 0.6542, |
|
"step": 49000 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 1.2159036907967688e-05, |
|
"loss": 0.6677, |
|
"step": 49500 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 1.207983526057342e-05, |
|
"loss": 0.6236, |
|
"step": 50000 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 1.2000633613179155e-05, |
|
"loss": 0.6123, |
|
"step": 50500 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_accuracy": 0.781352820974613, |
|
"eval_f1": 0.779187798721889, |
|
"eval_kappa": 0.7064476335084646, |
|
"eval_loss": 0.8750618696212769, |
|
"eval_precision": 0.7867337948169639, |
|
"eval_recall": 0.781352820974613, |
|
"eval_runtime": 303.7391, |
|
"eval_samples_per_second": 95.059, |
|
"eval_steps_per_second": 23.767, |
|
"step": 50504 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 1.1921431965784888e-05, |
|
"loss": 0.453, |
|
"step": 51000 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 1.1842230318390624e-05, |
|
"loss": 0.5324, |
|
"step": 51500 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 1.1763028670996357e-05, |
|
"loss": 0.4745, |
|
"step": 52000 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 1.1683827023602092e-05, |
|
"loss": 0.5104, |
|
"step": 52500 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 1.1604625376207826e-05, |
|
"loss": 0.5623, |
|
"step": 53000 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 1.1525423728813561e-05, |
|
"loss": 0.4998, |
|
"step": 53500 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 1.1446222081419293e-05, |
|
"loss": 0.5157, |
|
"step": 54000 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 1.136702043402503e-05, |
|
"loss": 0.4614, |
|
"step": 54500 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 1.1287818786630762e-05, |
|
"loss": 0.5011, |
|
"step": 55000 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 1.1208617139236497e-05, |
|
"loss": 0.5621, |
|
"step": 55500 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 1.112941549184223e-05, |
|
"loss": 0.517, |
|
"step": 56000 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 1.1050213844447966e-05, |
|
"loss": 0.5152, |
|
"step": 56500 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 1.0971012197053699e-05, |
|
"loss": 0.5192, |
|
"step": 57000 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 1.0891810549659434e-05, |
|
"loss": 0.474, |
|
"step": 57500 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 1.0812608902265168e-05, |
|
"loss": 0.5443, |
|
"step": 58000 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 1.0733407254870903e-05, |
|
"loss": 0.4717, |
|
"step": 58500 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 1.0654205607476635e-05, |
|
"loss": 0.5105, |
|
"step": 59000 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 1.057500396008237e-05, |
|
"loss": 0.5299, |
|
"step": 59500 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 1.0495802312688104e-05, |
|
"loss": 0.5105, |
|
"step": 60000 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 1.0416600665293839e-05, |
|
"loss": 0.5919, |
|
"step": 60500 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 1.0337399017899574e-05, |
|
"loss": 0.5423, |
|
"step": 61000 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 1.0258197370505308e-05, |
|
"loss": 0.4779, |
|
"step": 61500 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 1.0178995723111043e-05, |
|
"loss": 0.5086, |
|
"step": 62000 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 1.0099794075716776e-05, |
|
"loss": 0.5192, |
|
"step": 62500 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 1.0020592428322511e-05, |
|
"loss": 0.5096, |
|
"step": 63000 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 9.941390780928243e-06, |
|
"loss": 0.5558, |
|
"step": 63500 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 9.862189133533979e-06, |
|
"loss": 0.5074, |
|
"step": 64000 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 9.782987486139712e-06, |
|
"loss": 0.5131, |
|
"step": 64500 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 9.703785838745447e-06, |
|
"loss": 0.5521, |
|
"step": 65000 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 9.62458419135118e-06, |
|
"loss": 0.4883, |
|
"step": 65500 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 9.545382543956914e-06, |
|
"loss": 0.5142, |
|
"step": 66000 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 9.46618089656265e-06, |
|
"loss": 0.5255, |
|
"step": 66500 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 9.386979249168383e-06, |
|
"loss": 0.5168, |
|
"step": 67000 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 9.307777601774117e-06, |
|
"loss": 0.5303, |
|
"step": 67500 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 9.228575954379852e-06, |
|
"loss": 0.4962, |
|
"step": 68000 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 9.149374306985585e-06, |
|
"loss": 0.531, |
|
"step": 68500 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 9.07017265959132e-06, |
|
"loss": 0.5689, |
|
"step": 69000 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 8.990971012197054e-06, |
|
"loss": 0.547, |
|
"step": 69500 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 8.911769364802788e-06, |
|
"loss": 0.5206, |
|
"step": 70000 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 8.832567717408523e-06, |
|
"loss": 0.5492, |
|
"step": 70500 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 8.753366070014258e-06, |
|
"loss": 0.5338, |
|
"step": 71000 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 8.674164422619991e-06, |
|
"loss": 0.5102, |
|
"step": 71500 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 8.594962775225727e-06, |
|
"loss": 0.5333, |
|
"step": 72000 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 8.51576112783146e-06, |
|
"loss": 0.4985, |
|
"step": 72500 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 8.436559480437194e-06, |
|
"loss": 0.4929, |
|
"step": 73000 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 8.357357833042929e-06, |
|
"loss": 0.4884, |
|
"step": 73500 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 8.278156185648662e-06, |
|
"loss": 0.5154, |
|
"step": 74000 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 8.198954538254396e-06, |
|
"loss": 0.5431, |
|
"step": 74500 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 8.119752890860131e-06, |
|
"loss": 0.4838, |
|
"step": 75000 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 8.040551243465865e-06, |
|
"loss": 0.4605, |
|
"step": 75500 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_accuracy": 0.7850587053648738, |
|
"eval_f1": 0.7838888247253561, |
|
"eval_kappa": 0.7119236186647375, |
|
"eval_loss": 0.9557045102119446, |
|
"eval_precision": 0.7851328181797262, |
|
"eval_recall": 0.7850587053648738, |
|
"eval_runtime": 304.9153, |
|
"eval_samples_per_second": 94.692, |
|
"eval_steps_per_second": 23.675, |
|
"step": 75756 |
|
} |
|
], |
|
"max_steps": 126260, |
|
"num_train_epochs": 5, |
|
"total_flos": 7.766798662066938e+16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|