masatochi commited on
Commit
ebf81dc
1 Parent(s): 4a58de6

Training in progress, step 40, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ff0260d1e3631bfaf3aad882c37dfe986418a9da16b6a4e18e4aa30e93aa2502
3
  size 59827904
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc401528be5fc9de4bb49982a7dc2a42745c3f4f115ade208627dbca7071c7f8
3
  size 59827904
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5661af4bb46bd4690466f1c6547cb2605855d735bc15f0f625b75eddd477b21b
3
  size 30875540
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f6bf2cb94ce5eba23af2a09df9e2a5daa15283bd4cbf0db5a31caaac8e36bb51
3
  size 30875540
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0c9d6305112455934908c7bfd9606d4794d06684c6273dfde936a82028d5a005
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4eb55320a08e50ed394cd960f98070f7af223236ab26cee8318395f23a36d7e2
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:666e3907648b0318630be4f0e9270d7d902986930986134493537abd333e959c
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:35e2941b1419d36fedcd8eb55488740cb386508ea401393ade4c1f5fd25ff6c8
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.017115960633290545,
5
  "eval_steps": 34,
6
- "global_step": 35,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -268,6 +268,41 @@
268
  "learning_rate": 0.00019957341762950344,
269
  "loss": 10.3859,
270
  "step": 35
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
271
  }
272
  ],
273
  "logging_steps": 1,
@@ -287,7 +322,7 @@
287
  "attributes": {}
288
  }
289
  },
290
- "total_flos": 7.715678865850368e+16,
291
  "train_batch_size": 3,
292
  "trial_name": null,
293
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.019561097866617763,
5
  "eval_steps": 34,
6
+ "global_step": 40,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
268
  "learning_rate": 0.00019957341762950344,
269
  "loss": 10.3859,
270
  "step": 35
271
+ },
272
+ {
273
+ "epoch": 0.017604988079955988,
274
+ "grad_norm": Infinity,
275
+ "learning_rate": 0.0001993859136895274,
276
+ "loss": 8.1513,
277
+ "step": 36
278
+ },
279
+ {
280
+ "epoch": 0.018094015526621432,
281
+ "grad_norm": 7.999064190058758e+18,
282
+ "learning_rate": 0.00019916446955107428,
283
+ "loss": 9.6598,
284
+ "step": 37
285
+ },
286
+ {
287
+ "epoch": 0.018583042973286876,
288
+ "grad_norm": 2.2595998866116444e+18,
289
+ "learning_rate": 0.0001989091608371146,
290
+ "loss": 8.1904,
291
+ "step": 38
292
+ },
293
+ {
294
+ "epoch": 0.01907207041995232,
295
+ "grad_norm": Infinity,
296
+ "learning_rate": 0.00019862007473534025,
297
+ "loss": 9.4891,
298
+ "step": 39
299
+ },
300
+ {
301
+ "epoch": 0.019561097866617763,
302
+ "grad_norm": Infinity,
303
+ "learning_rate": 0.0001982973099683902,
304
+ "loss": 9.481,
305
+ "step": 40
306
  }
307
  ],
308
  "logging_steps": 1,
 
322
  "attributes": {}
323
  }
324
  },
325
+ "total_flos": 8.817918703828992e+16,
326
  "train_batch_size": 3,
327
  "trial_name": null,
328
  "trial_params": null