masatochi commited on
Commit
a478881
1 Parent(s): 05162b7

Training in progress, step 45, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bc401528be5fc9de4bb49982a7dc2a42745c3f4f115ade208627dbca7071c7f8
3
  size 59827904
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b39424892d2b750e22eaa5c0216a3d5de439c4f797acbb4f9c456ba34c8a355e
3
  size 59827904
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f6bf2cb94ce5eba23af2a09df9e2a5daa15283bd4cbf0db5a31caaac8e36bb51
3
  size 30875540
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:802ed36b7afa5017c93a6fd3510a67856790006ca5398e785e74ab898f545ef4
3
  size 30875540
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4eb55320a08e50ed394cd960f98070f7af223236ab26cee8318395f23a36d7e2
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26d6b797df47639e9d56f9b6739dd1a2f36c0e6d770b9652f614b93b6b7b4046
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:35e2941b1419d36fedcd8eb55488740cb386508ea401393ade4c1f5fd25ff6c8
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8074483b54111bf60815c86124b01e5486a0c97c3c0bcc7642d6609b17e4381b
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.019561097866617763,
5
  "eval_steps": 34,
6
- "global_step": 40,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -303,6 +303,41 @@
303
  "learning_rate": 0.0001982973099683902,
304
  "loss": 9.481,
305
  "step": 40
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
306
  }
307
  ],
308
  "logging_steps": 1,
@@ -322,7 +357,7 @@
322
  "attributes": {}
323
  }
324
  },
325
- "total_flos": 8.817918703828992e+16,
326
  "train_batch_size": 3,
327
  "trial_name": null,
328
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.022006235099944985,
5
  "eval_steps": 34,
6
+ "global_step": 45,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
303
  "learning_rate": 0.0001982973099683902,
304
  "loss": 9.481,
305
  "step": 40
306
+ },
307
+ {
308
+ "epoch": 0.020050125313283207,
309
+ "grad_norm": Infinity,
310
+ "learning_rate": 0.0001979409767601366,
311
+ "loss": 9.684,
312
+ "step": 41
313
+ },
314
+ {
315
+ "epoch": 0.02053915275994865,
316
+ "grad_norm": 1.0845123650277868e+18,
317
+ "learning_rate": 0.00019755119679804367,
318
+ "loss": 9.2871,
319
+ "step": 42
320
+ },
321
+ {
322
+ "epoch": 0.021028180206614098,
323
+ "grad_norm": Infinity,
324
+ "learning_rate": 0.0001971281031916114,
325
+ "loss": 9.423,
326
+ "step": 43
327
+ },
328
+ {
329
+ "epoch": 0.02151720765327954,
330
+ "grad_norm": Infinity,
331
+ "learning_rate": 0.00019667184042691875,
332
+ "loss": 8.6827,
333
+ "step": 44
334
+ },
335
+ {
336
+ "epoch": 0.022006235099944985,
337
+ "grad_norm": Infinity,
338
+ "learning_rate": 0.00019618256431728194,
339
+ "loss": 9.9237,
340
+ "step": 45
341
  }
342
  ],
343
  "logging_steps": 1,
 
357
  "attributes": {}
358
  }
359
  },
360
+ "total_flos": 9.920158541807616e+16,
361
  "train_batch_size": 3,
362
  "trial_name": null,
363
  "trial_params": null