masatochi commited on
Commit
3ef69b7
1 Parent(s): ffad08f

Training in progress, step 200, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1e414d493c7b174cfc9f15e8c8f498a62290ef39418bff71646044ba53766d2f
3
  size 22573704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d78ac10e647d69703b938d7abcdf8f79189782ab487755938d17f4e0e6f36e0
3
  size 22573704
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:746384d1c74a736559be2e4296fa1ef73756a741e53afa3adbe66178b2abe282
3
  size 11710970
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ef026f68d94ee545f3aa1f5074b5b437c9f7b3ab3b920f537c479f981c3a67e
3
  size 11710970
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:333f2a71b3e34575844f8fd8a45254433fb8a5c1731e62ba395b530a464ea2d4
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b17ba4933d91121a2dbebb5ae4436f28fce7ee82ad22fbfd4c0dd47fb8c808e
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a35430a05f2b9748f37dd11667a782564c85a35d840d60cbaddfa2c905ab7c0a
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca9a25c72339c898b564e0c464a3f6fc75bbeec408008928b7ed05533156b98c
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.0953603520997616,
5
  "eval_steps": 34,
6
- "global_step": 195,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -1420,6 +1420,41 @@
1420
  "learning_rate": 4.2658237049655323e-07,
1421
  "loss": 0.9653,
1422
  "step": 195
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1423
  }
1424
  ],
1425
  "logging_steps": 1,
@@ -1434,12 +1469,12 @@
1434
  "should_evaluate": false,
1435
  "should_log": false,
1436
  "should_save": true,
1437
- "should_training_stop": false
1438
  },
1439
  "attributes": {}
1440
  }
1441
  },
1442
- "total_flos": 1.1257530122502144e+17,
1443
  "train_batch_size": 3,
1444
  "trial_name": null,
1445
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.09780548933308882,
5
  "eval_steps": 34,
6
+ "global_step": 200,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
1420
  "learning_rate": 4.2658237049655323e-07,
1421
  "loss": 0.9653,
1422
  "step": 195
1423
+ },
1424
+ {
1425
+ "epoch": 0.09584937954642704,
1426
+ "grad_norm": 1.8938212394714355,
1427
+ "learning_rate": 2.7308266142119785e-07,
1428
+ "loss": 1.4515,
1429
+ "step": 196
1430
+ },
1431
+ {
1432
+ "epoch": 0.09633840699309248,
1433
+ "grad_norm": 1.7014609575271606,
1434
+ "learning_rate": 1.5363960325660565e-07,
1435
+ "loss": 1.3232,
1436
+ "step": 197
1437
+ },
1438
+ {
1439
+ "epoch": 0.09682743443975793,
1440
+ "grad_norm": 1.6950466632843018,
1441
+ "learning_rate": 6.829398569770939e-08,
1442
+ "loss": 1.3605,
1443
+ "step": 198
1444
+ },
1445
+ {
1446
+ "epoch": 0.09731646188642337,
1447
+ "grad_norm": 1.5391230583190918,
1448
+ "learning_rate": 1.7074954194729044e-08,
1449
+ "loss": 1.155,
1450
+ "step": 199
1451
+ },
1452
+ {
1453
+ "epoch": 0.09780548933308882,
1454
+ "grad_norm": 1.8160505294799805,
1455
+ "learning_rate": 0.0,
1456
+ "loss": 1.3849,
1457
+ "step": 200
1458
  }
1459
  ],
1460
  "logging_steps": 1,
 
1469
  "should_evaluate": false,
1470
  "should_log": false,
1471
  "should_save": true,
1472
+ "should_training_stop": true
1473
  },
1474
  "attributes": {}
1475
  }
1476
  },
1477
+ "total_flos": 1.154618474102784e+17,
1478
  "train_batch_size": 3,
1479
  "trial_name": null,
1480
  "trial_params": null