masatochi commited on
Commit
5e01acc
1 Parent(s): a2b7ed8

Training in progress, step 130, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1080c2dcf21c0b41e7ea16ba47e965a1a0bc3cba9fab02708c6b5bd830c931b2
3
  size 22573704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:81c631790391462074f7254948e2937b8d20f618f4411a2eb2bd8919e480a821
3
  size 22573704
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8e8cc1a90bf0d92d37ea0e0431dd8f7cb23cc35f1625edfceb841ebbccf62d51
3
  size 11710970
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ecbe9f1bd7b567eef8ff0b47d85ebde84bef5328a324b5eb82318b5f61dffdd4
3
  size 11710970
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:86748f5994a634f58ece1ba5cddd089ee5a870e039bbcbeb2cee4742ee257abe
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ebcfbe921966314e6f2c12a21ae76e083d0bfa3c62fc82a12af734431287d92
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b577939e8ae09a93269bdd1ffbcc4ef41ec4027476aa914ab19034c5a6ebf492
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb82cfc852da5eb6970b061692451307a303bfd2a3160c6d6a29266f8bb6adef
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.061128430833180515,
5
  "eval_steps": 34,
6
- "global_step": 125,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -914,6 +914,41 @@
914
  "learning_rate": 8.162504821834295e-05,
915
  "loss": 1.2018,
916
  "step": 125
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
917
  }
918
  ],
919
  "logging_steps": 1,
@@ -933,7 +968,7 @@
933
  "attributes": {}
934
  }
935
  },
936
- "total_flos": 7.2163654631424e+16,
937
  "train_batch_size": 3,
938
  "trial_name": null,
939
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.06357356806650773,
5
  "eval_steps": 34,
6
+ "global_step": 130,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
914
  "learning_rate": 8.162504821834295e-05,
915
  "loss": 1.2018,
916
  "step": 125
917
+ },
918
+ {
919
+ "epoch": 0.06161745827984596,
920
+ "grad_norm": 1.4820826053619385,
921
+ "learning_rate": 7.9811759084299e-05,
922
+ "loss": 1.102,
923
+ "step": 126
924
+ },
925
+ {
926
+ "epoch": 0.0621064857265114,
927
+ "grad_norm": 1.699189305305481,
928
+ "learning_rate": 7.800536421603317e-05,
929
+ "loss": 1.161,
930
+ "step": 127
931
+ },
932
+ {
933
+ "epoch": 0.06259551317317684,
934
+ "grad_norm": 1.9508938789367676,
935
+ "learning_rate": 7.620648049573815e-05,
936
+ "loss": 1.1853,
937
+ "step": 128
938
+ },
939
+ {
940
+ "epoch": 0.06308454061984228,
941
+ "grad_norm": 1.7746766805648804,
942
+ "learning_rate": 7.441572224055644e-05,
943
+ "loss": 1.1611,
944
+ "step": 129
945
+ },
946
+ {
947
+ "epoch": 0.06357356806650773,
948
+ "grad_norm": 1.4383032321929932,
949
+ "learning_rate": 7.263370099279172e-05,
950
+ "loss": 1.0862,
951
+ "step": 130
952
  }
953
  ],
954
  "logging_steps": 1,
 
968
  "attributes": {}
969
  }
970
  },
971
+ "total_flos": 7.505020081668096e+16,
972
  "train_batch_size": 3,
973
  "trial_name": null,
974
  "trial_params": null