masatochi commited on
Commit
85b242e
1 Parent(s): a1846d5

Training in progress, step 100, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1c0eeae5d0bad680cc137a8478659f61e169009131b874e94497482c82943a5b
3
  size 48679352
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:34410a29025509af716b74f8f8e5e594616ec96042ec3c844e69738ed53ee00d
3
  size 48679352
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9f1cd2840a4f0b9b6f6c26974d02c6b354c89f90c34c21d56ba51ebe0799f308
3
  size 25152500
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0bfc7fa8494a7d0b19676a339c8110ea263a3bc6118c004bbafd199930580826
3
  size 25152500
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f0c3752ce98bfc3ec197e65d611692cfd440b987d3ddcbb16981cae0698d7e0b
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a039ad76f0746f8ff052e22511b41b8352036173c7caa7a0153574149fda9e0
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3160dd3097641f3bf4d4036c0ddfd8673184925120f088b5ebecc6a1e5c953dd
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:07b8a5b9aaab52c529e8225a8efdb5ee4b1c103decfe1101828eac45305d759c
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.048460116049225274,
5
  "eval_steps": 34,
6
- "global_step": 95,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -696,6 +696,41 @@
696
  "learning_rate": 0.00013612416661871533,
697
  "loss": 0.4754,
698
  "step": 95
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
699
  }
700
  ],
701
  "logging_steps": 1,
@@ -715,7 +750,7 @@
715
  "attributes": {}
716
  }
717
  },
718
- "total_flos": 1.586248544157696e+17,
719
  "train_batch_size": 3,
720
  "trial_name": null,
721
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.05101064847286871,
5
  "eval_steps": 34,
6
+ "global_step": 100,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
696
  "learning_rate": 0.00013612416661871533,
697
  "loss": 0.4754,
698
  "step": 95
699
+ },
700
+ {
701
+ "epoch": 0.048970222533953965,
702
+ "grad_norm": 0.3616621494293213,
703
+ "learning_rate": 0.00013439489186339282,
704
+ "loss": 0.6012,
705
+ "step": 96
706
+ },
707
+ {
708
+ "epoch": 0.04948032901868265,
709
+ "grad_norm": 0.44421911239624023,
710
+ "learning_rate": 0.0001326538712840083,
711
+ "loss": 0.4534,
712
+ "step": 97
713
+ },
714
+ {
715
+ "epoch": 0.04999043550341134,
716
+ "grad_norm": 0.35028621554374695,
717
+ "learning_rate": 0.00013090169943749476,
718
+ "loss": 0.5802,
719
+ "step": 98
720
+ },
721
+ {
722
+ "epoch": 0.050500541988140024,
723
+ "grad_norm": 0.38989895582199097,
724
+ "learning_rate": 0.00012913897468893248,
725
+ "loss": 0.5838,
726
+ "step": 99
727
+ },
728
+ {
729
+ "epoch": 0.05101064847286871,
730
+ "grad_norm": 0.4552973210811615,
731
+ "learning_rate": 0.0001273662990072083,
732
+ "loss": 0.7394,
733
+ "step": 100
734
  }
735
  ],
736
  "logging_steps": 1,
 
750
  "attributes": {}
751
  }
752
  },
753
+ "total_flos": 1.66973530963968e+17,
754
  "train_batch_size": 3,
755
  "trial_name": null,
756
  "trial_params": null