joe611 commited on
Commit
b7acc33
1 Parent(s): 73b772f

Training in progress, epoch 149, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a47898b6387ef2cfdf4a2494cac6d643001aa350b662f1ca1cff6c6341846cae
3
  size 166496880
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e9790a957a17076b2fd3e81a46a61514d45e3fd83828e7c71e313b269d2d43c
3
  size 166496880
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9fd2c0a7bf22601834502a8f53f12a6d808f34646649b87e6eb1f54f7f0d7f05
3
  size 330495866
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bda7aaccfa295b45aedc9af5b9bfb50bc9bb8fdffc88c815357e96fd66ca4fd2
3
  size 330495866
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:622595e15d8d4adca4c438019feffee921460c3fc84414d13543954d20d79148
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a396222dff59b097bed0889ab42f3d144ca156561494a7c8362451cc383f6f2
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:426b4b133f941dbc1b63fbe118d222c9115fbe2fd29b895828a7fd399389ce5d
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d97cfcad85d9182f57a6d2bdf768ab9899061715a0c3f6d191db6d91a27ccfef
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.29718613624572754,
3
  "best_model_checkpoint": "chickens-composite-201616161616-150-epochs-wo-transform-metrics-test-shfld/checkpoint-49500",
4
- "epoch": 148.0,
5
  "eval_steps": 500,
6
- "global_step": 74000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -21117,6 +21117,151 @@
21117
  "eval_samples_per_second": 15.103,
21118
  "eval_steps_per_second": 1.963,
21119
  "step": 74000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
21120
  }
21121
  ],
21122
  "logging_steps": 30,
@@ -21136,7 +21281,7 @@
21136
  "attributes": {}
21137
  }
21138
  },
21139
- "total_flos": 2.54577103552512e+19,
21140
  "train_batch_size": 2,
21141
  "trial_name": null,
21142
  "trial_params": null
 
1
  {
2
  "best_metric": 0.29718613624572754,
3
  "best_model_checkpoint": "chickens-composite-201616161616-150-epochs-wo-transform-metrics-test-shfld/checkpoint-49500",
4
+ "epoch": 149.0,
5
  "eval_steps": 500,
6
+ "global_step": 74500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
21117
  "eval_samples_per_second": 15.103,
21118
  "eval_steps_per_second": 1.963,
21119
  "step": 74000
21120
+ },
21121
+ {
21122
+ "epoch": 148.02,
21123
+ "grad_norm": 43.992610931396484,
21124
+ "learning_rate": 4.298583608501328e-09,
21125
+ "loss": 0.2141,
21126
+ "step": 74010
21127
+ },
21128
+ {
21129
+ "epoch": 148.08,
21130
+ "grad_norm": 37.40137481689453,
21131
+ "learning_rate": 4.042045240927883e-09,
21132
+ "loss": 0.2215,
21133
+ "step": 74040
21134
+ },
21135
+ {
21136
+ "epoch": 148.14,
21137
+ "grad_norm": 64.64765930175781,
21138
+ "learning_rate": 3.793396172895314e-09,
21139
+ "loss": 0.2091,
21140
+ "step": 74070
21141
+ },
21142
+ {
21143
+ "epoch": 148.2,
21144
+ "grad_norm": 88.0441665649414,
21145
+ "learning_rate": 3.5526367970539765e-09,
21146
+ "loss": 0.204,
21147
+ "step": 74100
21148
+ },
21149
+ {
21150
+ "epoch": 148.26,
21151
+ "grad_norm": 53.6719856262207,
21152
+ "learning_rate": 3.31976749359586e-09,
21153
+ "loss": 0.1983,
21154
+ "step": 74130
21155
+ },
21156
+ {
21157
+ "epoch": 148.32,
21158
+ "grad_norm": 64.46839141845703,
21159
+ "learning_rate": 3.094788630254031e-09,
21160
+ "loss": 0.2306,
21161
+ "step": 74160
21162
+ },
21163
+ {
21164
+ "epoch": 148.38,
21165
+ "grad_norm": 34.077728271484375,
21166
+ "learning_rate": 2.8777005622998567e-09,
21167
+ "loss": 0.2127,
21168
+ "step": 74190
21169
+ },
21170
+ {
21171
+ "epoch": 148.44,
21172
+ "grad_norm": 71.48238372802734,
21173
+ "learning_rate": 2.6685036325457826e-09,
21174
+ "loss": 0.1978,
21175
+ "step": 74220
21176
+ },
21177
+ {
21178
+ "epoch": 148.5,
21179
+ "grad_norm": 103.37129974365234,
21180
+ "learning_rate": 2.4671981713420003e-09,
21181
+ "loss": 0.2089,
21182
+ "step": 74250
21183
+ },
21184
+ {
21185
+ "epoch": 148.56,
21186
+ "grad_norm": 31.161420822143555,
21187
+ "learning_rate": 2.2737844965775578e-09,
21188
+ "loss": 0.1992,
21189
+ "step": 74280
21190
+ },
21191
+ {
21192
+ "epoch": 148.62,
21193
+ "grad_norm": 60.29501724243164,
21194
+ "learning_rate": 2.088262913679251e-09,
21195
+ "loss": 0.1955,
21196
+ "step": 74310
21197
+ },
21198
+ {
21199
+ "epoch": 148.68,
21200
+ "grad_norm": 55.193145751953125,
21201
+ "learning_rate": 1.9106337156099553e-09,
21202
+ "loss": 0.2197,
21203
+ "step": 74340
21204
+ },
21205
+ {
21206
+ "epoch": 148.74,
21207
+ "grad_norm": 51.890506744384766,
21208
+ "learning_rate": 1.740897182871404e-09,
21209
+ "loss": 0.2325,
21210
+ "step": 74370
21211
+ },
21212
+ {
21213
+ "epoch": 148.8,
21214
+ "grad_norm": 81.51058197021484,
21215
+ "learning_rate": 1.5790535835003006e-09,
21216
+ "loss": 0.206,
21217
+ "step": 74400
21218
+ },
21219
+ {
21220
+ "epoch": 148.86,
21221
+ "grad_norm": 72.36231231689453,
21222
+ "learning_rate": 1.425103173069986e-09,
21223
+ "loss": 0.2108,
21224
+ "step": 74430
21225
+ },
21226
+ {
21227
+ "epoch": 148.92,
21228
+ "grad_norm": 67.69371795654297,
21229
+ "learning_rate": 1.2790461946887712e-09,
21230
+ "loss": 0.2046,
21231
+ "step": 74460
21232
+ },
21233
+ {
21234
+ "epoch": 148.98,
21235
+ "grad_norm": 54.43954086303711,
21236
+ "learning_rate": 1.1408828790010484e-09,
21237
+ "loss": 0.1978,
21238
+ "step": 74490
21239
+ },
21240
+ {
21241
+ "epoch": 149.0,
21242
+ "eval_loss": 0.30586719512939453,
21243
+ "eval_map": 0.8044,
21244
+ "eval_map_50": 0.9405,
21245
+ "eval_map_75": 0.9024,
21246
+ "eval_map_chicken": 0.7936,
21247
+ "eval_map_duck": 0.7475,
21248
+ "eval_map_large": 0.7843,
21249
+ "eval_map_medium": 0.8141,
21250
+ "eval_map_plant": 0.8722,
21251
+ "eval_map_small": 0.2979,
21252
+ "eval_mar_1": 0.3221,
21253
+ "eval_mar_10": 0.8382,
21254
+ "eval_mar_100": 0.8419,
21255
+ "eval_mar_100_chicken": 0.844,
21256
+ "eval_mar_100_duck": 0.7804,
21257
+ "eval_mar_100_plant": 0.9012,
21258
+ "eval_mar_large": 0.8145,
21259
+ "eval_mar_medium": 0.8546,
21260
+ "eval_mar_small": 0.3829,
21261
+ "eval_runtime": 7.0692,
21262
+ "eval_samples_per_second": 14.146,
21263
+ "eval_steps_per_second": 1.839,
21264
+ "step": 74500
21265
  }
21266
  ],
21267
  "logging_steps": 30,
 
21281
  "attributes": {}
21282
  }
21283
  },
21284
+ "total_flos": 2.56297219117056e+19,
21285
  "train_batch_size": 2,
21286
  "trial_name": null,
21287
  "trial_params": null