Training in progress, epoch 149, checkpoint
Browse files
last-checkpoint/model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 166496880
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7e9790a957a17076b2fd3e81a46a61514d45e3fd83828e7c71e313b269d2d43c
|
3 |
size 166496880
|
last-checkpoint/optimizer.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 330495866
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bda7aaccfa295b45aedc9af5b9bfb50bc9bb8fdffc88c815357e96fd66ca4fd2
|
3 |
size 330495866
|
last-checkpoint/rng_state.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14244
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7a396222dff59b097bed0889ab42f3d144ca156561494a7c8362451cc383f6f2
|
3 |
size 14244
|
last-checkpoint/scheduler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1064
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d97cfcad85d9182f57a6d2bdf768ab9899061715a0c3f6d191db6d91a27ccfef
|
3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
{
|
2 |
"best_metric": 0.29718613624572754,
|
3 |
"best_model_checkpoint": "chickens-composite-201616161616-150-epochs-wo-transform-metrics-test-shfld/checkpoint-49500",
|
4 |
-
"epoch":
|
5 |
"eval_steps": 500,
|
6 |
-
"global_step":
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
@@ -21117,6 +21117,151 @@
|
|
21117 |
"eval_samples_per_second": 15.103,
|
21118 |
"eval_steps_per_second": 1.963,
|
21119 |
"step": 74000
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
21120 |
}
|
21121 |
],
|
21122 |
"logging_steps": 30,
|
@@ -21136,7 +21281,7 @@
|
|
21136 |
"attributes": {}
|
21137 |
}
|
21138 |
},
|
21139 |
-
"total_flos": 2.
|
21140 |
"train_batch_size": 2,
|
21141 |
"trial_name": null,
|
21142 |
"trial_params": null
|
|
|
1 |
{
|
2 |
"best_metric": 0.29718613624572754,
|
3 |
"best_model_checkpoint": "chickens-composite-201616161616-150-epochs-wo-transform-metrics-test-shfld/checkpoint-49500",
|
4 |
+
"epoch": 149.0,
|
5 |
"eval_steps": 500,
|
6 |
+
"global_step": 74500,
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
|
|
21117 |
"eval_samples_per_second": 15.103,
|
21118 |
"eval_steps_per_second": 1.963,
|
21119 |
"step": 74000
|
21120 |
+
},
|
21121 |
+
{
|
21122 |
+
"epoch": 148.02,
|
21123 |
+
"grad_norm": 43.992610931396484,
|
21124 |
+
"learning_rate": 4.298583608501328e-09,
|
21125 |
+
"loss": 0.2141,
|
21126 |
+
"step": 74010
|
21127 |
+
},
|
21128 |
+
{
|
21129 |
+
"epoch": 148.08,
|
21130 |
+
"grad_norm": 37.40137481689453,
|
21131 |
+
"learning_rate": 4.042045240927883e-09,
|
21132 |
+
"loss": 0.2215,
|
21133 |
+
"step": 74040
|
21134 |
+
},
|
21135 |
+
{
|
21136 |
+
"epoch": 148.14,
|
21137 |
+
"grad_norm": 64.64765930175781,
|
21138 |
+
"learning_rate": 3.793396172895314e-09,
|
21139 |
+
"loss": 0.2091,
|
21140 |
+
"step": 74070
|
21141 |
+
},
|
21142 |
+
{
|
21143 |
+
"epoch": 148.2,
|
21144 |
+
"grad_norm": 88.0441665649414,
|
21145 |
+
"learning_rate": 3.5526367970539765e-09,
|
21146 |
+
"loss": 0.204,
|
21147 |
+
"step": 74100
|
21148 |
+
},
|
21149 |
+
{
|
21150 |
+
"epoch": 148.26,
|
21151 |
+
"grad_norm": 53.6719856262207,
|
21152 |
+
"learning_rate": 3.31976749359586e-09,
|
21153 |
+
"loss": 0.1983,
|
21154 |
+
"step": 74130
|
21155 |
+
},
|
21156 |
+
{
|
21157 |
+
"epoch": 148.32,
|
21158 |
+
"grad_norm": 64.46839141845703,
|
21159 |
+
"learning_rate": 3.094788630254031e-09,
|
21160 |
+
"loss": 0.2306,
|
21161 |
+
"step": 74160
|
21162 |
+
},
|
21163 |
+
{
|
21164 |
+
"epoch": 148.38,
|
21165 |
+
"grad_norm": 34.077728271484375,
|
21166 |
+
"learning_rate": 2.8777005622998567e-09,
|
21167 |
+
"loss": 0.2127,
|
21168 |
+
"step": 74190
|
21169 |
+
},
|
21170 |
+
{
|
21171 |
+
"epoch": 148.44,
|
21172 |
+
"grad_norm": 71.48238372802734,
|
21173 |
+
"learning_rate": 2.6685036325457826e-09,
|
21174 |
+
"loss": 0.1978,
|
21175 |
+
"step": 74220
|
21176 |
+
},
|
21177 |
+
{
|
21178 |
+
"epoch": 148.5,
|
21179 |
+
"grad_norm": 103.37129974365234,
|
21180 |
+
"learning_rate": 2.4671981713420003e-09,
|
21181 |
+
"loss": 0.2089,
|
21182 |
+
"step": 74250
|
21183 |
+
},
|
21184 |
+
{
|
21185 |
+
"epoch": 148.56,
|
21186 |
+
"grad_norm": 31.161420822143555,
|
21187 |
+
"learning_rate": 2.2737844965775578e-09,
|
21188 |
+
"loss": 0.1992,
|
21189 |
+
"step": 74280
|
21190 |
+
},
|
21191 |
+
{
|
21192 |
+
"epoch": 148.62,
|
21193 |
+
"grad_norm": 60.29501724243164,
|
21194 |
+
"learning_rate": 2.088262913679251e-09,
|
21195 |
+
"loss": 0.1955,
|
21196 |
+
"step": 74310
|
21197 |
+
},
|
21198 |
+
{
|
21199 |
+
"epoch": 148.68,
|
21200 |
+
"grad_norm": 55.193145751953125,
|
21201 |
+
"learning_rate": 1.9106337156099553e-09,
|
21202 |
+
"loss": 0.2197,
|
21203 |
+
"step": 74340
|
21204 |
+
},
|
21205 |
+
{
|
21206 |
+
"epoch": 148.74,
|
21207 |
+
"grad_norm": 51.890506744384766,
|
21208 |
+
"learning_rate": 1.740897182871404e-09,
|
21209 |
+
"loss": 0.2325,
|
21210 |
+
"step": 74370
|
21211 |
+
},
|
21212 |
+
{
|
21213 |
+
"epoch": 148.8,
|
21214 |
+
"grad_norm": 81.51058197021484,
|
21215 |
+
"learning_rate": 1.5790535835003006e-09,
|
21216 |
+
"loss": 0.206,
|
21217 |
+
"step": 74400
|
21218 |
+
},
|
21219 |
+
{
|
21220 |
+
"epoch": 148.86,
|
21221 |
+
"grad_norm": 72.36231231689453,
|
21222 |
+
"learning_rate": 1.425103173069986e-09,
|
21223 |
+
"loss": 0.2108,
|
21224 |
+
"step": 74430
|
21225 |
+
},
|
21226 |
+
{
|
21227 |
+
"epoch": 148.92,
|
21228 |
+
"grad_norm": 67.69371795654297,
|
21229 |
+
"learning_rate": 1.2790461946887712e-09,
|
21230 |
+
"loss": 0.2046,
|
21231 |
+
"step": 74460
|
21232 |
+
},
|
21233 |
+
{
|
21234 |
+
"epoch": 148.98,
|
21235 |
+
"grad_norm": 54.43954086303711,
|
21236 |
+
"learning_rate": 1.1408828790010484e-09,
|
21237 |
+
"loss": 0.1978,
|
21238 |
+
"step": 74490
|
21239 |
+
},
|
21240 |
+
{
|
21241 |
+
"epoch": 149.0,
|
21242 |
+
"eval_loss": 0.30586719512939453,
|
21243 |
+
"eval_map": 0.8044,
|
21244 |
+
"eval_map_50": 0.9405,
|
21245 |
+
"eval_map_75": 0.9024,
|
21246 |
+
"eval_map_chicken": 0.7936,
|
21247 |
+
"eval_map_duck": 0.7475,
|
21248 |
+
"eval_map_large": 0.7843,
|
21249 |
+
"eval_map_medium": 0.8141,
|
21250 |
+
"eval_map_plant": 0.8722,
|
21251 |
+
"eval_map_small": 0.2979,
|
21252 |
+
"eval_mar_1": 0.3221,
|
21253 |
+
"eval_mar_10": 0.8382,
|
21254 |
+
"eval_mar_100": 0.8419,
|
21255 |
+
"eval_mar_100_chicken": 0.844,
|
21256 |
+
"eval_mar_100_duck": 0.7804,
|
21257 |
+
"eval_mar_100_plant": 0.9012,
|
21258 |
+
"eval_mar_large": 0.8145,
|
21259 |
+
"eval_mar_medium": 0.8546,
|
21260 |
+
"eval_mar_small": 0.3829,
|
21261 |
+
"eval_runtime": 7.0692,
|
21262 |
+
"eval_samples_per_second": 14.146,
|
21263 |
+
"eval_steps_per_second": 1.839,
|
21264 |
+
"step": 74500
|
21265 |
}
|
21266 |
],
|
21267 |
"logging_steps": 30,
|
|
|
21281 |
"attributes": {}
|
21282 |
}
|
21283 |
},
|
21284 |
+
"total_flos": 2.56297219117056e+19,
|
21285 |
"train_batch_size": 2,
|
21286 |
"trial_name": null,
|
21287 |
"trial_params": null
|