elsayedissa commited on
Commit
fefba33
1 Parent(s): 96fdeb1

Training in progress, step 5000

Browse files
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:071c5f4b0d6b94f41e4ec3d1d9e0d746d356fee79d390bbf5696fdec12d09eee
3
  size 12347192855
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae1044aaa1ff0119e85e5a0d94822e141e873a1f2c86ff72ffd19663c3603950
3
  size 12347192855
last-checkpoint/pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2d13c1515c5bec3b1b6df88233162654239b2f0a7c1119dddff788e8c045bd24
3
  size 6173655480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10329c118d1ff193245b323dfb24913485913d730509ac1e75a72918eceefd52
3
  size 6173655480
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ec70788b682d89359da85d225e858c6024d3fd2773541d2690bca2b322a42f06
3
  size 14575
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:51a6d2be8295de6c5e778332840e9e6dea5c8558aaf65c1f10a5f7fcbf6bc26c
3
  size 14575
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d92186bd8b7a36cbab743326563ee77d960ade57f12c3c59d4d213b20af4bd62
3
  size 557
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b99c6a597945a355f19ec7e8be26c1ddbe903987bb1e029b16374ec80b93d036
3
  size 557
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f8e131db79df8f2c3d12190a618e4f984ac87c57b6930e6ec08f7edadfe13646
3
  size 627
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:15ee7b0a16db565d640aec25ccf9707932c3a303f7fa345a261173fe1e99fbfc
3
  size 627
last-checkpoint/trainer_state.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.11536111854119403,
3
  "best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-spanish/checkpoint-1000",
4
- "epoch": 0.13884549967024193,
5
- "global_step": 4000,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -1002,11 +1002,260 @@
1002
  "eval_steps_per_second": 0.062,
1003
  "eval_wer": 0.11463792380367409,
1004
  "step": 4000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1005
  }
1006
  ],
1007
  "max_steps": 25000,
1008
  "num_train_epochs": 1,
1009
- "total_flos": 1.358829453312e+20,
1010
  "trial_name": null,
1011
  "trial_params": null
1012
  }
 
1
  {
2
  "best_metric": 0.11536111854119403,
3
  "best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-spanish/checkpoint-1000",
4
+ "epoch": 0.17355687458780242,
5
+ "global_step": 5000,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
1002
  "eval_steps_per_second": 0.062,
1003
  "eval_wer": 0.11463792380367409,
1004
  "step": 4000
1005
+ },
1006
+ {
1007
+ "epoch": 0.14,
1008
+ "learning_rate": 8.564081632653062e-06,
1009
+ "loss": 0.2078,
1010
+ "step": 4025
1011
+ },
1012
+ {
1013
+ "epoch": 0.14,
1014
+ "learning_rate": 8.55387755102041e-06,
1015
+ "loss": 0.1935,
1016
+ "step": 4050
1017
+ },
1018
+ {
1019
+ "epoch": 0.14,
1020
+ "learning_rate": 8.543673469387755e-06,
1021
+ "loss": 0.1966,
1022
+ "step": 4075
1023
+ },
1024
+ {
1025
+ "epoch": 0.14,
1026
+ "learning_rate": 8.533469387755102e-06,
1027
+ "loss": 0.2027,
1028
+ "step": 4100
1029
+ },
1030
+ {
1031
+ "epoch": 0.14,
1032
+ "learning_rate": 8.52326530612245e-06,
1033
+ "loss": 0.1985,
1034
+ "step": 4125
1035
+ },
1036
+ {
1037
+ "epoch": 0.14,
1038
+ "learning_rate": 8.513061224489797e-06,
1039
+ "loss": 0.1822,
1040
+ "step": 4150
1041
+ },
1042
+ {
1043
+ "epoch": 0.14,
1044
+ "learning_rate": 8.502857142857143e-06,
1045
+ "loss": 0.2102,
1046
+ "step": 4175
1047
+ },
1048
+ {
1049
+ "epoch": 0.15,
1050
+ "learning_rate": 8.49265306122449e-06,
1051
+ "loss": 0.2183,
1052
+ "step": 4200
1053
+ },
1054
+ {
1055
+ "epoch": 0.15,
1056
+ "learning_rate": 8.482448979591838e-06,
1057
+ "loss": 0.2299,
1058
+ "step": 4225
1059
+ },
1060
+ {
1061
+ "epoch": 0.15,
1062
+ "learning_rate": 8.472244897959184e-06,
1063
+ "loss": 0.1802,
1064
+ "step": 4250
1065
+ },
1066
+ {
1067
+ "epoch": 0.15,
1068
+ "learning_rate": 8.462040816326531e-06,
1069
+ "loss": 0.1931,
1070
+ "step": 4275
1071
+ },
1072
+ {
1073
+ "epoch": 0.15,
1074
+ "learning_rate": 8.451836734693877e-06,
1075
+ "loss": 0.1847,
1076
+ "step": 4300
1077
+ },
1078
+ {
1079
+ "epoch": 0.15,
1080
+ "learning_rate": 8.441632653061225e-06,
1081
+ "loss": 0.221,
1082
+ "step": 4325
1083
+ },
1084
+ {
1085
+ "epoch": 0.15,
1086
+ "learning_rate": 8.431428571428572e-06,
1087
+ "loss": 0.1894,
1088
+ "step": 4350
1089
+ },
1090
+ {
1091
+ "epoch": 0.15,
1092
+ "learning_rate": 8.42122448979592e-06,
1093
+ "loss": 0.1987,
1094
+ "step": 4375
1095
+ },
1096
+ {
1097
+ "epoch": 0.15,
1098
+ "learning_rate": 8.411020408163266e-06,
1099
+ "loss": 0.1714,
1100
+ "step": 4400
1101
+ },
1102
+ {
1103
+ "epoch": 0.15,
1104
+ "learning_rate": 8.400816326530613e-06,
1105
+ "loss": 0.2167,
1106
+ "step": 4425
1107
+ },
1108
+ {
1109
+ "epoch": 0.15,
1110
+ "learning_rate": 8.39061224489796e-06,
1111
+ "loss": 0.1895,
1112
+ "step": 4450
1113
+ },
1114
+ {
1115
+ "epoch": 0.16,
1116
+ "learning_rate": 8.380408163265306e-06,
1117
+ "loss": 0.1955,
1118
+ "step": 4475
1119
+ },
1120
+ {
1121
+ "epoch": 0.16,
1122
+ "learning_rate": 8.370204081632654e-06,
1123
+ "loss": 0.1764,
1124
+ "step": 4500
1125
+ },
1126
+ {
1127
+ "epoch": 0.16,
1128
+ "learning_rate": 8.36e-06,
1129
+ "loss": 0.2209,
1130
+ "step": 4525
1131
+ },
1132
+ {
1133
+ "epoch": 0.16,
1134
+ "learning_rate": 8.349795918367347e-06,
1135
+ "loss": 0.1681,
1136
+ "step": 4550
1137
+ },
1138
+ {
1139
+ "epoch": 0.16,
1140
+ "learning_rate": 8.339591836734695e-06,
1141
+ "loss": 0.218,
1142
+ "step": 4575
1143
+ },
1144
+ {
1145
+ "epoch": 0.16,
1146
+ "learning_rate": 8.329387755102042e-06,
1147
+ "loss": 0.1725,
1148
+ "step": 4600
1149
+ },
1150
+ {
1151
+ "epoch": 0.16,
1152
+ "learning_rate": 8.31918367346939e-06,
1153
+ "loss": 0.1768,
1154
+ "step": 4625
1155
+ },
1156
+ {
1157
+ "epoch": 0.16,
1158
+ "learning_rate": 8.308979591836735e-06,
1159
+ "loss": 0.1893,
1160
+ "step": 4650
1161
+ },
1162
+ {
1163
+ "epoch": 0.16,
1164
+ "learning_rate": 8.298775510204083e-06,
1165
+ "loss": 0.2104,
1166
+ "step": 4675
1167
+ },
1168
+ {
1169
+ "epoch": 0.16,
1170
+ "learning_rate": 8.288571428571429e-06,
1171
+ "loss": 0.1708,
1172
+ "step": 4700
1173
+ },
1174
+ {
1175
+ "epoch": 0.16,
1176
+ "learning_rate": 8.278367346938776e-06,
1177
+ "loss": 0.2091,
1178
+ "step": 4725
1179
+ },
1180
+ {
1181
+ "epoch": 0.16,
1182
+ "learning_rate": 8.268163265306124e-06,
1183
+ "loss": 0.2185,
1184
+ "step": 4750
1185
+ },
1186
+ {
1187
+ "epoch": 0.17,
1188
+ "learning_rate": 8.25795918367347e-06,
1189
+ "loss": 0.2086,
1190
+ "step": 4775
1191
+ },
1192
+ {
1193
+ "epoch": 0.17,
1194
+ "learning_rate": 8.247755102040817e-06,
1195
+ "loss": 0.1724,
1196
+ "step": 4800
1197
+ },
1198
+ {
1199
+ "epoch": 0.17,
1200
+ "learning_rate": 8.237551020408164e-06,
1201
+ "loss": 0.2082,
1202
+ "step": 4825
1203
+ },
1204
+ {
1205
+ "epoch": 0.17,
1206
+ "learning_rate": 8.227346938775512e-06,
1207
+ "loss": 0.1546,
1208
+ "step": 4850
1209
+ },
1210
+ {
1211
+ "epoch": 0.17,
1212
+ "learning_rate": 8.217142857142858e-06,
1213
+ "loss": 0.2014,
1214
+ "step": 4875
1215
+ },
1216
+ {
1217
+ "epoch": 0.17,
1218
+ "learning_rate": 8.206938775510205e-06,
1219
+ "loss": 0.1666,
1220
+ "step": 4900
1221
+ },
1222
+ {
1223
+ "epoch": 0.17,
1224
+ "learning_rate": 8.196734693877551e-06,
1225
+ "loss": 0.1793,
1226
+ "step": 4925
1227
+ },
1228
+ {
1229
+ "epoch": 0.17,
1230
+ "learning_rate": 8.186530612244898e-06,
1231
+ "loss": 0.1734,
1232
+ "step": 4950
1233
+ },
1234
+ {
1235
+ "epoch": 0.17,
1236
+ "learning_rate": 8.176326530612246e-06,
1237
+ "loss": 0.2033,
1238
+ "step": 4975
1239
+ },
1240
+ {
1241
+ "epoch": 0.17,
1242
+ "learning_rate": 8.166122448979592e-06,
1243
+ "loss": 0.1656,
1244
+ "step": 5000
1245
+ },
1246
+ {
1247
+ "epoch": 0.17,
1248
+ "eval_loss": 0.20018751919269562,
1249
+ "eval_runtime": 30831.2936,
1250
+ "eval_samples_per_second": 1.007,
1251
+ "eval_steps_per_second": 0.063,
1252
+ "eval_wer": 0.10732672221066432,
1253
+ "step": 5000
1254
  }
1255
  ],
1256
  "max_steps": 25000,
1257
  "num_train_epochs": 1,
1258
+ "total_flos": 1.69853681664e+20,
1259
  "trial_name": null,
1260
  "trial_params": null
1261
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2d13c1515c5bec3b1b6df88233162654239b2f0a7c1119dddff788e8c045bd24
3
  size 6173655480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10329c118d1ff193245b323dfb24913485913d730509ac1e75a72918eceefd52
3
  size 6173655480
runs/Dec26_19-57-19_gpu07.cyverse.org/events.out.tfevents.1672109970.gpu07.cyverse.org.126369.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3be876d39b9ffab345bd37299ad53f20801693f40e6411525ec37c8d2d55b4dd
3
- size 30781
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1de67b1e2d6be047867041a9be230f9e7c3ad6c837579ee7e86bfee1abf00cb1
3
+ size 37379