w11wo commited on
Commit
0705270
1 Parent(s): 4728655

End of training

Browse files
README.md CHANGED
@@ -20,11 +20,11 @@ should probably proofread and complete it, then remove this comment. -->
20
 
21
  This model is a fine-tuned version of [indobenchmark/indobert-large-p1](https://huggingface.co/indobenchmark/indobert-large-p1) on an unknown dataset.
22
  It achieves the following results on the evaluation set:
23
- - Loss: 0.5739
24
- - Accuracy: 0.8619
25
- - F1: 0.7218
26
- - Precision: 0.7273
27
- - Recall: 0.7164
28
 
29
  ## Model description
30
 
 
20
 
21
  This model is a fine-tuned version of [indobenchmark/indobert-large-p1](https://huggingface.co/indobenchmark/indobert-large-p1) on an unknown dataset.
22
  It achieves the following results on the evaluation set:
23
+ - Loss: 0.3207
24
+ - Accuracy: 0.8643
25
+ - F1: 0.7160
26
+ - Precision: 0.7480
27
+ - Recall: 0.6866
28
 
29
  ## Model description
30
 
all_results.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 7.0,
3
+ "eval_accuracy": 0.8643122676579925,
4
+ "eval_f1": 0.7159533073929961,
5
+ "eval_loss": 0.32070210576057434,
6
+ "eval_precision": 0.7479674796747967,
7
+ "eval_recall": 0.6865671641791045,
8
+ "eval_runtime": 6.8805,
9
+ "eval_samples": 268,
10
+ "eval_samples_per_second": 78.193,
11
+ "eval_steps_per_second": 1.308,
12
+ "train_loss": 0.20290469026450095,
13
+ "train_runtime": 601.0116,
14
+ "train_samples": 1878,
15
+ "train_samples_per_second": 312.473,
16
+ "train_steps_per_second": 9.817
17
+ }
eval_results.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 7.0,
3
+ "eval_accuracy": 0.8643122676579925,
4
+ "eval_f1": 0.7159533073929961,
5
+ "eval_loss": 0.32070210576057434,
6
+ "eval_precision": 0.7479674796747967,
7
+ "eval_recall": 0.6865671641791045,
8
+ "eval_runtime": 6.8805,
9
+ "eval_samples": 268,
10
+ "eval_samples_per_second": 78.193,
11
+ "eval_steps_per_second": 1.308
12
+ }
predict_results.txt ADDED
@@ -0,0 +1,539 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ index prediction
2
+ 0 1
3
+ 1 0
4
+ 2 0
5
+ 3 0
6
+ 4 1
7
+ 5 0
8
+ 6 1
9
+ 7 1
10
+ 8 0
11
+ 9 0
12
+ 10 0
13
+ 11 0
14
+ 12 1
15
+ 13 0
16
+ 14 0
17
+ 15 1
18
+ 16 0
19
+ 17 0
20
+ 18 0
21
+ 19 0
22
+ 20 0
23
+ 21 0
24
+ 22 0
25
+ 23 1
26
+ 24 0
27
+ 25 0
28
+ 26 1
29
+ 27 1
30
+ 28 0
31
+ 29 0
32
+ 30 0
33
+ 31 0
34
+ 32 0
35
+ 33 1
36
+ 34 0
37
+ 35 0
38
+ 36 0
39
+ 37 0
40
+ 38 0
41
+ 39 0
42
+ 40 0
43
+ 41 0
44
+ 42 0
45
+ 43 0
46
+ 44 0
47
+ 45 0
48
+ 46 0
49
+ 47 0
50
+ 48 0
51
+ 49 0
52
+ 50 0
53
+ 51 0
54
+ 52 0
55
+ 53 0
56
+ 54 0
57
+ 55 0
58
+ 56 0
59
+ 57 0
60
+ 58 0
61
+ 59 0
62
+ 60 0
63
+ 61 0
64
+ 62 0
65
+ 63 0
66
+ 64 0
67
+ 65 0
68
+ 66 0
69
+ 67 0
70
+ 68 0
71
+ 69 0
72
+ 70 1
73
+ 71 1
74
+ 72 0
75
+ 73 1
76
+ 74 0
77
+ 75 1
78
+ 76 0
79
+ 77 0
80
+ 78 0
81
+ 79 0
82
+ 80 0
83
+ 81 0
84
+ 82 0
85
+ 83 1
86
+ 84 0
87
+ 85 0
88
+ 86 0
89
+ 87 0
90
+ 88 0
91
+ 89 0
92
+ 90 0
93
+ 91 0
94
+ 92 1
95
+ 93 0
96
+ 94 0
97
+ 95 0
98
+ 96 0
99
+ 97 0
100
+ 98 0
101
+ 99 0
102
+ 100 0
103
+ 101 0
104
+ 102 0
105
+ 103 0
106
+ 104 0
107
+ 105 0
108
+ 106 1
109
+ 107 0
110
+ 108 1
111
+ 109 0
112
+ 110 0
113
+ 111 0
114
+ 112 0
115
+ 113 0
116
+ 114 0
117
+ 115 0
118
+ 116 0
119
+ 117 1
120
+ 118 0
121
+ 119 1
122
+ 120 0
123
+ 121 0
124
+ 122 0
125
+ 123 0
126
+ 124 1
127
+ 125 0
128
+ 126 0
129
+ 127 0
130
+ 128 0
131
+ 129 0
132
+ 130 1
133
+ 131 0
134
+ 132 0
135
+ 133 0
136
+ 134 0
137
+ 135 0
138
+ 136 1
139
+ 137 0
140
+ 138 0
141
+ 139 0
142
+ 140 1
143
+ 141 0
144
+ 142 1
145
+ 143 0
146
+ 144 0
147
+ 145 0
148
+ 146 0
149
+ 147 1
150
+ 148 0
151
+ 149 0
152
+ 150 1
153
+ 151 0
154
+ 152 0
155
+ 153 0
156
+ 154 1
157
+ 155 0
158
+ 156 0
159
+ 157 0
160
+ 158 0
161
+ 159 0
162
+ 160 0
163
+ 161 0
164
+ 162 0
165
+ 163 0
166
+ 164 0
167
+ 165 0
168
+ 166 0
169
+ 167 0
170
+ 168 0
171
+ 169 0
172
+ 170 1
173
+ 171 1
174
+ 172 0
175
+ 173 0
176
+ 174 1
177
+ 175 0
178
+ 176 1
179
+ 177 0
180
+ 178 0
181
+ 179 0
182
+ 180 1
183
+ 181 0
184
+ 182 0
185
+ 183 0
186
+ 184 0
187
+ 185 1
188
+ 186 1
189
+ 187 0
190
+ 188 0
191
+ 189 0
192
+ 190 0
193
+ 191 0
194
+ 192 0
195
+ 193 0
196
+ 194 0
197
+ 195 0
198
+ 196 1
199
+ 197 0
200
+ 198 1
201
+ 199 0
202
+ 200 0
203
+ 201 0
204
+ 202 1
205
+ 203 1
206
+ 204 0
207
+ 205 1
208
+ 206 1
209
+ 207 0
210
+ 208 0
211
+ 209 0
212
+ 210 0
213
+ 211 0
214
+ 212 1
215
+ 213 1
216
+ 214 0
217
+ 215 0
218
+ 216 0
219
+ 217 0
220
+ 218 1
221
+ 219 0
222
+ 220 0
223
+ 221 0
224
+ 222 0
225
+ 223 1
226
+ 224 0
227
+ 225 0
228
+ 226 1
229
+ 227 0
230
+ 228 1
231
+ 229 0
232
+ 230 0
233
+ 231 1
234
+ 232 1
235
+ 233 0
236
+ 234 1
237
+ 235 0
238
+ 236 0
239
+ 237 1
240
+ 238 0
241
+ 239 0
242
+ 240 0
243
+ 241 0
244
+ 242 0
245
+ 243 0
246
+ 244 0
247
+ 245 0
248
+ 246 0
249
+ 247 0
250
+ 248 0
251
+ 249 1
252
+ 250 0
253
+ 251 0
254
+ 252 0
255
+ 253 1
256
+ 254 1
257
+ 255 0
258
+ 256 0
259
+ 257 1
260
+ 258 0
261
+ 259 1
262
+ 260 1
263
+ 261 0
264
+ 262 0
265
+ 263 0
266
+ 264 0
267
+ 265 0
268
+ 266 0
269
+ 267 0
270
+ 268 0
271
+ 269 0
272
+ 270 0
273
+ 271 1
274
+ 272 1
275
+ 273 0
276
+ 274 0
277
+ 275 0
278
+ 276 0
279
+ 277 0
280
+ 278 1
281
+ 279 0
282
+ 280 0
283
+ 281 0
284
+ 282 0
285
+ 283 0
286
+ 284 0
287
+ 285 1
288
+ 286 0
289
+ 287 0
290
+ 288 0
291
+ 289 0
292
+ 290 0
293
+ 291 0
294
+ 292 0
295
+ 293 0
296
+ 294 1
297
+ 295 0
298
+ 296 0
299
+ 297 1
300
+ 298 0
301
+ 299 0
302
+ 300 0
303
+ 301 1
304
+ 302 1
305
+ 303 1
306
+ 304 1
307
+ 305 0
308
+ 306 1
309
+ 307 0
310
+ 308 0
311
+ 309 0
312
+ 310 0
313
+ 311 0
314
+ 312 0
315
+ 313 0
316
+ 314 0
317
+ 315 0
318
+ 316 0
319
+ 317 0
320
+ 318 0
321
+ 319 0
322
+ 320 1
323
+ 321 1
324
+ 322 0
325
+ 323 0
326
+ 324 0
327
+ 325 0
328
+ 326 1
329
+ 327 0
330
+ 328 0
331
+ 329 0
332
+ 330 0
333
+ 331 1
334
+ 332 0
335
+ 333 0
336
+ 334 0
337
+ 335 0
338
+ 336 0
339
+ 337 0
340
+ 338 0
341
+ 339 0
342
+ 340 1
343
+ 341 1
344
+ 342 0
345
+ 343 0
346
+ 344 0
347
+ 345 0
348
+ 346 1
349
+ 347 0
350
+ 348 0
351
+ 349 0
352
+ 350 0
353
+ 351 0
354
+ 352 0
355
+ 353 1
356
+ 354 0
357
+ 355 0
358
+ 356 0
359
+ 357 0
360
+ 358 0
361
+ 359 1
362
+ 360 1
363
+ 361 0
364
+ 362 1
365
+ 363 1
366
+ 364 1
367
+ 365 0
368
+ 366 1
369
+ 367 0
370
+ 368 1
371
+ 369 0
372
+ 370 0
373
+ 371 1
374
+ 372 0
375
+ 373 1
376
+ 374 0
377
+ 375 0
378
+ 376 0
379
+ 377 0
380
+ 378 1
381
+ 379 0
382
+ 380 1
383
+ 381 0
384
+ 382 0
385
+ 383 0
386
+ 384 1
387
+ 385 0
388
+ 386 0
389
+ 387 0
390
+ 388 0
391
+ 389 0
392
+ 390 1
393
+ 391 0
394
+ 392 0
395
+ 393 0
396
+ 394 1
397
+ 395 0
398
+ 396 0
399
+ 397 0
400
+ 398 1
401
+ 399 0
402
+ 400 1
403
+ 401 0
404
+ 402 0
405
+ 403 1
406
+ 404 0
407
+ 405 1
408
+ 406 0
409
+ 407 1
410
+ 408 0
411
+ 409 0
412
+ 410 0
413
+ 411 0
414
+ 412 0
415
+ 413 0
416
+ 414 0
417
+ 415 0
418
+ 416 0
419
+ 417 1
420
+ 418 0
421
+ 419 0
422
+ 420 1
423
+ 421 0
424
+ 422 1
425
+ 423 1
426
+ 424 1
427
+ 425 0
428
+ 426 1
429
+ 427 0
430
+ 428 0
431
+ 429 0
432
+ 430 0
433
+ 431 0
434
+ 432 0
435
+ 433 0
436
+ 434 0
437
+ 435 0
438
+ 436 0
439
+ 437 0
440
+ 438 0
441
+ 439 0
442
+ 440 0
443
+ 441 1
444
+ 442 0
445
+ 443 0
446
+ 444 0
447
+ 445 0
448
+ 446 0
449
+ 447 0
450
+ 448 1
451
+ 449 1
452
+ 450 0
453
+ 451 0
454
+ 452 0
455
+ 453 0
456
+ 454 0
457
+ 455 0
458
+ 456 0
459
+ 457 0
460
+ 458 0
461
+ 459 0
462
+ 460 0
463
+ 461 0
464
+ 462 0
465
+ 463 1
466
+ 464 0
467
+ 465 0
468
+ 466 1
469
+ 467 1
470
+ 468 0
471
+ 469 0
472
+ 470 0
473
+ 471 1
474
+ 472 0
475
+ 473 0
476
+ 474 0
477
+ 475 1
478
+ 476 0
479
+ 477 0
480
+ 478 0
481
+ 479 0
482
+ 480 1
483
+ 481 0
484
+ 482 0
485
+ 483 0
486
+ 484 0
487
+ 485 0
488
+ 486 1
489
+ 487 1
490
+ 488 0
491
+ 489 0
492
+ 490 0
493
+ 491 0
494
+ 492 1
495
+ 493 1
496
+ 494 1
497
+ 495 1
498
+ 496 1
499
+ 497 0
500
+ 498 0
501
+ 499 0
502
+ 500 0
503
+ 501 0
504
+ 502 0
505
+ 503 0
506
+ 504 0
507
+ 505 0
508
+ 506 0
509
+ 507 0
510
+ 508 0
511
+ 509 0
512
+ 510 0
513
+ 511 0
514
+ 512 0
515
+ 513 0
516
+ 514 0
517
+ 515 0
518
+ 516 0
519
+ 517 1
520
+ 518 0
521
+ 519 0
522
+ 520 0
523
+ 521 0
524
+ 522 1
525
+ 523 0
526
+ 524 0
527
+ 525 0
528
+ 526 0
529
+ 527 0
530
+ 528 0
531
+ 529 0
532
+ 530 1
533
+ 531 0
534
+ 532 1
535
+ 533 0
536
+ 534 0
537
+ 535 0
538
+ 536 1
539
+ 537 1
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 7.0,
3
+ "train_loss": 0.20290469026450095,
4
+ "train_runtime": 601.0116,
5
+ "train_samples": 1878,
6
+ "train_samples_per_second": 312.473,
7
+ "train_steps_per_second": 9.817
8
+ }
trainer_state.json ADDED
@@ -0,0 +1,156 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.7328244274809161,
3
+ "best_model_checkpoint": "outputs/indobert-large-p1-twitter-indonesia-sarcastic/checkpoint-236",
4
+ "epoch": 7.0,
5
+ "eval_steps": 500,
6
+ "global_step": 413,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 1.0,
13
+ "learning_rate": 9.997777306367356e-06,
14
+ "loss": 0.5836,
15
+ "step": 59
16
+ },
17
+ {
18
+ "epoch": 1.0,
19
+ "eval_accuracy": 0.8059701492537313,
20
+ "eval_f1": 0.5737704918032788,
21
+ "eval_loss": 0.41532421112060547,
22
+ "eval_precision": 0.6363636363636364,
23
+ "eval_recall": 0.5223880597014925,
24
+ "eval_runtime": 3.3886,
25
+ "eval_samples_per_second": 79.088,
26
+ "eval_steps_per_second": 1.476,
27
+ "step": 59
28
+ },
29
+ {
30
+ "epoch": 2.0,
31
+ "learning_rate": 9.990628789297524e-06,
32
+ "loss": 0.3766,
33
+ "step": 118
34
+ },
35
+ {
36
+ "epoch": 2.0,
37
+ "eval_accuracy": 0.8432835820895522,
38
+ "eval_f1": 0.5961538461538461,
39
+ "eval_loss": 0.33532315492630005,
40
+ "eval_precision": 0.8378378378378378,
41
+ "eval_recall": 0.4626865671641791,
42
+ "eval_runtime": 3.3898,
43
+ "eval_samples_per_second": 79.062,
44
+ "eval_steps_per_second": 1.475,
45
+ "step": 118
46
+ },
47
+ {
48
+ "epoch": 3.0,
49
+ "learning_rate": 9.978800734821829e-06,
50
+ "loss": 0.2476,
51
+ "step": 177
52
+ },
53
+ {
54
+ "epoch": 3.0,
55
+ "eval_accuracy": 0.8619402985074627,
56
+ "eval_f1": 0.694214876033058,
57
+ "eval_loss": 0.3113671541213989,
58
+ "eval_precision": 0.7777777777777778,
59
+ "eval_recall": 0.6268656716417911,
60
+ "eval_runtime": 3.3876,
61
+ "eval_samples_per_second": 79.111,
62
+ "eval_steps_per_second": 1.476,
63
+ "step": 177
64
+ },
65
+ {
66
+ "epoch": 4.0,
67
+ "learning_rate": 9.961896985969137e-06,
68
+ "loss": 0.1356,
69
+ "step": 236
70
+ },
71
+ {
72
+ "epoch": 4.0,
73
+ "eval_accuracy": 0.8694029850746269,
74
+ "eval_f1": 0.7328244274809161,
75
+ "eval_loss": 0.327924907207489,
76
+ "eval_precision": 0.75,
77
+ "eval_recall": 0.7164179104477612,
78
+ "eval_runtime": 3.3876,
79
+ "eval_samples_per_second": 79.113,
80
+ "eval_steps_per_second": 1.476,
81
+ "step": 236
82
+ },
83
+ {
84
+ "epoch": 5.0,
85
+ "learning_rate": 9.940096443848375e-06,
86
+ "loss": 0.0536,
87
+ "step": 295
88
+ },
89
+ {
90
+ "epoch": 5.0,
91
+ "eval_accuracy": 0.8582089552238806,
92
+ "eval_f1": 0.7164179104477613,
93
+ "eval_loss": 0.42652806639671326,
94
+ "eval_precision": 0.7164179104477612,
95
+ "eval_recall": 0.7164179104477612,
96
+ "eval_runtime": 3.3976,
97
+ "eval_samples_per_second": 78.88,
98
+ "eval_steps_per_second": 1.472,
99
+ "step": 295
100
+ },
101
+ {
102
+ "epoch": 6.0,
103
+ "learning_rate": 9.913420622962606e-06,
104
+ "loss": 0.0157,
105
+ "step": 354
106
+ },
107
+ {
108
+ "epoch": 6.0,
109
+ "eval_accuracy": 0.8619402985074627,
110
+ "eval_f1": 0.6666666666666667,
111
+ "eval_loss": 0.644801914691925,
112
+ "eval_precision": 0.8409090909090909,
113
+ "eval_recall": 0.5522388059701493,
114
+ "eval_runtime": 3.3901,
115
+ "eval_samples_per_second": 79.054,
116
+ "eval_steps_per_second": 1.475,
117
+ "step": 354
118
+ },
119
+ {
120
+ "epoch": 7.0,
121
+ "learning_rate": 9.881895849126432e-06,
122
+ "loss": 0.0076,
123
+ "step": 413
124
+ },
125
+ {
126
+ "epoch": 7.0,
127
+ "eval_accuracy": 0.8619402985074627,
128
+ "eval_f1": 0.7218045112781953,
129
+ "eval_loss": 0.5739138722419739,
130
+ "eval_precision": 0.7272727272727273,
131
+ "eval_recall": 0.7164179104477612,
132
+ "eval_runtime": 3.3841,
133
+ "eval_samples_per_second": 79.193,
134
+ "eval_steps_per_second": 1.477,
135
+ "step": 413
136
+ },
137
+ {
138
+ "epoch": 7.0,
139
+ "step": 413,
140
+ "total_flos": 3062792425577472.0,
141
+ "train_loss": 0.20290469026450095,
142
+ "train_runtime": 601.0116,
143
+ "train_samples_per_second": 312.473,
144
+ "train_steps_per_second": 9.817
145
+ }
146
+ ],
147
+ "logging_steps": 500,
148
+ "max_steps": 5900,
149
+ "num_input_tokens_seen": 0,
150
+ "num_train_epochs": 100,
151
+ "save_steps": 500,
152
+ "total_flos": 3062792425577472.0,
153
+ "train_batch_size": 32,
154
+ "trial_name": null,
155
+ "trial_params": null
156
+ }