Jingmei commited on
Commit
5923c8c
1 Parent(s): a5b5771

Training in progress, step 10

Browse files
adapter_config.json CHANGED
@@ -20,8 +20,8 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "v_proj",
24
- "q_proj"
25
  ],
26
  "task_type": "CAUSAL_LM",
27
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "q_proj",
24
+ "v_proj"
25
  ],
26
  "task_type": "CAUSAL_LM",
27
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bcf53e97cd04631eb055583e3a68ea8399fcd1d1baab28a3c28bea2034d0ae72
3
  size 16794200
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc02e3853cde0b9c4efd88d7e0ca3794dd681f76a2ad0cb8a966972ab1629c6b
3
  size 16794200
trainer_peft.log CHANGED
@@ -1,50 +1,20 @@
1
- 2024-05-31 16:32 - Cuda check
2
- 2024-05-31 16:32 - True
3
- 2024-05-31 16:32 - 3
4
- 2024-05-31 16:32 - Configue Model and tokenizer
5
- 2024-05-31 16:32 - Cuda check
6
- 2024-05-31 16:32 - True
7
- 2024-05-31 16:32 - 3
8
- 2024-05-31 16:32 - Configue Model and tokenizer
9
- 2024-05-31 16:32 - Cuda check
10
- 2024-05-31 16:32 - True
11
- 2024-05-31 16:32 - 3
12
- 2024-05-31 16:32 - Configue Model and tokenizer
13
- 2024-05-31 16:32 - Cuda check
14
- 2024-05-31 16:32 - True
15
- 2024-05-31 16:32 - 3
16
- 2024-05-31 16:32 - Configue Model and tokenizer
17
- 2024-05-31 16:32 - Cuda check
18
- 2024-05-31 16:32 - True
19
- 2024-05-31 16:32 - 3
20
- 2024-05-31 16:32 - Configue Model and tokenizer
21
- 2024-05-31 16:32 - Cuda check
22
- 2024-05-31 16:32 - True
23
- 2024-05-31 16:32 - 3
24
- 2024-05-31 16:32 - Configue Model and tokenizer
25
- 2024-05-31 16:33 - Memory usage in 9.05 GB
26
- 2024-05-31 16:33 - Memory usage in 8.03 GB
27
- 2024-05-31 16:33 - Memory usage in 8.03 GB
28
- 2024-05-31 16:33 - Dataset loaded successfully:
29
  train-Jingmei/Pandemic_Wiki
30
- test -Jingmei/Pandemic_WHO
31
- 2024-05-31 16:33 - Dataset loaded successfully:
32
  train-Jingmei/Pandemic_Wiki
33
- test -Jingmei/Pandemic_WHO
34
- 2024-05-31 16:33 - Dataset loaded successfully:
35
- train-Jingmei/Pandemic_Wiki
36
- test -Jingmei/Pandemic_WHO
37
- 2024-05-31 16:34 - Tokenize data: DatasetDict({
38
- train: Dataset({
39
- features: ['input_ids', 'attention_mask'],
40
- num_rows: 2152
41
- })
42
- test: Dataset({
43
- features: ['input_ids', 'attention_mask'],
44
- num_rows: 8264
45
- })
46
- })
47
- 2024-05-31 16:34 - Tokenize data: DatasetDict({
48
  train: Dataset({
49
  features: ['input_ids', 'attention_mask'],
50
  num_rows: 2152
@@ -54,7 +24,7 @@
54
  num_rows: 8264
55
  })
56
  })
57
- 2024-05-31 16:34 - Tokenize data: DatasetDict({
58
  train: Dataset({
59
  features: ['input_ids', 'attention_mask'],
60
  num_rows: 2152
@@ -64,470 +34,91 @@
64
  num_rows: 8264
65
  })
66
  })
67
- 2024-05-31 16:37 - Cuda check
68
- 2024-05-31 16:37 - True
69
- 2024-05-31 16:37 - 3
70
- 2024-05-31 16:37 - Configue Model and tokenizer
71
- 2024-05-31 16:37 - Cuda check
72
- 2024-05-31 16:37 - True
73
- 2024-05-31 16:37 - 3
74
- 2024-05-31 16:37 - Configue Model and tokenizer
75
- 2024-05-31 16:37 - Cuda check
76
- 2024-05-31 16:37 - True
77
- 2024-05-31 16:37 - 3
78
- 2024-05-31 16:37 - Configue Model and tokenizer
79
- 2024-05-31 16:37 - Memory usage in 9.05 GB
80
- 2024-05-31 16:37 - Memory usage in 8.03 GB
81
- 2024-05-31 16:37 - Memory usage in 8.03 GB
82
- 2024-05-31 16:38 - Dataset loaded successfully:
83
- train-Jingmei/Pandemic_Wiki
84
- test -Jingmei/Pandemic_WHO
85
- 2024-05-31 16:38 - Dataset loaded successfully:
86
- train-Jingmei/Pandemic_Wiki
87
- test -Jingmei/Pandemic_WHO
88
- 2024-05-31 16:38 - Dataset loaded successfully:
89
- train-Jingmei/Pandemic_Wiki
90
- test -Jingmei/Pandemic_WHO
91
- 2024-05-31 16:38 - Tokenize data: DatasetDict({
92
- train: Dataset({
93
- features: ['input_ids', 'attention_mask'],
94
- num_rows: 2152
95
- })
96
- test: Dataset({
97
- features: ['input_ids', 'attention_mask'],
98
- num_rows: 100
99
- })
100
- })
101
- 2024-05-31 16:38 - Tokenize data: DatasetDict({
102
- train: Dataset({
103
- features: ['input_ids', 'attention_mask'],
104
- num_rows: 2152
105
- })
106
- test: Dataset({
107
- features: ['input_ids', 'attention_mask'],
108
- num_rows: 100
109
- })
110
- })
111
- 2024-05-31 16:38 - Split data into chunks:DatasetDict({
112
  train: Dataset({
113
  features: ['input_ids', 'attention_mask'],
114
  num_rows: 24863
115
  })
116
  test: Dataset({
117
  features: ['input_ids', 'attention_mask'],
118
- num_rows: 2868
119
  })
120
  })
121
- 2024-05-31 16:38 - Setup PEFT
122
- 2024-05-31 16:38 - Split data into chunks:DatasetDict({
 
123
  train: Dataset({
124
  features: ['input_ids', 'attention_mask'],
125
  num_rows: 24863
126
  })
127
  test: Dataset({
128
  features: ['input_ids', 'attention_mask'],
129
- num_rows: 2868
130
- })
131
- })
132
- 2024-05-31 16:38 - Setup PEFT
133
- 2024-05-31 16:38 - Tokenize data: DatasetDict({
134
- train: Dataset({
135
- features: ['input_ids', 'attention_mask'],
136
- num_rows: 2152
137
- })
138
- test: Dataset({
139
- features: ['input_ids', 'attention_mask'],
140
- num_rows: 100
141
  })
142
  })
143
- 2024-05-31 16:38 - Split data into chunks:DatasetDict({
144
- train: Dataset({
145
- features: ['input_ids', 'attention_mask'],
146
- num_rows: 24863
147
- })
148
- test: Dataset({
149
- features: ['input_ids', 'attention_mask'],
150
- num_rows: 2868
151
- })
152
- })
153
- 2024-05-31 16:38 - Setup PEFT
154
- 2024-05-31 16:38 - Setup optimizer
155
- 2024-05-31 16:38 - Setup optimizer
156
- 2024-05-31 16:38 - Setup optimizer
157
- 2024-05-31 16:38 - Start training!!
158
- 2024-05-31 16:38 - Start training!!
159
- 2024-05-31 16:38 - Start training!!
160
- 2024-05-31 16:38 - Cuda check
161
- 2024-05-31 16:38 - True
162
- 2024-05-31 16:38 - 3
163
- 2024-05-31 16:38 - Configue Model and tokenizer
164
- 2024-05-31 16:38 - Cuda check
165
- 2024-05-31 16:38 - True
166
- 2024-05-31 16:38 - 3
167
- 2024-05-31 16:38 - Configue Model and tokenizer
168
- 2024-05-31 16:38 - Cuda check
169
- 2024-05-31 16:38 - True
170
- 2024-05-31 16:38 - 3
171
- 2024-05-31 16:38 - Configue Model and tokenizer
172
- 2024-05-31 16:39 - Memory usage in 0.00 GB
173
- 2024-05-31 16:39 - Memory usage in 0.00 GB
174
- 2024-05-31 16:39 - Memory usage in 0.00 GB
175
- 2024-05-31 16:39 - Dataset loaded successfully:
176
  train-Jingmei/Pandemic_Wiki
177
- test -Jingmei/Pandemic_WHO
178
- 2024-05-31 16:39 - Tokenize data: DatasetDict({
179
  train: Dataset({
180
  features: ['input_ids', 'attention_mask'],
181
  num_rows: 2152
182
  })
183
  test: Dataset({
184
  features: ['input_ids', 'attention_mask'],
185
- num_rows: 100
186
- })
187
- })
188
- 2024-05-31 16:39 - Split data into chunks:DatasetDict({
189
- train: Dataset({
190
- features: ['input_ids', 'attention_mask'],
191
- num_rows: 24863
192
- })
193
- test: Dataset({
194
- features: ['input_ids', 'attention_mask'],
195
- num_rows: 2868
196
- })
197
- })
198
- 2024-05-31 16:39 - Setup PEFT
199
- 2024-05-31 16:39 - Setup optimizer
200
- 2024-05-31 16:39 - Dataset loaded successfully:
201
- train-Jingmei/Pandemic_Wiki
202
- test -Jingmei/Pandemic_WHO
203
- 2024-05-31 16:39 - Dataset loaded successfully:
204
- train-Jingmei/Pandemic_Wiki
205
- test -Jingmei/Pandemic_WHO
206
- 2024-05-31 16:39 - Tokenize data: DatasetDict({
207
- train: Dataset({
208
- features: ['input_ids', 'attention_mask'],
209
- num_rows: 2152
210
- })
211
- test: Dataset({
212
- features: ['input_ids', 'attention_mask'],
213
- num_rows: 100
214
- })
215
- })
216
- 2024-05-31 16:39 - Tokenize data: DatasetDict({
217
- train: Dataset({
218
- features: ['input_ids', 'attention_mask'],
219
- num_rows: 2152
220
- })
221
- test: Dataset({
222
- features: ['input_ids', 'attention_mask'],
223
- num_rows: 100
224
- })
225
- })
226
- 2024-05-31 16:39 - Split data into chunks:DatasetDict({
227
- train: Dataset({
228
- features: ['input_ids', 'attention_mask'],
229
- num_rows: 24863
230
- })
231
- test: Dataset({
232
- features: ['input_ids', 'attention_mask'],
233
- num_rows: 2868
234
- })
235
- })
236
- 2024-05-31 16:39 - Setup PEFT
237
- 2024-05-31 16:39 - Split data into chunks:DatasetDict({
238
- train: Dataset({
239
- features: ['input_ids', 'attention_mask'],
240
- num_rows: 24863
241
- })
242
- test: Dataset({
243
- features: ['input_ids', 'attention_mask'],
244
- num_rows: 2868
245
- })
246
- })
247
- 2024-05-31 16:39 - Setup PEFT
248
- 2024-05-31 16:39 - Setup optimizer
249
- 2024-05-31 16:39 - Setup optimizer
250
- 2024-05-31 16:39 - Start training!!
251
- 2024-05-31 16:39 - Start training!!
252
- 2024-05-31 16:39 - Start training!!
253
- 2024-05-31 16:42 - Cuda check
254
- 2024-05-31 16:42 - True
255
- 2024-05-31 16:42 - 3
256
- 2024-05-31 16:42 - Configue Model and tokenizer
257
- 2024-05-31 16:42 - Cuda check
258
- 2024-05-31 16:42 - True
259
- 2024-05-31 16:42 - 3
260
- 2024-05-31 16:42 - Configue Model and tokenizer
261
- 2024-05-31 16:42 - Cuda check
262
- 2024-05-31 16:42 - True
263
- 2024-05-31 16:42 - 3
264
- 2024-05-31 16:42 - Configue Model and tokenizer
265
- 2024-05-31 16:42 - Memory usage in 0.00 GB
266
- 2024-05-31 16:42 - Memory usage in 0.00 GB
267
- 2024-05-31 16:42 - Memory usage in 0.00 GB
268
- 2024-05-31 16:42 - Dataset loaded successfully:
269
- train-Jingmei/Pandemic_Wiki
270
- test -Jingmei/Pandemic_WHO
271
- 2024-05-31 16:42 - Tokenize data: DatasetDict({
272
- train: Dataset({
273
- features: ['input_ids', 'attention_mask'],
274
- num_rows: 2152
275
- })
276
- test: Dataset({
277
- features: ['input_ids', 'attention_mask'],
278
- num_rows: 100
279
- })
280
- })
281
- 2024-05-31 16:42 - Split data into chunks:DatasetDict({
282
- train: Dataset({
283
- features: ['input_ids', 'attention_mask'],
284
- num_rows: 24863
285
- })
286
- test: Dataset({
287
- features: ['input_ids', 'attention_mask'],
288
- num_rows: 2868
289
- })
290
- })
291
- 2024-05-31 16:42 - Setup PEFT
292
- 2024-05-31 16:42 - Dataset loaded successfully:
293
- train-Jingmei/Pandemic_Wiki
294
- test -Jingmei/Pandemic_WHO
295
- 2024-05-31 16:42 - Dataset loaded successfully:
296
- train-Jingmei/Pandemic_Wiki
297
- test -Jingmei/Pandemic_WHO
298
- 2024-05-31 16:42 - Tokenize data: DatasetDict({
299
- train: Dataset({
300
- features: ['input_ids', 'attention_mask'],
301
- num_rows: 2152
302
- })
303
- test: Dataset({
304
- features: ['input_ids', 'attention_mask'],
305
- num_rows: 100
306
- })
307
- })
308
- 2024-05-31 16:42 - Tokenize data: DatasetDict({
309
- train: Dataset({
310
- features: ['input_ids', 'attention_mask'],
311
- num_rows: 2152
312
- })
313
- test: Dataset({
314
- features: ['input_ids', 'attention_mask'],
315
- num_rows: 100
316
- })
317
- })
318
- 2024-05-31 16:42 - Split data into chunks:DatasetDict({
319
- train: Dataset({
320
- features: ['input_ids', 'attention_mask'],
321
- num_rows: 24863
322
- })
323
- test: Dataset({
324
- features: ['input_ids', 'attention_mask'],
325
- num_rows: 2868
326
- })
327
- })
328
- 2024-05-31 16:42 - Setup PEFT
329
- 2024-05-31 16:42 - Split data into chunks:DatasetDict({
330
- train: Dataset({
331
- features: ['input_ids', 'attention_mask'],
332
- num_rows: 24863
333
- })
334
- test: Dataset({
335
- features: ['input_ids', 'attention_mask'],
336
- num_rows: 2868
337
- })
338
- })
339
- 2024-05-31 16:42 - Setup PEFT
340
- 2024-05-31 16:42 - Setup optimizer
341
- 2024-05-31 16:42 - Setup optimizer
342
- 2024-05-31 16:42 - Setup optimizer
343
- 2024-05-31 16:42 - Start training!!
344
- 2024-05-31 16:42 - Start training!!
345
- 2024-05-31 16:42 - Start training!!
346
- 2024-05-31 16:44 - Cuda check
347
- 2024-05-31 16:44 - True
348
- 2024-05-31 16:44 - 3
349
- 2024-05-31 16:44 - Configue Model and tokenizer
350
- 2024-05-31 16:44 - Cuda check
351
- 2024-05-31 16:44 - True
352
- 2024-05-31 16:44 - 3
353
- 2024-05-31 16:44 - Configue Model and tokenizer
354
- 2024-05-31 16:44 - Cuda check
355
- 2024-05-31 16:44 - True
356
- 2024-05-31 16:44 - 3
357
- 2024-05-31 16:44 - Configue Model and tokenizer
358
- 2024-05-31 16:44 - Memory usage in 0.00 GB
359
- 2024-05-31 16:44 - Memory usage in 0.00 GB
360
- 2024-05-31 16:44 - Memory usage in 0.00 GB
361
- 2024-05-31 16:44 - Dataset loaded successfully:
362
- train-Jingmei/Pandemic_Wiki
363
- test -Jingmei/Pandemic_WHO
364
- 2024-05-31 16:44 - Tokenize data: DatasetDict({
365
- train: Dataset({
366
- features: ['input_ids', 'attention_mask'],
367
- num_rows: 2152
368
- })
369
- test: Dataset({
370
- features: ['input_ids', 'attention_mask'],
371
- num_rows: 100
372
- })
373
- })
374
- 2024-05-31 16:44 - Split data into chunks:DatasetDict({
375
- train: Dataset({
376
- features: ['input_ids', 'attention_mask'],
377
- num_rows: 24863
378
- })
379
- test: Dataset({
380
- features: ['input_ids', 'attention_mask'],
381
- num_rows: 2868
382
- })
383
- })
384
- 2024-05-31 16:44 - Setup PEFT
385
- 2024-05-31 16:44 - Dataset loaded successfully:
386
- train-Jingmei/Pandemic_Wiki
387
- test -Jingmei/Pandemic_WHO
388
- 2024-05-31 16:44 - Tokenize data: DatasetDict({
389
- train: Dataset({
390
- features: ['input_ids', 'attention_mask'],
391
- num_rows: 2152
392
- })
393
- test: Dataset({
394
- features: ['input_ids', 'attention_mask'],
395
- num_rows: 100
396
- })
397
- })
398
- 2024-05-31 16:44 - Split data into chunks:DatasetDict({
399
- train: Dataset({
400
- features: ['input_ids', 'attention_mask'],
401
- num_rows: 24863
402
- })
403
- test: Dataset({
404
- features: ['input_ids', 'attention_mask'],
405
- num_rows: 2868
406
  })
407
  })
408
- 2024-05-31 16:44 - Setup PEFT
409
- 2024-05-31 16:44 - Dataset loaded successfully:
410
  train-Jingmei/Pandemic_Wiki
411
- test -Jingmei/Pandemic_WHO
412
- 2024-05-31 16:44 - Tokenize data: DatasetDict({
413
- train: Dataset({
414
- features: ['input_ids', 'attention_mask'],
415
- num_rows: 2152
416
- })
417
- test: Dataset({
418
- features: ['input_ids', 'attention_mask'],
419
- num_rows: 100
420
- })
421
- })
422
- 2024-05-31 16:44 - Split data into chunks:DatasetDict({
423
  train: Dataset({
424
  features: ['input_ids', 'attention_mask'],
425
  num_rows: 24863
426
  })
427
  test: Dataset({
428
  features: ['input_ids', 'attention_mask'],
429
- num_rows: 2868
430
- })
431
- })
432
- 2024-05-31 16:44 - Setup PEFT
433
- 2024-05-31 16:44 - Setup optimizer
434
- 2024-05-31 16:44 - Setup optimizer
435
- 2024-05-31 16:44 - Setup optimizer
436
- 2024-05-31 16:44 - Start training!!
437
- 2024-05-31 16:44 - Start training!!
438
- 2024-05-31 16:44 - Start training!!
439
- 2024-05-31 16:47 - Cuda check
440
- 2024-05-31 16:47 - True
441
- 2024-05-31 16:47 - 3
442
- 2024-05-31 16:47 - Configue Model and tokenizer
443
- 2024-05-31 16:47 - Cuda check
444
- 2024-05-31 16:47 - True
445
- 2024-05-31 16:47 - 3
446
- 2024-05-31 16:47 - Configue Model and tokenizer
447
- 2024-05-31 16:47 - Cuda check
448
- 2024-05-31 16:47 - True
449
- 2024-05-31 16:47 - 3
450
- 2024-05-31 16:47 - Configue Model and tokenizer
451
- 2024-05-31 16:47 - Memory usage in 0.00 GB
452
- 2024-05-31 16:47 - Memory usage in 0.00 GB
453
- 2024-05-31 16:47 - Memory usage in 0.00 GB
454
- 2024-05-31 16:47 - Dataset loaded successfully:
455
- train-Jingmei/Pandemic_Wiki
456
- test -Jingmei/Pandemic_WHO
457
- 2024-05-31 16:47 - Dataset loaded successfully:
458
- train-Jingmei/Pandemic_Wiki
459
- test -Jingmei/Pandemic_WHO
460
- 2024-05-31 16:47 - Dataset loaded successfully:
461
- train-Jingmei/Pandemic_Wiki
462
- test -Jingmei/Pandemic_WHO
463
- 2024-05-31 16:47 - Tokenize data: DatasetDict({
464
- train: Dataset({
465
- features: ['input_ids', 'attention_mask'],
466
- num_rows: 2152
467
- })
468
- test: Dataset({
469
- features: ['input_ids', 'attention_mask'],
470
- num_rows: 100
471
- })
472
- })
473
- 2024-05-31 16:47 - Tokenize data: DatasetDict({
474
- train: Dataset({
475
- features: ['input_ids', 'attention_mask'],
476
- num_rows: 2152
477
- })
478
- test: Dataset({
479
- features: ['input_ids', 'attention_mask'],
480
- num_rows: 100
481
  })
482
  })
483
- 2024-05-31 16:47 - Tokenize data: DatasetDict({
 
484
  train: Dataset({
485
  features: ['input_ids', 'attention_mask'],
486
  num_rows: 2152
487
  })
488
  test: Dataset({
489
  features: ['input_ids', 'attention_mask'],
490
- num_rows: 100
491
- })
492
- })
493
- 2024-05-31 16:48 - Split data into chunks:DatasetDict({
494
- train: Dataset({
495
- features: ['input_ids', 'attention_mask'],
496
- num_rows: 24863
497
- })
498
- test: Dataset({
499
- features: ['input_ids', 'attention_mask'],
500
- num_rows: 2868
501
- })
502
- })
503
- 2024-05-31 16:48 - Setup PEFT
504
- 2024-05-31 16:48 - Split data into chunks:DatasetDict({
505
- train: Dataset({
506
- features: ['input_ids', 'attention_mask'],
507
- num_rows: 24863
508
- })
509
- test: Dataset({
510
- features: ['input_ids', 'attention_mask'],
511
- num_rows: 2868
512
  })
513
  })
514
- 2024-05-31 16:48 - Setup PEFT
515
- 2024-05-31 16:48 - Split data into chunks:DatasetDict({
516
  train: Dataset({
517
  features: ['input_ids', 'attention_mask'],
518
  num_rows: 24863
519
  })
520
  test: Dataset({
521
  features: ['input_ids', 'attention_mask'],
522
- num_rows: 2868
523
  })
524
  })
525
- 2024-05-31 16:48 - Setup PEFT
526
- 2024-05-31 16:48 - Setup optimizer
527
- 2024-05-31 16:48 - Setup optimizer
528
- 2024-05-31 16:48 - Setup optimizer
529
- 2024-05-31 16:48 - Start training!!
530
- 2024-05-31 16:48 - Start training!!
531
- 2024-05-31 16:48 - Start training!!
532
- 2024-05-31 17:28 - Training complete!!!
533
- 2024-05-31 17:28 - Training complete!!!
 
1
+ 2024-06-01 14:49 - Cuda check
2
+ 2024-06-01 14:49 - True
3
+ 2024-06-01 14:49 - 2
4
+ 2024-06-01 14:49 - Configue Model and tokenizer
5
+ 2024-06-01 14:49 - Cuda check
6
+ 2024-06-01 14:49 - True
7
+ 2024-06-01 14:49 - 2
8
+ 2024-06-01 14:49 - Configue Model and tokenizer
9
+ 2024-06-01 14:49 - Memory usage in 0.00 GB
10
+ 2024-06-01 14:49 - Memory usage in 0.00 GB
11
+ 2024-06-01 14:49 - Dataset loaded successfully:
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
12
  train-Jingmei/Pandemic_Wiki
13
+ test -Jingmei/Pandemic
14
+ 2024-06-01 14:49 - Dataset loaded successfully:
15
  train-Jingmei/Pandemic_Wiki
16
+ test -Jingmei/Pandemic
17
+ 2024-06-01 14:49 - Tokenize data: DatasetDict({
 
 
 
 
 
 
 
 
 
 
 
 
 
18
  train: Dataset({
19
  features: ['input_ids', 'attention_mask'],
20
  num_rows: 2152
 
24
  num_rows: 8264
25
  })
26
  })
27
+ 2024-06-01 14:49 - Tokenize data: DatasetDict({
28
  train: Dataset({
29
  features: ['input_ids', 'attention_mask'],
30
  num_rows: 2152
 
34
  num_rows: 8264
35
  })
36
  })
37
+ 2024-06-01 14:49 - Split data into chunks:DatasetDict({
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
38
  train: Dataset({
39
  features: ['input_ids', 'attention_mask'],
40
  num_rows: 24863
41
  })
42
  test: Dataset({
43
  features: ['input_ids', 'attention_mask'],
44
+ num_rows: 198964
45
  })
46
  })
47
+ 2024-06-01 14:49 - Setup PEFT
48
+ 2024-06-01 14:49 - Setup optimizer
49
+ 2024-06-01 14:49 - Split data into chunks:DatasetDict({
50
  train: Dataset({
51
  features: ['input_ids', 'attention_mask'],
52
  num_rows: 24863
53
  })
54
  test: Dataset({
55
  features: ['input_ids', 'attention_mask'],
56
+ num_rows: 198964
 
 
 
 
 
 
 
 
 
 
 
57
  })
58
  })
59
+ 2024-06-01 14:49 - Setup PEFT
60
+ 2024-06-01 14:49 - Setup optimizer
61
+ 2024-06-01 14:49 - Start training!!
62
+ 2024-06-01 14:49 - Start training!!
63
+ 2024-06-01 14:51 - Cuda check
64
+ 2024-06-01 14:51 - True
65
+ 2024-06-01 14:51 - 2
66
+ 2024-06-01 14:51 - Configue Model and tokenizer
67
+ 2024-06-01 14:51 - Cuda check
68
+ 2024-06-01 14:51 - True
69
+ 2024-06-01 14:51 - 2
70
+ 2024-06-01 14:51 - Configue Model and tokenizer
71
+ 2024-06-01 14:51 - Memory usage in 0.00 GB
72
+ 2024-06-01 14:51 - Memory usage in 0.00 GB
73
+ 2024-06-01 14:51 - Dataset loaded successfully:
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
74
  train-Jingmei/Pandemic_Wiki
75
+ test -Jingmei/Pandemic
76
+ 2024-06-01 14:51 - Tokenize data: DatasetDict({
77
  train: Dataset({
78
  features: ['input_ids', 'attention_mask'],
79
  num_rows: 2152
80
  })
81
  test: Dataset({
82
  features: ['input_ids', 'attention_mask'],
83
+ num_rows: 8264
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
84
  })
85
  })
86
+ 2024-06-01 14:51 - Dataset loaded successfully:
 
87
  train-Jingmei/Pandemic_Wiki
88
+ test -Jingmei/Pandemic
89
+ 2024-06-01 14:51 - Split data into chunks:DatasetDict({
 
 
 
 
 
 
 
 
 
 
90
  train: Dataset({
91
  features: ['input_ids', 'attention_mask'],
92
  num_rows: 24863
93
  })
94
  test: Dataset({
95
  features: ['input_ids', 'attention_mask'],
96
+ num_rows: 198964
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
97
  })
98
  })
99
+ 2024-06-01 14:51 - Setup PEFT
100
+ 2024-06-01 14:51 - Tokenize data: DatasetDict({
101
  train: Dataset({
102
  features: ['input_ids', 'attention_mask'],
103
  num_rows: 2152
104
  })
105
  test: Dataset({
106
  features: ['input_ids', 'attention_mask'],
107
+ num_rows: 8264
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
108
  })
109
  })
110
+ 2024-06-01 14:51 - Split data into chunks:DatasetDict({
 
111
  train: Dataset({
112
  features: ['input_ids', 'attention_mask'],
113
  num_rows: 24863
114
  })
115
  test: Dataset({
116
  features: ['input_ids', 'attention_mask'],
117
+ num_rows: 198964
118
  })
119
  })
120
+ 2024-06-01 14:51 - Setup PEFT
121
+ 2024-06-01 14:51 - Setup optimizer
122
+ 2024-06-01 14:51 - Setup optimizer
123
+ 2024-06-01 14:51 - Start training!!
124
+ 2024-06-01 14:51 - Start training!!
 
 
 
 
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:40d2f7de8feae613a30567af3919480984d8f1b9ce0ae696737a789c2949339b
3
  size 5176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44addaf9c8b314d3a88a7b01508035d34a4e18244af21ea2cde47f3d51ac0894
3
  size 5176