Jingmei commited on
Commit
a1a31f9
1 Parent(s): 69fead1

End of training

Browse files
Files changed (4) hide show
  1. README.md +1 -1
  2. adapter_config.json +2 -2
  3. trainer_peft.log +126 -0
  4. training_args.bin +1 -1
README.md CHANGED
@@ -12,7 +12,7 @@ model-index:
12
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
  should probably proofread and complete it, then remove this comment. -->
14
 
15
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/noc-lab/PMC_LLAMA2_7B_trainer_lora/runs/n09numov)
16
  # PMC_LLAMA2_7B_trainer_lora
17
 
18
  This model is a fine-tuned version of [chaoyi-wu/PMC_LLAMA_7B](https://huggingface.co/chaoyi-wu/PMC_LLAMA_7B) on an unknown dataset.
 
12
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
  should probably proofread and complete it, then remove this comment. -->
14
 
15
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/noc-lab/PMC_LLAMA2_7B_trainer_lora/runs/pvbcl0q5)
16
  # PMC_LLAMA2_7B_trainer_lora
17
 
18
  This model is a fine-tuned version of [chaoyi-wu/PMC_LLAMA_7B](https://huggingface.co/chaoyi-wu/PMC_LLAMA_7B) on an unknown dataset.
adapter_config.json CHANGED
@@ -20,8 +20,8 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "q_proj",
24
- "v_proj"
25
  ],
26
  "task_type": "CAUSAL_LM",
27
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "v_proj",
24
+ "q_proj"
25
  ],
26
  "task_type": "CAUSAL_LM",
27
  "use_dora": false,
trainer_peft.log CHANGED
@@ -123,3 +123,129 @@
123
  2024-06-01 14:51 - Start training!!
124
  2024-06-01 14:51 - Start training!!
125
  2024-06-01 15:49 - Training complete!!!
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
123
  2024-06-01 14:51 - Start training!!
124
  2024-06-01 14:51 - Start training!!
125
  2024-06-01 15:49 - Training complete!!!
126
+ 2024-06-01 15:49 - Training complete!!!
127
+ 2024-06-01 20:49 - Cuda check
128
+ 2024-06-01 20:49 - True
129
+ 2024-06-01 20:49 - 2
130
+ 2024-06-01 20:49 - Configue Model and tokenizer
131
+ 2024-06-01 20:49 - Cuda check
132
+ 2024-06-01 20:49 - True
133
+ 2024-06-01 20:49 - 2
134
+ 2024-06-01 20:49 - Configue Model and tokenizer
135
+ 2024-06-01 20:49 - Memory usage in 0.00 GB
136
+ 2024-06-01 20:49 - Memory usage in 0.00 GB
137
+ 2024-06-01 20:49 - Dataset loaded successfully:
138
+ train-Jingmei/Pandemic_Wiki
139
+ test -Jingmei/Pandemic
140
+ 2024-06-01 20:49 - Dataset loaded successfully:
141
+ train-Jingmei/Pandemic_Wiki
142
+ test -Jingmei/Pandemic
143
+ 2024-06-01 20:49 - Tokenize data: DatasetDict({
144
+ train: Dataset({
145
+ features: ['input_ids', 'attention_mask'],
146
+ num_rows: 2152
147
+ })
148
+ test: Dataset({
149
+ features: ['input_ids', 'attention_mask'],
150
+ num_rows: 8264
151
+ })
152
+ })
153
+ 2024-06-01 20:49 - Tokenize data: DatasetDict({
154
+ train: Dataset({
155
+ features: ['input_ids', 'attention_mask'],
156
+ num_rows: 2152
157
+ })
158
+ test: Dataset({
159
+ features: ['input_ids', 'attention_mask'],
160
+ num_rows: 8264
161
+ })
162
+ })
163
+ 2024-06-01 20:49 - Split data into chunks:DatasetDict({
164
+ train: Dataset({
165
+ features: ['input_ids', 'attention_mask'],
166
+ num_rows: 24863
167
+ })
168
+ test: Dataset({
169
+ features: ['input_ids', 'attention_mask'],
170
+ num_rows: 198964
171
+ })
172
+ })
173
+ 2024-06-01 20:49 - Setup PEFT
174
+ 2024-06-01 20:49 - Split data into chunks:DatasetDict({
175
+ train: Dataset({
176
+ features: ['input_ids', 'attention_mask'],
177
+ num_rows: 24863
178
+ })
179
+ test: Dataset({
180
+ features: ['input_ids', 'attention_mask'],
181
+ num_rows: 198964
182
+ })
183
+ })
184
+ 2024-06-01 20:49 - Setup PEFT
185
+ 2024-06-01 20:49 - Setup optimizer
186
+ 2024-06-01 20:49 - Setup optimizer
187
+ 2024-06-01 20:49 - Start training!!
188
+ 2024-06-01 20:49 - Start training!!
189
+ 2024-06-01 20:55 - Cuda check
190
+ 2024-06-01 20:55 - True
191
+ 2024-06-01 20:55 - 2
192
+ 2024-06-01 20:55 - Configue Model and tokenizer
193
+ 2024-06-01 20:55 - Cuda check
194
+ 2024-06-01 20:55 - True
195
+ 2024-06-01 20:55 - 2
196
+ 2024-06-01 20:55 - Configue Model and tokenizer
197
+ 2024-06-01 20:55 - Memory usage in 0.00 GB
198
+ 2024-06-01 20:55 - Memory usage in 0.00 GB
199
+ 2024-06-01 20:55 - Dataset loaded successfully:
200
+ train-Jingmei/Pandemic_Wiki
201
+ test -Jingmei/Pandemic
202
+ 2024-06-01 20:55 - Tokenize data: DatasetDict({
203
+ train: Dataset({
204
+ features: ['input_ids', 'attention_mask'],
205
+ num_rows: 2152
206
+ })
207
+ test: Dataset({
208
+ features: ['input_ids', 'attention_mask'],
209
+ num_rows: 8264
210
+ })
211
+ })
212
+ 2024-06-01 20:55 - Dataset loaded successfully:
213
+ train-Jingmei/Pandemic_Wiki
214
+ test -Jingmei/Pandemic
215
+ 2024-06-01 20:55 - Split data into chunks:DatasetDict({
216
+ train: Dataset({
217
+ features: ['input_ids', 'attention_mask'],
218
+ num_rows: 24863
219
+ })
220
+ test: Dataset({
221
+ features: ['input_ids', 'attention_mask'],
222
+ num_rows: 198964
223
+ })
224
+ })
225
+ 2024-06-01 20:55 - Setup PEFT
226
+ 2024-06-01 20:55 - Tokenize data: DatasetDict({
227
+ train: Dataset({
228
+ features: ['input_ids', 'attention_mask'],
229
+ num_rows: 2152
230
+ })
231
+ test: Dataset({
232
+ features: ['input_ids', 'attention_mask'],
233
+ num_rows: 8264
234
+ })
235
+ })
236
+ 2024-06-01 20:55 - Split data into chunks:DatasetDict({
237
+ train: Dataset({
238
+ features: ['input_ids', 'attention_mask'],
239
+ num_rows: 24863
240
+ })
241
+ test: Dataset({
242
+ features: ['input_ids', 'attention_mask'],
243
+ num_rows: 198964
244
+ })
245
+ })
246
+ 2024-06-01 20:55 - Setup PEFT
247
+ 2024-06-01 20:55 - Setup optimizer
248
+ 2024-06-01 20:55 - Setup optimizer
249
+ 2024-06-01 20:55 - Continue training!!
250
+ 2024-06-01 20:55 - Continue training!!
251
+ 2024-06-01 20:56 - Training complete!!!
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:44addaf9c8b314d3a88a7b01508035d34a4e18244af21ea2cde47f3d51ac0894
3
  size 5176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f504f95f70f5d19e6e4e48a5e120ce2676ce8241842d7143d46d087125307096
3
  size 5176