End of training
Browse files- README.md +1 -1
- adapter_config.json +2 -2
- trainer_peft.log +126 -0
- training_args.bin +1 -1
README.md
CHANGED
@@ -12,7 +12,7 @@ model-index:
|
|
12 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
13 |
should probably proofread and complete it, then remove this comment. -->
|
14 |
|
15 |
-
[<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/noc-lab/PMC_LLAMA2_7B_trainer_lora/runs/
|
16 |
# PMC_LLAMA2_7B_trainer_lora
|
17 |
|
18 |
This model is a fine-tuned version of [chaoyi-wu/PMC_LLAMA_7B](https://huggingface.co/chaoyi-wu/PMC_LLAMA_7B) on an unknown dataset.
|
|
|
12 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
13 |
should probably proofread and complete it, then remove this comment. -->
|
14 |
|
15 |
+
[<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/noc-lab/PMC_LLAMA2_7B_trainer_lora/runs/pvbcl0q5)
|
16 |
# PMC_LLAMA2_7B_trainer_lora
|
17 |
|
18 |
This model is a fine-tuned version of [chaoyi-wu/PMC_LLAMA_7B](https://huggingface.co/chaoyi-wu/PMC_LLAMA_7B) on an unknown dataset.
|
adapter_config.json
CHANGED
@@ -20,8 +20,8 @@
|
|
20 |
"rank_pattern": {},
|
21 |
"revision": null,
|
22 |
"target_modules": [
|
23 |
-
"
|
24 |
-
"
|
25 |
],
|
26 |
"task_type": "CAUSAL_LM",
|
27 |
"use_dora": false,
|
|
|
20 |
"rank_pattern": {},
|
21 |
"revision": null,
|
22 |
"target_modules": [
|
23 |
+
"v_proj",
|
24 |
+
"q_proj"
|
25 |
],
|
26 |
"task_type": "CAUSAL_LM",
|
27 |
"use_dora": false,
|
trainer_peft.log
CHANGED
@@ -123,3 +123,129 @@
|
|
123 |
2024-06-01 14:51 - Start training!!
|
124 |
2024-06-01 14:51 - Start training!!
|
125 |
2024-06-01 15:49 - Training complete!!!
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
123 |
2024-06-01 14:51 - Start training!!
|
124 |
2024-06-01 14:51 - Start training!!
|
125 |
2024-06-01 15:49 - Training complete!!!
|
126 |
+
2024-06-01 15:49 - Training complete!!!
|
127 |
+
2024-06-01 20:49 - Cuda check
|
128 |
+
2024-06-01 20:49 - True
|
129 |
+
2024-06-01 20:49 - 2
|
130 |
+
2024-06-01 20:49 - Configue Model and tokenizer
|
131 |
+
2024-06-01 20:49 - Cuda check
|
132 |
+
2024-06-01 20:49 - True
|
133 |
+
2024-06-01 20:49 - 2
|
134 |
+
2024-06-01 20:49 - Configue Model and tokenizer
|
135 |
+
2024-06-01 20:49 - Memory usage in 0.00 GB
|
136 |
+
2024-06-01 20:49 - Memory usage in 0.00 GB
|
137 |
+
2024-06-01 20:49 - Dataset loaded successfully:
|
138 |
+
train-Jingmei/Pandemic_Wiki
|
139 |
+
test -Jingmei/Pandemic
|
140 |
+
2024-06-01 20:49 - Dataset loaded successfully:
|
141 |
+
train-Jingmei/Pandemic_Wiki
|
142 |
+
test -Jingmei/Pandemic
|
143 |
+
2024-06-01 20:49 - Tokenize data: DatasetDict({
|
144 |
+
train: Dataset({
|
145 |
+
features: ['input_ids', 'attention_mask'],
|
146 |
+
num_rows: 2152
|
147 |
+
})
|
148 |
+
test: Dataset({
|
149 |
+
features: ['input_ids', 'attention_mask'],
|
150 |
+
num_rows: 8264
|
151 |
+
})
|
152 |
+
})
|
153 |
+
2024-06-01 20:49 - Tokenize data: DatasetDict({
|
154 |
+
train: Dataset({
|
155 |
+
features: ['input_ids', 'attention_mask'],
|
156 |
+
num_rows: 2152
|
157 |
+
})
|
158 |
+
test: Dataset({
|
159 |
+
features: ['input_ids', 'attention_mask'],
|
160 |
+
num_rows: 8264
|
161 |
+
})
|
162 |
+
})
|
163 |
+
2024-06-01 20:49 - Split data into chunks:DatasetDict({
|
164 |
+
train: Dataset({
|
165 |
+
features: ['input_ids', 'attention_mask'],
|
166 |
+
num_rows: 24863
|
167 |
+
})
|
168 |
+
test: Dataset({
|
169 |
+
features: ['input_ids', 'attention_mask'],
|
170 |
+
num_rows: 198964
|
171 |
+
})
|
172 |
+
})
|
173 |
+
2024-06-01 20:49 - Setup PEFT
|
174 |
+
2024-06-01 20:49 - Split data into chunks:DatasetDict({
|
175 |
+
train: Dataset({
|
176 |
+
features: ['input_ids', 'attention_mask'],
|
177 |
+
num_rows: 24863
|
178 |
+
})
|
179 |
+
test: Dataset({
|
180 |
+
features: ['input_ids', 'attention_mask'],
|
181 |
+
num_rows: 198964
|
182 |
+
})
|
183 |
+
})
|
184 |
+
2024-06-01 20:49 - Setup PEFT
|
185 |
+
2024-06-01 20:49 - Setup optimizer
|
186 |
+
2024-06-01 20:49 - Setup optimizer
|
187 |
+
2024-06-01 20:49 - Start training!!
|
188 |
+
2024-06-01 20:49 - Start training!!
|
189 |
+
2024-06-01 20:55 - Cuda check
|
190 |
+
2024-06-01 20:55 - True
|
191 |
+
2024-06-01 20:55 - 2
|
192 |
+
2024-06-01 20:55 - Configue Model and tokenizer
|
193 |
+
2024-06-01 20:55 - Cuda check
|
194 |
+
2024-06-01 20:55 - True
|
195 |
+
2024-06-01 20:55 - 2
|
196 |
+
2024-06-01 20:55 - Configue Model and tokenizer
|
197 |
+
2024-06-01 20:55 - Memory usage in 0.00 GB
|
198 |
+
2024-06-01 20:55 - Memory usage in 0.00 GB
|
199 |
+
2024-06-01 20:55 - Dataset loaded successfully:
|
200 |
+
train-Jingmei/Pandemic_Wiki
|
201 |
+
test -Jingmei/Pandemic
|
202 |
+
2024-06-01 20:55 - Tokenize data: DatasetDict({
|
203 |
+
train: Dataset({
|
204 |
+
features: ['input_ids', 'attention_mask'],
|
205 |
+
num_rows: 2152
|
206 |
+
})
|
207 |
+
test: Dataset({
|
208 |
+
features: ['input_ids', 'attention_mask'],
|
209 |
+
num_rows: 8264
|
210 |
+
})
|
211 |
+
})
|
212 |
+
2024-06-01 20:55 - Dataset loaded successfully:
|
213 |
+
train-Jingmei/Pandemic_Wiki
|
214 |
+
test -Jingmei/Pandemic
|
215 |
+
2024-06-01 20:55 - Split data into chunks:DatasetDict({
|
216 |
+
train: Dataset({
|
217 |
+
features: ['input_ids', 'attention_mask'],
|
218 |
+
num_rows: 24863
|
219 |
+
})
|
220 |
+
test: Dataset({
|
221 |
+
features: ['input_ids', 'attention_mask'],
|
222 |
+
num_rows: 198964
|
223 |
+
})
|
224 |
+
})
|
225 |
+
2024-06-01 20:55 - Setup PEFT
|
226 |
+
2024-06-01 20:55 - Tokenize data: DatasetDict({
|
227 |
+
train: Dataset({
|
228 |
+
features: ['input_ids', 'attention_mask'],
|
229 |
+
num_rows: 2152
|
230 |
+
})
|
231 |
+
test: Dataset({
|
232 |
+
features: ['input_ids', 'attention_mask'],
|
233 |
+
num_rows: 8264
|
234 |
+
})
|
235 |
+
})
|
236 |
+
2024-06-01 20:55 - Split data into chunks:DatasetDict({
|
237 |
+
train: Dataset({
|
238 |
+
features: ['input_ids', 'attention_mask'],
|
239 |
+
num_rows: 24863
|
240 |
+
})
|
241 |
+
test: Dataset({
|
242 |
+
features: ['input_ids', 'attention_mask'],
|
243 |
+
num_rows: 198964
|
244 |
+
})
|
245 |
+
})
|
246 |
+
2024-06-01 20:55 - Setup PEFT
|
247 |
+
2024-06-01 20:55 - Setup optimizer
|
248 |
+
2024-06-01 20:55 - Setup optimizer
|
249 |
+
2024-06-01 20:55 - Continue training!!
|
250 |
+
2024-06-01 20:55 - Continue training!!
|
251 |
+
2024-06-01 20:56 - Training complete!!!
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 5176
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f504f95f70f5d19e6e4e48a5e120ce2676ce8241842d7143d46d087125307096
|
3 |
size 5176
|