isaacOnline commited on
Commit
ea4048a
1 Parent(s): 605dca8

Training in progress, step 20

Browse files
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  library_name: peft
3
- base_model: mistralai/Mistral-7B-Instruct-v0.1
4
  ---
5
 
6
  # Model Card for Model ID
 
1
  ---
2
  library_name: peft
3
+ base_model: meta-llama/Llama-2-7b-chat-hf
4
  ---
5
 
6
  # Model Card for Model ID
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "meta-llama/Llama-2-7b-chat-hf",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
@@ -19,13 +19,13 @@
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
 
 
 
22
  "down_proj",
23
  "gate_proj",
24
- "up_proj",
25
- "v_proj",
26
- "k_proj",
27
  "q_proj",
28
- "o_proj"
29
  ],
30
  "task_type": "CAUSAL_LM",
31
  "use_rslora": false
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "mistralai/Mistral-7B-Instruct-v0.1",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
 
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
+ "k_proj",
23
+ "o_proj",
24
+ "up_proj",
25
  "down_proj",
26
  "gate_proj",
 
 
 
27
  "q_proj",
28
+ "v_proj"
29
  ],
30
  "task_type": "CAUSAL_LM",
31
  "use_rslora": false
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b4af5a9c1c16eb58f8f74295acee632c517069fb9d3d7a57f2fe3c471939e302
3
- size 319876032
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:16e622041813b3f1ced0c71d797aa55ded3eb83e7500cc861998bfa4a41798a3
3
+ size 335604696
qual_clasification.log CHANGED
@@ -10,3 +10,27 @@
10
  2024-02-12 23:54:52,801 - INFO - __main__ - Loaded Model ID: mistralai/Mistral-7B-Instruct-v0.1
11
  2024-02-12 23:54:53,906 - INFO - __main__ - Loaded LoRA Model
12
  2024-02-12 23:54:54,461 - INFO - __main__ - Instantiated Trainer
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
10
  2024-02-12 23:54:52,801 - INFO - __main__ - Loaded Model ID: mistralai/Mistral-7B-Instruct-v0.1
11
  2024-02-12 23:54:53,906 - INFO - __main__ - Loaded LoRA Model
12
  2024-02-12 23:54:54,461 - INFO - __main__ - Instantiated Trainer
13
+ 2024-02-13 00:27:59,465 - INFO - __main__ - Loaded Model ID: meta-llama/Llama-2-7b-chat-hf
14
+ 2024-02-13 00:28:01,163 - INFO - __main__ - Loaded LoRA Model
15
+ 2024-02-13 00:37:53,715 - INFO - __main__ - Instantiated Trainer
16
+ 2024-02-13 00:38:10,056 - INFO - __main__ - Completed fine-tuning
17
+ 2024-02-13 00:38:12,663 - INFO - __main__ - Saved model and tokenizer to machine_learning/llm_finetune_models/0
18
+ 2024-02-13 00:38:28,563 - INFO - __main__ - Saved model to hub
19
+ 2024-02-13 01:15:56,957 - INFO - __main__ - Loaded Model ID: meta-llama/Llama-2-7b-chat-hf
20
+ 2024-02-13 01:15:58,628 - INFO - __main__ - Loaded LoRA Model
21
+ 2024-02-13 01:16:00,388 - INFO - __main__ - Instantiated Trainer
22
+ 2024-02-13 01:16:17,164 - INFO - __main__ - Completed fine-tuning
23
+ 2024-02-13 01:16:19,646 - INFO - __main__ - Saved model and tokenizer to machine_learning/llm_finetune_models/0
24
+ 2024-02-13 01:16:33,544 - INFO - __main__ - Saved model to hub
25
+ 2024-02-13 01:25:41,929 - INFO - __main__ - Loaded Model ID: meta-llama/Llama-2-7b-chat-hf
26
+ 2024-02-13 01:25:43,508 - INFO - __main__ - Loaded LoRA Model
27
+ 2024-02-13 01:25:45,068 - INFO - __main__ - Instantiated Trainer
28
+ 2024-02-13 01:26:01,011 - INFO - __main__ - Completed fine-tuning
29
+ 2024-02-13 01:26:03,356 - INFO - __main__ - Saved model and tokenizer to machine_learning/llm_finetune_models/0
30
+ 2024-02-13 01:26:17,186 - INFO - __main__ - Saved model to hub
31
+ 2024-02-13 01:38:55,615 - INFO - __main__ - Loaded Model ID: mistralai/Mistral-7B-Instruct-v0.1
32
+ 2024-02-13 01:38:57,066 - INFO - __main__ - Loaded LoRA Model
33
+ 2024-02-13 01:38:57,952 - INFO - __main__ - Instantiated Trainer
34
+ 2024-02-13 01:41:15,404 - INFO - __main__ - Loaded Model ID: mistralai/Mistral-7B-Instruct-v0.1
35
+ 2024-02-13 01:41:16,464 - INFO - __main__ - Loaded LoRA Model
36
+ 2024-02-13 01:41:17,219 - INFO - __main__ - Instantiated Trainer
tokenizer.json CHANGED
@@ -2,7 +2,7 @@
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
- "max_length": 974,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
 
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
+ "max_length": 2048,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5c4c3ec799eed003079ecaa1a1d9418c5cde9bee9ec29b81402595f794d71db7
3
  size 4728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9643948292c3ba4fe7b9435299dc40f33fa2b9fc74029e8c153976f1ed62f47e
3
  size 4728