oz1115 commited on
Commit
c75073f
1 Parent(s): ce08871

oz1115/phi-3-mini-QLoRA-adapter

Browse files
README.md CHANGED
@@ -14,7 +14,6 @@ model-index:
14
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
  should probably proofread and complete it, then remove this comment. -->
16
 
17
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/gsitm/phi-3-mini-QLoRA/runs/a5amu4ck)
18
  # phi-3-mini-QLoRA
19
 
20
  This model is a fine-tuned version of [microsoft/Phi-3-mini-4k-instruct](https://huggingface.co/microsoft/Phi-3-mini-4k-instruct) on an unknown dataset.
@@ -53,8 +52,8 @@ The following hyperparameters were used during training:
53
 
54
  ### Framework versions
55
 
56
- - PEFT 0.11.1
57
- - Transformers 4.42.4
58
  - Pytorch 2.3.1+cu121
59
  - Datasets 2.20.0
60
  - Tokenizers 0.19.1
 
14
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
  should probably proofread and complete it, then remove this comment. -->
16
 
 
17
  # phi-3-mini-QLoRA
18
 
19
  This model is a fine-tuned version of [microsoft/Phi-3-mini-4k-instruct](https://huggingface.co/microsoft/Phi-3-mini-4k-instruct) on an unknown dataset.
 
52
 
53
  ### Framework versions
54
 
55
+ - PEFT 0.12.0
56
+ - Transformers 4.44.0
57
  - Pytorch 2.3.1+cu121
58
  - Datasets 2.20.0
59
  - Tokenizers 0.19.1
adapter_config.json CHANGED
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
 
 
 
23
  "k_proj",
24
  "o_proj",
25
- "v_proj",
26
- "down_proj",
27
- "up_proj",
28
- "gate_proj",
29
- "q_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "q_proj",
24
+ "up_proj",
25
+ "down_proj",
26
+ "v_proj",
27
  "k_proj",
28
  "o_proj",
29
+ "gate_proj"
 
 
 
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6adcb61565d84963985e2370a78e4d0f4439152c6e4bc1bb9b37d2823949ebd6
3
  size 35668592
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:31e8c33602d963b527aef02955084cd84f66c498c5179d364b9c8861dead66e0
3
  size 35668592
special_tokens_map.json CHANGED
@@ -13,13 +13,7 @@
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
- "pad_token": {
17
- "content": "<unk>",
18
- "lstrip": false,
19
- "normalized": false,
20
- "rstrip": false,
21
- "single_word": false
22
- },
23
  "unk_token": {
24
  "content": "<unk>",
25
  "lstrip": false,
 
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
+ "pad_token": "<unk>",
 
 
 
 
 
 
17
  "unk_token": {
18
  "content": "<unk>",
19
  "lstrip": false,
tokenizer_config.json CHANGED
@@ -121,15 +121,11 @@
121
  "clean_up_tokenization_spaces": false,
122
  "eos_token": "<|endoftext|>",
123
  "legacy": false,
124
- "max_length": 512,
125
  "model_max_length": 4096,
126
  "pad_token": "<unk>",
127
  "padding_side": "left",
128
  "sp_model_kwargs": {},
129
- "stride": 0,
130
  "tokenizer_class": "LlamaTokenizer",
131
- "truncation_side": "right",
132
- "truncation_strategy": "longest_first",
133
  "unk_token": "<unk>",
134
  "use_default_system_prompt": false
135
  }
 
121
  "clean_up_tokenization_spaces": false,
122
  "eos_token": "<|endoftext|>",
123
  "legacy": false,
 
124
  "model_max_length": 4096,
125
  "pad_token": "<unk>",
126
  "padding_side": "left",
127
  "sp_model_kwargs": {},
 
128
  "tokenizer_class": "LlamaTokenizer",
 
 
129
  "unk_token": "<unk>",
130
  "use_default_system_prompt": false
131
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f4a0eb564f9aa3880098ef1e9105a6f0b5a1efedce3886fca8da19de8b4887fa
3
- size 5368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1f3f1fba257ee86d8eb0b4b9caa8d2e295d0c38487b19895aa741c773d40529
3
+ size 5432