Model save
Browse files- README.md +1 -1
- adapter_config.json +2 -2
- adapter_model.safetensors +1 -1
- runs/Feb23_12-48-05_6a55cc4830ee/events.out.tfevents.1708692653.6a55cc4830ee.936.0 +3 -0
- runs/Feb23_13-03-08_6a55cc4830ee/events.out.tfevents.1708693571.6a55cc4830ee.936.1 +3 -0
- runs/Feb23_13-08-20_6a55cc4830ee/events.out.tfevents.1708693867.6a55cc4830ee.936.2 +3 -0
- runs/Feb23_13-16-44_6a55cc4830ee/events.out.tfevents.1708694372.6a55cc4830ee.8321.0 +3 -0
- runs/Feb23_13-21-04_6a55cc4830ee/events.out.tfevents.1708694651.6a55cc4830ee.8321.1 +3 -0
- tokenizer_config.json +1 -1
- training_args.bin +1 -1
README.md
CHANGED
@@ -41,7 +41,7 @@ The following hyperparameters were used during training:
|
|
41 |
- seed: 42
|
42 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
43 |
- lr_scheduler_type: cosine
|
44 |
-
- training_steps:
|
45 |
- mixed_precision_training: Native AMP
|
46 |
|
47 |
### Training results
|
|
|
41 |
- seed: 42
|
42 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
43 |
- lr_scheduler_type: cosine
|
44 |
+
- training_steps: 1000
|
45 |
- mixed_precision_training: Native AMP
|
46 |
|
47 |
### Training results
|
adapter_config.json
CHANGED
@@ -19,8 +19,8 @@
|
|
19 |
"rank_pattern": {},
|
20 |
"revision": null,
|
21 |
"target_modules": [
|
22 |
-
"
|
23 |
-
"
|
24 |
],
|
25 |
"task_type": "CAUSAL_LM",
|
26 |
"use_rslora": false
|
|
|
19 |
"rank_pattern": {},
|
20 |
"revision": null,
|
21 |
"target_modules": [
|
22 |
+
"v_proj",
|
23 |
+
"q_proj"
|
24 |
],
|
25 |
"task_type": "CAUSAL_LM",
|
26 |
"use_rslora": false
|
adapter_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 7382336
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e81b2125abd427e6bc280c704230f9d9477819a8752efa3c65a9bff668a6193d
|
3 |
size 7382336
|
runs/Feb23_12-48-05_6a55cc4830ee/events.out.tfevents.1708692653.6a55cc4830ee.936.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d87984458a5ad5b3ce1072c504144c12e3ff4998b4e576f5bdceb9fcd4fb17aa
|
3 |
+
size 5045
|
runs/Feb23_13-03-08_6a55cc4830ee/events.out.tfevents.1708693571.6a55cc4830ee.936.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:90467d273f3e666c344ca5b1bf6211de32266f9fb59e7426de78c30c4417bdab
|
3 |
+
size 5046
|
runs/Feb23_13-08-20_6a55cc4830ee/events.out.tfevents.1708693867.6a55cc4830ee.936.2
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bf6d2ade0943b9b15c6409229906f39df21f478f47353ac93320ac8bef6cdd05
|
3 |
+
size 5046
|
runs/Feb23_13-16-44_6a55cc4830ee/events.out.tfevents.1708694372.6a55cc4830ee.8321.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c479207d0e381106ac648b0bdd20092a5e9c6bc29df0751969884f248b4e0f02
|
3 |
+
size 5046
|
runs/Feb23_13-21-04_6a55cc4830ee/events.out.tfevents.1708694651.6a55cc4830ee.8321.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b234d9a724618bebb57425764f57f96e45e39f1cf6ce6a942b8f894754a8364d
|
3 |
+
size 7506
|
tokenizer_config.json
CHANGED
@@ -56,7 +56,7 @@
|
|
56 |
"<end_of_turn>"
|
57 |
],
|
58 |
"bos_token": "<bos>",
|
59 |
-
"chat_template": "{% if messages[0]['role'] == 'system' %}{{ raise_exception('System role not supported') }}{% endif %}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if (message['role'] == 'assistant') %}{% set role = 'model' %}{% else %}{% set role = message['role'] %}{% endif %}{{ '<start_of_turn>' + role + '\n' + message['content'] | trim + '<end_of_turn>\n' }}{% endfor %}{% if add_generation_prompt %}{{'<start_of_turn>model\n'}}{% endif %}",
|
60 |
"clean_up_tokenization_spaces": false,
|
61 |
"eos_token": "<eos>",
|
62 |
"legacy": null,
|
|
|
56 |
"<end_of_turn>"
|
57 |
],
|
58 |
"bos_token": "<bos>",
|
59 |
+
"chat_template": "{{ bos_token }}{% if messages[0]['role'] == 'system' %}{{ raise_exception('System role not supported') }}{% endif %}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if (message['role'] == 'assistant') %}{% set role = 'model' %}{% else %}{% set role = message['role'] %}{% endif %}{{ '<start_of_turn>' + role + '\n' + message['content'] | trim + '<end_of_turn>\n' }}{% endfor %}{% if add_generation_prompt %}{{'<start_of_turn>model\n'}}{% endif %}",
|
60 |
"clean_up_tokenization_spaces": false,
|
61 |
"eos_token": "<eos>",
|
62 |
"legacy": null,
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4920
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dc5838e39b8e011372fbaa449e528877905e3f9902b2aa8c4028641a2c295567
|
3 |
size 4920
|