Pavanb commited on
Commit
70c99d5
1 Parent(s): b75a1e8

hiranmai49/llama-2_totto_instructions

Browse files
README.md CHANGED
@@ -41,6 +41,7 @@ The following hyperparameters were used during training:
41
  - lr_scheduler_type: constant
42
  - lr_scheduler_warmup_ratio: 0.03
43
  - training_steps: 500
 
44
 
45
  ### Training results
46
 
@@ -48,7 +49,7 @@ The following hyperparameters were used during training:
48
 
49
  ### Framework versions
50
 
51
- - Transformers 4.33.2
52
  - Pytorch 2.0.0
53
  - Datasets 2.1.0
54
- - Tokenizers 0.13.3
 
41
  - lr_scheduler_type: constant
42
  - lr_scheduler_warmup_ratio: 0.03
43
  - training_steps: 500
44
+ - mixed_precision_training: Native AMP
45
 
46
  ### Training results
47
 
 
49
 
50
  ### Framework versions
51
 
52
+ - Transformers 4.35.2
53
  - Pytorch 2.0.0
54
  - Datasets 2.1.0
55
+ - Tokenizers 0.15.0
adapter_config.json CHANGED
@@ -8,16 +8,19 @@
8
  "init_lora_weights": true,
9
  "layers_pattern": null,
10
  "layers_to_transform": null,
 
11
  "lora_alpha": 16,
12
  "lora_dropout": 0.1,
 
 
13
  "modules_to_save": null,
14
  "peft_type": "LORA",
15
  "r": 64,
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
19
- "q_proj",
20
- "v_proj"
21
  ],
22
  "task_type": "CAUSAL_LM"
23
  }
 
8
  "init_lora_weights": true,
9
  "layers_pattern": null,
10
  "layers_to_transform": null,
11
+ "loftq_config": {},
12
  "lora_alpha": 16,
13
  "lora_dropout": 0.1,
14
+ "megatron_config": null,
15
+ "megatron_core": "megatron.core",
16
  "modules_to_save": null,
17
  "peft_type": "LORA",
18
  "r": 64,
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
+ "v_proj",
23
+ "q_proj"
24
  ],
25
  "task_type": "CAUSAL_LM"
26
  }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:549812a016e194282c7dd96822d57093e7319358443c9117b3ebdc1cf60873d6
3
+ size 134235048
runs/Dec02_10-23-16_a88a7964f57c/events.out.tfevents.1701512646.a88a7964f57c.80.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c01737a65b79db48e88bd2ecf86220d0bd2763db9c04b31a2182634d629951d0
3
+ size 4613
runs/Dec02_10-30-25_a88a7964f57c/events.out.tfevents.1701513035.a88a7964f57c.391.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ce6956ef5b679f90fe83054146aaff480560907cbdfab420a91210d3834f1a3
3
+ size 6531
tokenizer.json CHANGED
@@ -2,7 +2,7 @@
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
- "max_length": 512,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
 
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
+ "max_length": 256,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
tokenizer_config.json CHANGED
@@ -1,34 +1,40 @@
1
  {
2
- "bos_token": {
3
- "__type": "AddedToken",
4
- "content": "<s>",
5
- "lstrip": false,
6
- "normalized": false,
7
- "rstrip": false,
8
- "single_word": false
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9
  },
 
 
10
  "clean_up_tokenization_spaces": false,
11
- "eos_token": {
12
- "__type": "AddedToken",
13
- "content": "</s>",
14
- "lstrip": false,
15
- "normalized": false,
16
- "rstrip": false,
17
- "single_word": false
18
- },
19
  "legacy": false,
20
  "model_max_length": 1000000000000000019884624838656,
21
- "pad_token": null,
22
  "padding_side": "right",
23
  "sp_model_kwargs": {},
24
  "tokenizer_class": "LlamaTokenizer",
25
- "unk_token": {
26
- "__type": "AddedToken",
27
- "content": "<unk>",
28
- "lstrip": false,
29
- "normalized": false,
30
- "rstrip": false,
31
- "single_word": false
32
- },
33
- "use_default_system_prompt": true
34
  }
 
1
  {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<unk>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<s>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "</s>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ }
27
  },
28
+ "bos_token": "<s>",
29
+ "chat_template": "{% if messages[0]['role'] == 'system' %}{% set loop_messages = messages[1:] %}{% set system_message = messages[0]['content'] %}{% else %}{% set loop_messages = messages %}{% set system_message = false %}{% endif %}{% for message in loop_messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if loop.index0 == 0 and system_message != false %}{% set content = '<<SYS>>\\n' + system_message + '\\n<</SYS>>\\n\\n' + message['content'] %}{% else %}{% set content = message['content'] %}{% endif %}{% if message['role'] == 'user' %}{{ bos_token + '[INST] ' + content.strip() + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ ' ' + content.strip() + ' ' + eos_token }}{% endif %}{% endfor %}",
30
  "clean_up_tokenization_spaces": false,
31
+ "eos_token": "</s>",
 
 
 
 
 
 
 
32
  "legacy": false,
33
  "model_max_length": 1000000000000000019884624838656,
34
+ "pad_token": "</s>",
35
  "padding_side": "right",
36
  "sp_model_kwargs": {},
37
  "tokenizer_class": "LlamaTokenizer",
38
+ "unk_token": "<unk>",
39
+ "use_default_system_prompt": false
 
 
 
 
 
 
 
40
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:42b254bfb1cab8007ed7e5f3d53ea1ae197fde91458cd33b42a57427a8718849
3
- size 4027
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e2e63e93839eaa998b5e16cb20ee9e106c022ebcf9dce5f477a2fe77e595fea7
3
+ size 4155