Proverbial1 commited on
Commit
cedd0c2
1 Parent(s): e179096

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +165 -0
README.md ADDED
@@ -0,0 +1,165 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ license_name: mrl
4
+ language:
5
+ - en
6
+ tags:
7
+ - chat
8
+ pipeline_tag: text-generation
9
+ library_name: transformers
10
+ ---
11
+
12
+
13
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/658a46cbfb9c2bdfae75b3a6/PeLc_rlHB98Hw4eojizIi.png)
14
+
15
+
16
+ This is a series of models designed to replicate the prose quality of the Claude 3 models, specifically Sonnet and Opus.
17
+
18
+
19
+
20
+ This model is fine-tuned on top of [mistralai/Mistral-Large-Instruct-2407](https://huggingface.co/mistralai/Mistral-Large-Instruct-2407).
21
+
22
+ ## Prompting
23
+ A typical input would look like this:
24
+
25
+ ```py
26
+ <s>[INST] SYSTEM MESSAGE\nUSER MESSAGE[/INST] ASSISTANT MESSAGE</s>[INST] USER MESSAGE[/INST]
27
+ ```
28
+
29
+ ## SillyTavern templates
30
+
31
+ Below are Instruct and Context templates for use within SillyTavern.
32
+
33
+ <details><summary>context template</summary>
34
+
35
+ ```yaml
36
+ default SillyTavern template works fine
37
+ ```
38
+
39
+ </details><br>
40
+ <details><summary>instruct template</summary>
41
+
42
+ ```yaml
43
+ default SillyTavern template works fine
44
+ ```
45
+
46
+ </details><br>
47
+
48
+ ## Axolotl config
49
+
50
+ <details><summary>See axolotl config</summary>
51
+
52
+ ```yaml
53
+ base_model: mistralai/Mistral-Large-Instruct-2407
54
+ model_type: AutoModelForCausalLM
55
+ tokenizer_type: AutoTokenizer
56
+
57
+ plugins:
58
+ - axolotl.integrations.liger.LigerPlugin
59
+ liger_rope: true
60
+ liger_rms_norm: true
61
+ liger_swiglu: true
62
+ liger_fused_linear_cross_entropy: true
63
+
64
+ load_in_8bit: false
65
+ load_in_4bit: false
66
+ strict: false
67
+
68
+ datasets:
69
+ - path: anthracite-org/c2_logs_16k_mistral-large_v1.2
70
+ type: sharegpt
71
+ conversation: mistral
72
+ - path: anthracite-org/kalo-opus-instruct-22k-no-refusal
73
+ type: sharegpt
74
+ conversation: mistral
75
+ - path: lodrick-the-lafted/kalo-opus-instruct-3k-filtered
76
+ type: sharegpt
77
+ conversation: mistral
78
+ - path: anthracite-org/nopm_claude_writing_fixed
79
+ type: sharegpt
80
+ conversation: mistral
81
+ - path: anthracite-org/kalo_opus_misc_240827
82
+ type: sharegpt
83
+ conversation: mistral
84
+ - path: anthracite-org/kalo_misc_part2
85
+ type: sharegpt
86
+ conversation: mistral
87
+ #chat_template: chatml
88
+ shuffle_merged_datasets: true
89
+ #default_system_message: "You are an assistant that responds to the user."
90
+ dataset_prepared_path: ./data/magnum-123b-data
91
+ val_set_size: 0.0
92
+ output_dir: ./data/123b-fft-out
93
+
94
+ sequence_len: 16384
95
+ sample_packing: true
96
+ pad_to_sequence_len: true
97
+
98
+ adapter:
99
+ lora_model_dir:
100
+ lora_r:
101
+ lora_alpha:
102
+ lora_dropout:
103
+ lora_target_linear:
104
+ lora_fan_in_fan_out:
105
+
106
+ wandb_project: 123b-magnum-fft
107
+ wandb_entity:
108
+ wandb_watch:
109
+ wandb_name: alter-attempt-04
110
+ wandb_log_model:
111
+
112
+ gradient_accumulation_steps: 4
113
+ micro_batch_size: 1
114
+ num_epochs: 2
115
+ optimizer: adamw_bnb_8bit
116
+ lr_scheduler: cosine
117
+ learning_rate: 0.0000015
118
+
119
+ train_on_inputs: false
120
+ group_by_length: false
121
+ bf16: auto
122
+ fp16:
123
+ tf32: false
124
+
125
+ gradient_checkpointing: unsloth
126
+ early_stopping_patience:
127
+ resume_from_checkpoint:
128
+ local_rank:
129
+ logging_steps: 1
130
+ xformers_attention:
131
+ flash_attention: true
132
+
133
+ warmup_steps: 40
134
+ evals_per_epoch:
135
+ eval_table_size:
136
+ eval_max_new_tokens:
137
+ saves_per_epoch: 2
138
+ debug:
139
+ deepspeed: deepspeed_configs/zero3_bf16.json
140
+ weight_decay: 0.01
141
+ fsdp:
142
+ fsdp_config:
143
+ special_tokens:
144
+ ```
145
+ </details><br>
146
+
147
+ ## Credits
148
+ We'd like to thank [Eric Hartford](https://huggingface.co/ehartford) for sponsoring the compute for this train.
149
+ We would also like to thank all members of Anthracite who made this finetune possible.
150
+
151
+ ## Datasets
152
+ - [anthracite-org/c2_logs_16k_mistral-large_v1.2](https://huggingface.co/datasets/anthracite-org/c2_logs_16k_mistral-large_v1.2)
153
+ - [anthracite-org/kalo-opus-instruct-22k-no-refusal](https://huggingface.co/datasets/anthracite-org/kalo-opus-instruct-22k-no-refusal)
154
+ - [lodrick-the-lafted/kalo-opus-instruct-3k-filtered](https://huggingface.co/datasets/lodrick-the-lafted/kalo-opus-instruct-3k-filtered)
155
+ - [anthracite-org/nopm_claude_writing_fixed](https://huggingface.co/datasets/anthracite-org/nopm_claude_writing_fixed)
156
+ - [anthracite-org/kalo_opus_misc_240827](https://huggingface.co/datasets/anthracite-org/kalo_opus_misc_240827)
157
+ - [anthracite-org/kalo_misc_part2](https://huggingface.co/datasets/anthracite-org/kalo_misc_part2)
158
+
159
+ ## Training
160
+ We used 8x mi300x GPUs graciously provided by [Eric Hartford](https://huggingface.co/ehartford) for the full-parameter fine-tuning of the model.
161
+
162
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
163
+
164
+ ## Safety
165
+ ...