kingbri commited on
Commit
cbc8e32
1 Parent(s): f2e4da6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +90 -88
README.md CHANGED
@@ -1,88 +1,90 @@
1
- ---
2
- base_model:
3
- - nothingiisreal/MN-12B-Celeste-V1.9
4
- - anthracite-org/magnum-v2.5-12b-kto
5
- - Sao10K/MN-12B-Lyra-v1
6
- library_name: transformers
7
- tags:
8
- - mergekit
9
- - merge
10
- language:
11
- - en
12
- ---
13
-
14
- # MN-LooseCannon-12B-v2
15
-
16
- This is a [Mistral Nemo](https://huggingface.co/mistralai/Mistral-Nemo-Instruct-2407)-based model consisting of a merge between:
17
- - [nothingiisreal/MN-12B-Celeste-V1.9](https://huggingface.co/nothingiisreal/MN-12B-Celeste-V1.9)
18
- - [anthracite-org/magnum-v2.5-12b-kto](https://huggingface.co/anthracite-org/magnum-v2.5-12b-kto)
19
- - [Sao10K/MN-12B-Lyra-v1](https://huggingface.co/Sao10K/MN-12B-Lyra-v1)
20
-
21
- This merge was performed with permission from the v1 creator (NGalrion). THe merge was performed in 2 steps.
22
-
23
- The first is an intermediate starcannon v3 merge, but using magnum-v2.5-12b-kto:
24
-
25
- ```yaml
26
- models:
27
- - model: anthracite-org/magnum-v2.5-12b-kto
28
- parameters:
29
- density: 0.3
30
- weight: 0.5
31
- - model: nothingiisreal/MN-12B-Celeste-V1.9
32
- parameters:
33
- density: 0.7
34
- weight: 0.5
35
-
36
- merge_method: ties
37
- base_model: nothingiisreal/MN-12B-Celeste-V1.9
38
- parameters:
39
- normalize: true
40
- int8_mask: true
41
- dtype: bfloat16
42
- ```
43
-
44
- The second is a merge between the intermediate starcannon v3 and Lyra, which reproduces LooseCannon.
45
-
46
- ```yaml
47
- models:
48
- - model: ./MN-LooseCannon-12B-v2-step1
49
- parameters:
50
- density: 0.3
51
- weight: 0.75
52
- - model: Sao10K/MN-12B-Lyra-v1
53
- parameters:
54
- density: 0.7
55
- weight: 0.25
56
-
57
- merge_method: ties
58
- base_model: ./MN-LooseCannon-12B-v2-step1
59
- parameters:
60
- normalize: true
61
- dtype: bfloat16
62
- ```
63
-
64
- ## Usage
65
- This model will follow the ChatML instruct format without the system prompt:
66
-
67
- ```
68
- <|im_start|>user
69
- {prompt}<|im_end|>
70
- <|im_start|>assistant
71
- ```
72
-
73
- It can also follow the base Mistral Nemo instruct format, but ChatML is recommended.
74
-
75
- ## Bias, Risks, and Limitations
76
-
77
- The model will show biases similar to those observed in niche roleplaying forums on the Internet, besides those exhibited by the base model. It is not intended for supplying factual information or advice in any form.
78
-
79
- ## Training Details
80
-
81
- This model is a merge. Please refer to the linked repositories of the merged models for details.
82
-
83
- ## Donate?
84
-
85
- All my infrastructure and cloud expenses are paid out of pocket. If you'd like to donate, you can do so here: https://ko-fi.com/kingbri
86
-
87
- You should not feel obligated to donate, but if you do, I'd appreciate it.
88
- ---
 
 
 
1
+ ---
2
+ base_model:
3
+ - nothingiisreal/MN-12B-Celeste-V1.9
4
+ - anthracite-org/magnum-v2.5-12b-kto
5
+ - Sao10K/MN-12B-Lyra-v1
6
+ library_name: transformers
7
+ tags:
8
+ - mergekit
9
+ - merge
10
+ language:
11
+ - en
12
+ ---
13
+
14
+ # MN-LooseCannon-12B-v2
15
+
16
+ This is a [Mistral Nemo](https://huggingface.co/mistralai/Mistral-Nemo-Instruct-2407)-based model consisting of a merge between:
17
+ - [nothingiisreal/MN-12B-Celeste-V1.9](https://huggingface.co/nothingiisreal/MN-12B-Celeste-V1.9)
18
+ - [anthracite-org/magnum-v2.5-12b-kto](https://huggingface.co/anthracite-org/magnum-v2.5-12b-kto)
19
+ - [Sao10K/MN-12B-Lyra-v1](https://huggingface.co/Sao10K/MN-12B-Lyra-v1)
20
+
21
+ This merge was performed with permission from the v1 creator (NGalrion). THe merge was performed in 2 steps.
22
+
23
+ The first is an intermediate starcannon v3 merge, but using magnum-v2.5-12b-kto:
24
+
25
+ ```yaml
26
+ models:
27
+ - model: anthracite-org/magnum-v2.5-12b-kto
28
+ parameters:
29
+ density: 0.3
30
+ weight: 0.5
31
+ - model: nothingiisreal/MN-12B-Celeste-V1.9
32
+ parameters:
33
+ density: 0.7
34
+ weight: 0.5
35
+
36
+ merge_method: ties
37
+ base_model: nothingiisreal/MN-12B-Celeste-V1.9
38
+ parameters:
39
+ normalize: true
40
+ int8_mask: true
41
+ dtype: bfloat16
42
+ ```
43
+
44
+ The second is a merge between the intermediate starcannon v3 and Lyra, which reproduces LooseCannon.
45
+
46
+ ```yaml
47
+ models:
48
+ - model: ./MN-LooseCannon-12B-v2-step1
49
+ parameters:
50
+ density: 0.3
51
+ weight: 0.75
52
+ - model: Sao10K/MN-12B-Lyra-v1
53
+ parameters:
54
+ density: 0.7
55
+ weight: 0.25
56
+
57
+ merge_method: ties
58
+ base_model: ./MN-LooseCannon-12B-v2-step1
59
+ parameters:
60
+ normalize: true
61
+ dtype: bfloat16
62
+ ```
63
+
64
+ ## Usage
65
+ This model will follow the ChatML instruct format:
66
+
67
+ ```
68
+ <|im_start|>system
69
+ You are a helpful assistant.<|im_end|>
70
+ <|im_start|>user
71
+ {prompt}<|im_end|>
72
+ <|im_start|>assistant
73
+ ```
74
+
75
+ It can also follow the base Mistral Nemo instruct format, but ChatML is recommended.
76
+
77
+ ## Bias, Risks, and Limitations
78
+
79
+ The model will show biases similar to those observed in niche roleplaying forums on the Internet, besides those exhibited by the base model. It is not intended for supplying factual information or advice in any form.
80
+
81
+ ## Training Details
82
+
83
+ This model is a merge. Please refer to the linked repositories of the merged models for details.
84
+
85
+ ## Donate?
86
+
87
+ All my infrastructure and cloud expenses are paid out of pocket. If you'd like to donate, you can do so here: https://ko-fi.com/kingbri
88
+
89
+ You should not feel obligated to donate, but if you do, I'd appreciate it.
90
+ ---