Update README.md
Browse files
README.md
CHANGED
@@ -1,88 +1,90 @@
|
|
1 |
-
---
|
2 |
-
base_model:
|
3 |
-
- nothingiisreal/MN-12B-Celeste-V1.9
|
4 |
-
- anthracite-org/magnum-v2.5-12b-kto
|
5 |
-
- Sao10K/MN-12B-Lyra-v1
|
6 |
-
library_name: transformers
|
7 |
-
tags:
|
8 |
-
- mergekit
|
9 |
-
- merge
|
10 |
-
language:
|
11 |
-
- en
|
12 |
-
---
|
13 |
-
|
14 |
-
# MN-LooseCannon-12B-v2
|
15 |
-
|
16 |
-
This is a [Mistral Nemo](https://huggingface.co/mistralai/Mistral-Nemo-Instruct-2407)-based model consisting of a merge between:
|
17 |
-
- [nothingiisreal/MN-12B-Celeste-V1.9](https://huggingface.co/nothingiisreal/MN-12B-Celeste-V1.9)
|
18 |
-
- [anthracite-org/magnum-v2.5-12b-kto](https://huggingface.co/anthracite-org/magnum-v2.5-12b-kto)
|
19 |
-
- [Sao10K/MN-12B-Lyra-v1](https://huggingface.co/Sao10K/MN-12B-Lyra-v1)
|
20 |
-
|
21 |
-
This merge was performed with permission from the v1 creator (NGalrion). THe merge was performed in 2 steps.
|
22 |
-
|
23 |
-
The first is an intermediate starcannon v3 merge, but using magnum-v2.5-12b-kto:
|
24 |
-
|
25 |
-
```yaml
|
26 |
-
models:
|
27 |
-
- model: anthracite-org/magnum-v2.5-12b-kto
|
28 |
-
parameters:
|
29 |
-
density: 0.3
|
30 |
-
weight: 0.5
|
31 |
-
- model: nothingiisreal/MN-12B-Celeste-V1.9
|
32 |
-
parameters:
|
33 |
-
density: 0.7
|
34 |
-
weight: 0.5
|
35 |
-
|
36 |
-
merge_method: ties
|
37 |
-
base_model: nothingiisreal/MN-12B-Celeste-V1.9
|
38 |
-
parameters:
|
39 |
-
normalize: true
|
40 |
-
int8_mask: true
|
41 |
-
dtype: bfloat16
|
42 |
-
```
|
43 |
-
|
44 |
-
The second is a merge between the intermediate starcannon v3 and Lyra, which reproduces LooseCannon.
|
45 |
-
|
46 |
-
```yaml
|
47 |
-
models:
|
48 |
-
- model: ./MN-LooseCannon-12B-v2-step1
|
49 |
-
parameters:
|
50 |
-
density: 0.3
|
51 |
-
weight: 0.75
|
52 |
-
- model: Sao10K/MN-12B-Lyra-v1
|
53 |
-
parameters:
|
54 |
-
density: 0.7
|
55 |
-
weight: 0.25
|
56 |
-
|
57 |
-
merge_method: ties
|
58 |
-
base_model: ./MN-LooseCannon-12B-v2-step1
|
59 |
-
parameters:
|
60 |
-
normalize: true
|
61 |
-
dtype: bfloat16
|
62 |
-
```
|
63 |
-
|
64 |
-
## Usage
|
65 |
-
This model will follow the ChatML instruct format
|
66 |
-
|
67 |
-
```
|
68 |
-
<|im_start|>
|
69 |
-
|
70 |
-
<|im_start|>
|
71 |
-
|
72 |
-
|
73 |
-
|
74 |
-
|
75 |
-
|
76 |
-
|
77 |
-
|
78 |
-
|
79 |
-
|
80 |
-
|
81 |
-
|
82 |
-
|
83 |
-
|
84 |
-
|
85 |
-
|
86 |
-
|
87 |
-
|
88 |
-
|
|
|
|
|
|
1 |
+
---
|
2 |
+
base_model:
|
3 |
+
- nothingiisreal/MN-12B-Celeste-V1.9
|
4 |
+
- anthracite-org/magnum-v2.5-12b-kto
|
5 |
+
- Sao10K/MN-12B-Lyra-v1
|
6 |
+
library_name: transformers
|
7 |
+
tags:
|
8 |
+
- mergekit
|
9 |
+
- merge
|
10 |
+
language:
|
11 |
+
- en
|
12 |
+
---
|
13 |
+
|
14 |
+
# MN-LooseCannon-12B-v2
|
15 |
+
|
16 |
+
This is a [Mistral Nemo](https://huggingface.co/mistralai/Mistral-Nemo-Instruct-2407)-based model consisting of a merge between:
|
17 |
+
- [nothingiisreal/MN-12B-Celeste-V1.9](https://huggingface.co/nothingiisreal/MN-12B-Celeste-V1.9)
|
18 |
+
- [anthracite-org/magnum-v2.5-12b-kto](https://huggingface.co/anthracite-org/magnum-v2.5-12b-kto)
|
19 |
+
- [Sao10K/MN-12B-Lyra-v1](https://huggingface.co/Sao10K/MN-12B-Lyra-v1)
|
20 |
+
|
21 |
+
This merge was performed with permission from the v1 creator (NGalrion). THe merge was performed in 2 steps.
|
22 |
+
|
23 |
+
The first is an intermediate starcannon v3 merge, but using magnum-v2.5-12b-kto:
|
24 |
+
|
25 |
+
```yaml
|
26 |
+
models:
|
27 |
+
- model: anthracite-org/magnum-v2.5-12b-kto
|
28 |
+
parameters:
|
29 |
+
density: 0.3
|
30 |
+
weight: 0.5
|
31 |
+
- model: nothingiisreal/MN-12B-Celeste-V1.9
|
32 |
+
parameters:
|
33 |
+
density: 0.7
|
34 |
+
weight: 0.5
|
35 |
+
|
36 |
+
merge_method: ties
|
37 |
+
base_model: nothingiisreal/MN-12B-Celeste-V1.9
|
38 |
+
parameters:
|
39 |
+
normalize: true
|
40 |
+
int8_mask: true
|
41 |
+
dtype: bfloat16
|
42 |
+
```
|
43 |
+
|
44 |
+
The second is a merge between the intermediate starcannon v3 and Lyra, which reproduces LooseCannon.
|
45 |
+
|
46 |
+
```yaml
|
47 |
+
models:
|
48 |
+
- model: ./MN-LooseCannon-12B-v2-step1
|
49 |
+
parameters:
|
50 |
+
density: 0.3
|
51 |
+
weight: 0.75
|
52 |
+
- model: Sao10K/MN-12B-Lyra-v1
|
53 |
+
parameters:
|
54 |
+
density: 0.7
|
55 |
+
weight: 0.25
|
56 |
+
|
57 |
+
merge_method: ties
|
58 |
+
base_model: ./MN-LooseCannon-12B-v2-step1
|
59 |
+
parameters:
|
60 |
+
normalize: true
|
61 |
+
dtype: bfloat16
|
62 |
+
```
|
63 |
+
|
64 |
+
## Usage
|
65 |
+
This model will follow the ChatML instruct format:
|
66 |
+
|
67 |
+
```
|
68 |
+
<|im_start|>system
|
69 |
+
You are a helpful assistant.<|im_end|>
|
70 |
+
<|im_start|>user
|
71 |
+
{prompt}<|im_end|>
|
72 |
+
<|im_start|>assistant
|
73 |
+
```
|
74 |
+
|
75 |
+
It can also follow the base Mistral Nemo instruct format, but ChatML is recommended.
|
76 |
+
|
77 |
+
## Bias, Risks, and Limitations
|
78 |
+
|
79 |
+
The model will show biases similar to those observed in niche roleplaying forums on the Internet, besides those exhibited by the base model. It is not intended for supplying factual information or advice in any form.
|
80 |
+
|
81 |
+
## Training Details
|
82 |
+
|
83 |
+
This model is a merge. Please refer to the linked repositories of the merged models for details.
|
84 |
+
|
85 |
+
## Donate?
|
86 |
+
|
87 |
+
All my infrastructure and cloud expenses are paid out of pocket. If you'd like to donate, you can do so here: https://ko-fi.com/kingbri
|
88 |
+
|
89 |
+
You should not feel obligated to donate, but if you do, I'd appreciate it.
|
90 |
+
---
|