File size: 2,403 Bytes
3435b27
 
 
 
 
 
 
43a764d
 
 
779d9b1
 
4830f99
2ef1e4e
18730ee
 
3435b27
2ab3371
3435b27
18730ee
3435b27
 
 
 
18730ee
3435b27
 
18730ee
322f49c
f535aac
322f49c
18730ee
3435b27
 
 
888baa8
3435b27
f2daf9e
6806840
450e58f
6806840
450e58f
3435b27
7787729
 
3435b27
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2423354
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
---
base_model: []
library_name: transformers
tags:
- mergekit
- merge
---

![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/6550b16f7490049d6237f200/WcVRcgD6jSzZbQR3LIQpJ.jpeg)

![image/png](https://cdn-uploads.huggingface.co/production/uploads/6550b16f7490049d6237f200/MizKCCYxd29iH_3cEYq2T.png)

# V4.0 is the best one, use that one.

## Information
### Description

My main goal is to merge the smartness of the base Instruct Nemo with the better prose from the different roleplaying fine-tunes. This is version v0.2, still to be tested. Not sure if it's better than v1.0. All credits and thanks go to Intervitens, Mistralai, and NeverSleep for providing amazing models used in the merge.

### Instruct

Mistral Instruct.

```
<s>[INST] {system} [/INST]{assistant}</s>[INST] {user} [/INST]
```

### Settings

Lower Temperature of 0.35 recommended, although I had luck with Temperatures above one (1.0-1.2) if you crank up the Min P (0.01-0.1). Run with base DRY of 0.8/1.75/2/0 and you're good to go.

### GGUF

https://huggingface.co/MarinaraSpaghetti/Nemomix-v2.0-12B-GGUF

### Other Versions

V1: https://huggingface.co/MarinaraSpaghetti/Nemomix-v1.0-12B

V2: https://huggingface.co/MarinaraSpaghetti/Nemomix-v2.0-12B

V3: https://huggingface.co/MarinaraSpaghetti/Nemomix-v3.0-12B

V4: https://huggingface.co/MarinaraSpaghetti/Nemomix-v4.0-12B

# Nemomix-v2.0-12B

This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).

## Merge Details
### Merge Method

This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using F:\mergekit\mistralaiMistral-Nemo-Base-2407 as a base.

### Models Merged

The following models were included in the merge:
* F:\mergekit\intervitens_mini-magnum-12b-v1.1
* F:\mergekit\NeverSleep_Lumimaid-v0.2-12B
* F:\mergekit\mistralaiMistral-Nemo-Instruct-2407

### Configuration

The following YAML configuration was used to produce this model:

```yaml
models:
  - model: F:\mergekit\NeverSleep_Lumimaid-v0.2-12B
  - model: F:\mergekit\intervitens_mini-magnum-12b-v1.1
  - model: F:\mergekit\mistralaiMistral-Nemo-Instruct-2407
merge_method: model_stock
base_model: F:\mergekit\mistralaiMistral-Nemo-Base-2407
parameters:
  filter_wise: false
dtype: bfloat16
```

## Ko-fi
### Enjoying what I do? Consider donating here, thank you!
https://ko-fi.com/spicy_marinara