File size: 2,403 Bytes
3435b27 43a764d 779d9b1 4830f99 2ef1e4e 18730ee 3435b27 2ab3371 3435b27 18730ee 3435b27 18730ee 3435b27 18730ee 322f49c f535aac 322f49c 18730ee 3435b27 888baa8 3435b27 f2daf9e 6806840 450e58f 6806840 450e58f 3435b27 7787729 3435b27 2423354 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 |
---
base_model: []
library_name: transformers
tags:
- mergekit
- merge
---
![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/6550b16f7490049d6237f200/WcVRcgD6jSzZbQR3LIQpJ.jpeg)
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6550b16f7490049d6237f200/MizKCCYxd29iH_3cEYq2T.png)
# V4.0 is the best one, use that one.
## Information
### Description
My main goal is to merge the smartness of the base Instruct Nemo with the better prose from the different roleplaying fine-tunes. This is version v0.2, still to be tested. Not sure if it's better than v1.0. All credits and thanks go to Intervitens, Mistralai, and NeverSleep for providing amazing models used in the merge.
### Instruct
Mistral Instruct.
```
<s>[INST] {system} [/INST]{assistant}</s>[INST] {user} [/INST]
```
### Settings
Lower Temperature of 0.35 recommended, although I had luck with Temperatures above one (1.0-1.2) if you crank up the Min P (0.01-0.1). Run with base DRY of 0.8/1.75/2/0 and you're good to go.
### GGUF
https://huggingface.co/MarinaraSpaghetti/Nemomix-v2.0-12B-GGUF
### Other Versions
V1: https://huggingface.co/MarinaraSpaghetti/Nemomix-v1.0-12B
V2: https://huggingface.co/MarinaraSpaghetti/Nemomix-v2.0-12B
V3: https://huggingface.co/MarinaraSpaghetti/Nemomix-v3.0-12B
V4: https://huggingface.co/MarinaraSpaghetti/Nemomix-v4.0-12B
# Nemomix-v2.0-12B
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
## Merge Details
### Merge Method
This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using F:\mergekit\mistralaiMistral-Nemo-Base-2407 as a base.
### Models Merged
The following models were included in the merge:
* F:\mergekit\intervitens_mini-magnum-12b-v1.1
* F:\mergekit\NeverSleep_Lumimaid-v0.2-12B
* F:\mergekit\mistralaiMistral-Nemo-Instruct-2407
### Configuration
The following YAML configuration was used to produce this model:
```yaml
models:
- model: F:\mergekit\NeverSleep_Lumimaid-v0.2-12B
- model: F:\mergekit\intervitens_mini-magnum-12b-v1.1
- model: F:\mergekit\mistralaiMistral-Nemo-Instruct-2407
merge_method: model_stock
base_model: F:\mergekit\mistralaiMistral-Nemo-Base-2407
parameters:
filter_wise: false
dtype: bfloat16
```
## Ko-fi
### Enjoying what I do? Consider donating here, thank you!
https://ko-fi.com/spicy_marinara |