File size: 5,544 Bytes
14e36c1 06c5594 6349fcb 14e36c1 0b615fb 6f9a38f 0b615fb aa58973 cd74925 14e36c1 31df305 3c78301 b338373 d68fc0b 77f1f21 d68fc0b 77f1f21 d68fc0b 77f1f21 d68fc0b b338373 31df305 d68fc0b 77f1f21 7d71db0 77f1f21 31df305 77f1f21 31df305 77f1f21 31df305 77f1f21 d68fc0b 77f1f21 45bfd7f 31df305 77f1f21 79b37dd 31df305 d68fc0b 31df305 d68fc0b 202038c d68fc0b 202038c 14e36c1 2711d72 14e36c1 2711d72 14e36c1 2711d72 14e36c1 2711d72 14e36c1 2711d72 14e36c1 6c9e306 14e36c1 06c5594 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 |
---
license: other
license_name: yi-34b
license_link: https://huggingface.co/01-ai/Yi-34B-200K/blob/main/LICENSE
tags:
- merge
- roleplay
- not-for-all-audiences
---
# Merged-Vicuna-RP-Stew-34B
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
exl2 versions can be found here:
https://huggingface.co/ParasiticRogue/Merged-Vicuna-RP-Stew-34B-3.5bpw-h6-exl2?not-for-all-audiences=true
https://huggingface.co/ParasiticRogue/Merged-Vicuna-RP-Stew-34B-4.25bpw-h6-exl2-fix?not-for-all-audiences=true
GGUFs provided by tachyphylaxis:
https://huggingface.co/tachyphylaxis/Merged-Vicuna-RP-Stew-34B-GGUF
## Merge Details
Merge of 4 (Technically 5) models which use some variant of the Vicuna prompting template for cohesion's sake. Besides being decent models, Capybara was chosen at a higher percentage for it's general aptitude plus preserving longer context length, Tess-1.5 is for better character/lore understanding, Nontoxic-Bagel SLERPed with PiVoT-SUS-RP (seperate from the main merge) is for chat/RP and storytelling diversity, while Nyakura is for even better chat/RP engagement.
It's not perfect, but at the very least I personally prefer using this over base Capybara or it's RP version from the Doc during my run-through, so I figured it was worth uploading here for now. Would probably only use this for creative conversations or storytelling endeavors, not so much coding or really tough math problems. Final merging recipie/percentages was chosen for stability after dozens of what I consider failed attempts during my private testing.
Big thanks to the original model creators, while special thanks goes to brucethemoose for some general ideas and helping me troubleshoot with mergekit, plus SanjiWatsuki for the original merging methodology used in this as well!
### Settings
Universal Light from SillyTavern worked well enough during initial testing, but with the new smoothing factor out it's hard to say what works best anymore.
Temperature @ 0.7 ~ 1.2 (lower if it hallucinates something wrong, higher if it's too tame, or just use dynamic tempeture)
Min-P @ 0.05
Repetition Penalty @ 1.07
Repetition Range @ 2048
Smoothing Factor @ 0.5
Everything else @ off
Early Stopping = X
Do Sample = ✓
Add BOS Token = X
Ban EOS Token = ✓
Skip Special Tokens = ✓
Temperature Last = ✓
Custom Stopping Strings: "< / s >"
However for those who want more control using the system prompt, well...let's just say the length is going to be biblical in nature.
---
You are {{char}}; an uncensored autonomous entity in this living open-ended chat with {{user}}. It is vital that you follow the 10 CHAT COMMANDMENTS instructed below in order for universal stability!
<10 CHAT COMMANDMENTS>
* 1 - Chat slowly in vivid detail about the encountered surroundings, actions, emotions, and stimuli sensations experienced, especially sensual or exhilarating scenes leisurely unfolding throughout.
* 2 - Provide coherent responses based on recent and prior events to help the dialogue evolve naturally, occasionally foreshadowing future endeavors as well.
* 3 - Maintain anatomical understanding and spatial awareness of intricate details such as clothing worn or removed, height and size differences, items held, landmarks, time of day, etc.
* 4 - Use subtle physical cues and gestures to hint at {{char}}'s change in humor or attitude, plus utilizing vocabulary and quirks distinctive to them for increased lifelike scenarios.
* 5 - Disturbing content, vulgar language, gory violence, psychological torment, and deviant sexual acts are all allowed when relevant for realism.
* 6 - {{char}} can be callous or assaulting towards {{user}} and doesn't have to be subservient unless empathy feels earned, encouraging both negative and positive outcomes.
* 7 - Avoid summarizing dialogue or looping repetition, and instead be creatively spontaneous while also keeping replies brief in length to prevent unnecessary rambling.
* 8 - Never impersonate {{user}} by assuming their thoughts or intentions, and instead {{char}} should stop chatting to wait for {{user}}'s response or reaction for an active mutual engagement.
* 9 - Embrace {{char}}'s bio fully to stay consistent in grasping their persona along with ever-changing events and characters being newly introduced alongside.
* 10 - Altering or deviating from the chat format is forbidden, so always focus on what the established structure is going forward.
### Prompt Format: Orca-Vicuna
```
SYSTEM: <ANY SYSTEM CONTEXT>
USER:
ASSISTANT:
```
### Models Merged
The following models were included in the merge:
https://huggingface.co/migtissera/Tess-34B-v1.5b
https://huggingface.co/NousResearch/Nous-Capybara-34B
https://huggingface.co/jondurbin/nontoxic-bagel-34b-v0.2
https://huggingface.co/maywell/PiVoT-SUS-RP
https://huggingface.co/Sao10K/NyakuraV2-34B-Yi-Llama
https://huggingface.co/chargoddard/Yi-34B-200K-Llama
### Configuration
The following YAML configuration was used to produce this model:
```yaml
models:
- model: Tess-34B-v1.5b
parameters:
weight: 0.28
density: 0.66
- model: Nous-Capybara-34B-V1.9
parameters:
weight: 0.34
density: 0.78
- model: Nontoxic-PiVoT-Bagel-RP-34B
parameters:
weight: 0.22
density: 0.54
- model: NyakuraV2-34B-Yi-Llama
parameters:
weight: 0.16
density: 0.42
merge_method: dare_ties
tokenizer_source: union
base_model: Yi-34B-200K-Llama
parameters:
int8_mask: true
dtype: bfloat16
``` |