--- license: cc-by-nc-4.0 tags: - not-for-all-audiences - nsfw --- [HIGHLY EXPERIMENTAL] Just try it for a good laugh. Need testing. ```shell The plan : Open-Orca/OpenOrcaxOpenChat-Preview2-13B PygmalionAI/pygmalion-2-13b Undi95/MLewd-L2-13B-v2-3 jondurbin/spicyboros-13b-2.2 lemonilia/limarp-llama2-v2 Step 1: Merge OpenOrcaxOpenChat-Preview2-13B with pygmalion-2-13b => OpenOrcaPyg2 Step 2: Merge MLewd with Spicyboros => MLewdBorosPlus Step 3: In the layer side, replace the layer 0 to 8 with MLewd, and the layer 16 to 20 with Spicyboros of the first merge => OpenOrcaPyg2-Layered Step 4: In the layer side, replace the layer 0 to 8 with MLewd, and the layer 16 to 20 with Spicyboros of the second merge => MLewdBorosPlus-Layered Step 5: Merge OpenOrcaPyg2-Layered with MLewdBorosPlus-Layered => OpenRPBase Step 6: Apply Limarp2 at 0.5 weight at the end => OpenRP Goal: making Orca a RP model with Pyg2 dataset and MLewd+Spicyboros 100% layer accross the merge and avoid censoring It will be diluted to ~25% in other layer, SLERP do the dirty job The LoRA is here to redirect to RP writing ``` Don't ask me why this model work. I'm a blind scientist. It seems a little obsessed with the game "Garry's mod" tho. Be patient with him. SuperCOT applied : https://huggingface.co/Undi95/OpenRP-13B-SuperCOT-GGUF