--- license: apache-2.0 pipeline_tag: text-generation tags: - chat - mistral - roleplay - creative-writing base_model: - nbeerbower/mistral-nemo-bophades-12B - anthracite-org/magnum-v2-12b - Sao10K/MN-12B-Lyra-v3 - Gryphe/Pantheon-RP-1.6-12b-Nemo language: - en --- ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6303fa71fc783bfc7443e7ae/c3ddWBoz-lINEykUDCoXy.png) ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6303fa71fc783bfc7443e7ae/hOpgDxJS2sDO7HzuC9e18.png) # StarDust-12b-v2 ## Quants - GGUF: Not Available Yet. - weighted/imatrix GGUF: Not Available Yet. - exl2: [lucyknada/Luni_StarDust-12b-v2-exl2](https://huggingface.co/lucyknada/Luni_StarDust-12b-v2-exl2) ## Description | Usecase - The result of this merge is in my opinion a more vibrant and less generic sonnet inspired prose, it's able to be gentle and harsh where asked. - The v2 uses the non-kto magnum which tends to have less "claudeism" (making the story feel rather repetitive) - Note on Non-Kto: There is a very big gap between people preferring and disliking the KTO. To make things easier, you can still use [Luni/StarDust-12b-v1](https://huggingface.co/Luni/StarDust-12b-v1) which has the KTO version. - In early testing users have reported a much better experience in longer roleplays and a abillity to add a creative touch to the stable experiencve. Just like with v1: - This model is intended to be used as a Role-playing model. - Its direct conversational output is... I can't even say it's luck, it's just not made for it. - Extension to Conversational output: The Model is designed for roleplay, direct instructing or general purpose is NOT recommended. ## Prompting ### Edit: ChatML has proven to be the BEST choice. Both Mistral and ChatML should work though I had better results with ChatML: ChatML Example: ```py """<|im_start|>user Hi there!<|im_end|> <|im_start|>assistant Nice to meet you!<|im_end|> <|im_start|>user Can I ask a question?<|im_end|> <|im_start|>assistant """ ``` ## Merge Details ### Merge Method This model was merged using the [DARE](https://arxiv.org/abs/2311.03099) [TIES](https://arxiv.org/abs/2306.01708) merge method using [Sao10K/MN-12B-Lyra-v3](https://huggingface.co/Sao10K/MN-12B-Lyra-v3) as a base. ### Models Merged The following models were included in the merge: * [nbeerbower/mistral-nemo-bophades-12B](https://huggingface.co/nbeerbower/mistral-nemo-bophades-12B) * [anthracite-org/magnum-v2-12b](https://huggingface.co/anthracite-org/magnum-v2-12b) * [Gryphe/Pantheon-RP-1.6-12b-Nemo](https://huggingface.co/Gryphe/Pantheon-RP-1.6-12b-Nemo) * [Sao10K/MN-12B-Lyra-v3](https://huggingface.co/Sao10K/MN-12B-Lyra-v3) ### Special Thanks Special thanks to the SillyTilly and myself for helping me find the energy to finish this.