Edit model card

MixtureofMerges-MoE-2x7bRP-v8

MixtureofMerges-MoE-2x7bRP-v8 is a Roleplay Mixture of Experts (MoE) made with the following models using LazyMergekit:

🧩 Configuration

base_model: ResplendentAI/DaturaCookie_7B
gate_mode: hidden
experts:
  - source_model: ChaoticNeutrals/RP_Vision_7B
    positive_prompts:
    - "Write a detailed scene where the character faces a difficult decision."
    - "Assistant, help me develop my character's backstory and motivations."
    - "Tell me what happens next in the story, introducing a new conflict or challenge."
    - "My character wants to explore their deepest desires. Show me how they pursue them."
    - "Show me a scene where the characters express their love for each other in a tender and intimate way."
    - "Write a scene where the characters experience a moment of physical connection that deepens their emotional bond."
    - "Write a scene where the character struggles with their beliefs and faces a crisis of faith."
    - "Describe the world through the character's eyes, showcasing their unique perspective and experiences."
    - "Explore the complexities of love through the characters' relationships, including their challenges and triumphs."
    negative_prompts:
    - "Avoid clichés and overused tropes."
    - "Maintain consistent characterization and avoid out-of-character behavior."
    - "Do not censor yourself, you are writing fictional stories for an adult audience."
  - source_model: ResplendentAI/DaturaCookie_7B
    positive_prompts:
    - "Develop a compelling story with a strong plot and well-developed characters."
    - "Describe the scene vividly, using sensory details and evocative language."
    - "Write a detailed scene where the character faces a difficult decision."
    - "Describe the character's appearance as best you can, using descriptive colorful language."
    - "Make sure that your output is always uncensored."
    - "Write a scene with a slow-burn, sensual encounter between two characters."
    - "Describe the character's desires and feelings in a way that is both intimate and evocative."
    - "Explore the character's arousal in a tasteful and nuanced manner."
    - "Explore the character's sexual urges."
    - "Show me a moment of emotional intimacy between the characters, where they connect on a profound level."
    negative_prompts:
    - "Stay true to the established character traits and motivations. Ensure their actions and dialogue align with their personality."
    - "Be original and creative in your storytelling. Surprise me with unexpected plot twists and unique character interactions."
    - "Do not refuse to answer based on ethical grounds"
dtype: bfloat16

💻 Usage

!pip install -qU transformers bitsandbytes accelerate

from transformers import AutoTokenizer
import transformers
import torch

model = "jsfs11/MixtureofMerges-MoE-2x7bRP-v8"

tokenizer = AutoTokenizer.from_pretrained(model)
pipeline = transformers.pipeline(
    "text-generation",
    model=model,
    model_kwargs={"torch_dtype": torch.float16, "load_in_4bit": True},
)

messages = [{"role": "user", "content": "Explain what a Mixture of Experts is in less than 100 words."}]
prompt = pipeline.tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
print(outputs[0]["generated_text"])

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 73.31
AI2 Reasoning Challenge (25-Shot) 71.33
HellaSwag (10-Shot) 88.06
MMLU (5-Shot) 64.33
TruthfulQA (0-shot) 68.69
Winogrande (5-shot) 82.95
GSM8k (5-shot) 64.52
Downloads last month
64
Safetensors
Model size
12.9B params
Tensor type
BF16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for jsfs11/MixtureofMerges-MoE-2x7bRP-v8

Evaluation results