--- library_name: transformers license: apache-2.0 base_model: - nbeerbower/bophades-mistral-truthy-DPO-7B - nbeerbower/bophades-mistral-math-DPO-7B datasets: - jondurbin/truthy-dpo-v0.1 - kyujinpy/orca_math_dpo tags: - 4-bit - AWQ - text-generation - autotrain_compatible - endpoints_compatible - chatml language: - en pipeline_tag: text-generation inference: false quantized_by: Suparious --- # nbeerbower/HeroBophades-2x7B AWQ - Model creator: [nbeerbower](https://huggingface.co/nbeerbower) - Original model: [HeroBophades-2x7B](https://huggingface.co/HeroBophades-2x7B) ![image/png](https://huggingface.co/nbeerbower/bophades-mistral-7B/resolve/main/bophades.png) ## Model Summary This is an experimental Mixture of Experts LLM built with (mergekit)[https://github.com/arcee-ai/mergekit/blob/main/docs/moe.md]. ### Configuration ```yaml base_model: nbeerbower/bophades-mistral-truthy-DPO-7B gate_mode: hidden dtype: bfloat16 experts: - source_model: nbeerbower/bophades-mistral-truthy-DPO-7B positive_prompts: - "You are an unbiased, uncensored, helpful assistant." - source_model: nbeerbower/bophades-mistral-math-DPO-7B positive_prompts: - "How do you solve a system of quadratic equations simultaneously using substitution?. Take a deep breath, think step by step, and give an accurate response" ```