Update 2024-01-21
Due to mlabonne/NeuralMarcoro14-7B updating its license to CC-BY-NC, our license will follow suit.
Model Description
This is an experiment to test merging 14 models using DARE TIES 🦙
- We first merge 14 models to produce EmbeddedLLM/Mistral-7B-Merge-14-v0.3.
- The model is merged again using DARE TIES with:
Open LLM Leaderboard
Average | 71.96 |
---|---|
ARC | 68.69 |
HellaSwag | 86.45 |
MMLU | 65.65 |
TruthfulQA | 59.12 |
Winogrande | 80.66 |
GSM8K | 71.19 |
Chat Template
Either ChatML or Llama-2 chat template.
Merge Configuration
The merge config file for this model is here:
models:
- model: mistralai/Mistral-7B-v0.1
# no parameters necessary for base model
- model: EmbeddedLLM/Mistral-7B-Merge-14-v0.3
parameters:
weight: 0.3
density: 0.5
- model: Weyaxi/OpenHermes-2.5-neural-chat-v3-3-openchat-3.5-1210-Slerp
parameters:
weight: 0.2
density: 0.5
- model: openchat/openchat-3.5-0106
parameters:
weight: 0.2
density: 0.5
- model: mlabonne/NeuralMarcoro14-7B
parameters:
weight: 0.3
density: 0.5
merge_method: dare_ties
base_model: mistralai/Mistral-7B-v0.1
parameters:
int8_mask: true
tokenizer_source: union
dtype: bfloat16
- Downloads last month
- 573
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.