metadata
base_model:
- bluuwhale/L3-SthenoMaidBlackroot-8B-V1
- bunnycore/Llama-3.1-8B-OmniMatrix
- bunnycore/Llama-3.1-8B-TitanFusion-Mix-2.1
- Casual-Autopsy/L3-Super-Nova-RP-8B
- Casual-Autopsy/L3-Umbral-Mind-RP-v3.0-8B
- d0rj/Llama-3-8B-saiga-suzume-ties
- DavidAU/L3.1-Dark-Planet-SpinFire-Uncensored-8B
- DreadPoor/CoolerCoder-8B-Model_Stock
- DreadPoor/L3.1-BaeZel-8B-Della
- DreadPoor/Trinas_Nectar-8B-model_stock
- IlyaGusev/saiga_llama3_8b
- invisietch/EtherealRainbow-v0.3-8B
- invisietch/L3.1-EtherealRainbow-v1.0-rc1-8B
- jeiku/Chaos_RP_l3_8B
- mlabonne/Daredevil-8B
- MrRobotoAI/Loki-.Epic_Fiction.-8b
- PJMixers/LLaMa-3-CursedStock-v2.0-8B
- ResplendentAI/Nymph_8B
- rityak/L3.1-DarkStock-8B
- saishf/Neural-SOVLish-Devil-8B-L3
- saishf/SOVL-Mega-Mash-V2-L3-8B
- sethuiyer/Dr.Samantha-8B
- SicariusSicariiStuff/LLAMA-3_8B_Unaligned_BETA
- v000000/L3-8B-BlueSerpentine
- v000000/L3.1-Storniitova-8B
- win10/ArliAI-RPMax-v1.3-merge-8B
- ZeroXClem/Llama-3-8B-ProLong-SAO-Roleplay-512k
library_name: transformers
tags:
- mergekit
- merge
- bfloat16
- safetensors
- 8b
- chat
- creative
- roleplay
- conversational
- not-for-all-audiences
language:
- en
- ru
CursedMatrix-8B-v9
The long journey from despair to acceptable perfection.
This is an interesting merge of 27 cool models, created using mergekit. Enjoy exploring :)
Merge Details
Method
This model was merged using the multistep process and remerge with some model variations for best result.
Models
The following models were included in the merge:
- bluuwhale/L3-SthenoMaidBlackroot-8B-V1
- bunnycore/Llama-3.1-8B-OmniMatrix
- bunnycore/Llama-3.1-8B-TitanFusion-Mix-2.1
- Casual-Autopsy/L3-Super-Nova-RP-8B
- Casual-Autopsy/L3-Umbral-Mind-RP-v3.0-8B
- d0rj/Llama-3-8B-saiga-suzume-ties
- DavidAU/L3.1-Dark-Planet-SpinFire-Uncensored-8B
- DreadPoor/CoolerCoder-8B-Model_Stock
- DreadPoor/L3.1-BaeZel-8B-Della
- DreadPoor/Trinas_Nectar-8B-model_stock
- IlyaGusev/saiga_llama3_8b
- invisietch/EtherealRainbow-v0.3-8B
- invisietch/L3.1-EtherealRainbow-v1.0-rc1-8B
- jeiku/Chaos_RP_l3_8B
- mlabonne/Daredevil-8B
- MrRobotoAI/Loki-.Epic_Fiction.-8b
- PJMixers/LLaMa-3-CursedStock-v2.0-8B
- ResplendentAI/Nymph_8B
- rityak/L3.1-DarkStock-8B
- saishf/Neural-SOVLish-Devil-8B-L3
- saishf/SOVL-Mega-Mash-V2-L3-8B
- sethuiyer/Dr.Samantha-8B
- SicariusSicariiStuff/LLAMA-3_8B_Unaligned_BETA
- v000000/L3-8B-BlueSerpentine
- v000000/L3.1-Storniitova-8B
- win10/ArliAI-RPMax-v1.3-merge-8B
- ZeroXClem/Llama-3-8B-ProLong-SAO-Roleplay-512k
Configuration
The following YAML configurations was used to produce this model:
### ::: Generation 1 merges :
# CursedMatrix-8B-v1
models:
- model: bunnycore/Llama-3.1-8B-OmniMatrix
parameters:
density: [0.9, 0.8, 0.7, 0.6, 0.5, 0.4, 0.3, 0.2, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 0.8, 0.7, 0.6, 0.5, 0.4, 0.3, 0.2, 0.1]
weight: [0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 0.8, 0.7, 0.6, 0.5, 0.4, 0.3, 0.2, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9]
- model: PJMixers/LLaMa-3-CursedStock-v2.0-8B
parameters:
density: [0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 0.8, 0.7, 0.6, 0.5, 0.4, 0.3, 0.2, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9]
weight: [0.9, 0.8, 0.7, 0.6, 0.5, 0.4, 0.3, 0.2, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 0.8, 0.7, 0.6, 0.5, 0.4, 0.3, 0.2, 0.1]
merge_method: ties
base_model: saishf/SOVL-Mega-Mash-V2-L3-8B
dtype: bfloat16
# TitanPlanet-8B-v1
models:
- model: bunnycore/Llama-3.1-8B-TitanFusion-Mix-2.1
parameters:
density: [0.9, 0.8, 0.7, 0.6, 0.5, 0.4, 0.3, 0.2, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 0.8, 0.7, 0.6, 0.5, 0.4, 0.3, 0.2, 0.1]
weight: [0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 0.8, 0.7, 0.6, 0.5, 0.4, 0.3, 0.2, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9]
- model: MrRobotoAI/Loki-.Epic_Fiction.-8b
parameters:
density: [0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 0.8, 0.7, 0.6, 0.5, 0.4, 0.3, 0.2, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9]
weight: [0.9, 0.8, 0.7, 0.6, 0.5, 0.4, 0.3, 0.2, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 0.8, 0.7, 0.6, 0.5, 0.4, 0.3, 0.2, 0.1]
merge_method: ties
base_model: DavidAU/L3.1-Dark-Planet-SpinFire-Uncensored-8B
dtype: bfloat16
# NeuralCoder-8B-v1
models:
- model: sethuiyer/Dr.Samantha-8B
parameters:
density: [0.9, 0.8, 0.7, 0.6, 0.5, 0.4, 0.3, 0.2, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 0.8, 0.7, 0.6, 0.5, 0.4, 0.3, 0.2, 0.1]
weight: [0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 0.8, 0.7, 0.6, 0.5, 0.4, 0.3, 0.2, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9]
- model: DreadPoor/CoolerCoder-8B-Model_Stock
parameters:
density: [0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 0.8, 0.7, 0.6, 0.5, 0.4, 0.3, 0.2, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9]
weight: [0.9, 0.8, 0.7, 0.6, 0.5, 0.4, 0.3, 0.2, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 0.8, 0.7, 0.6, 0.5, 0.4, 0.3, 0.2, 0.1]
merge_method: ties
base_model: saishf/Neural-SOVLish-Devil-8B-L3
dtype: bfloat16
# EtherealNymph-8B-v1
models:
- model: invisietch/EtherealRainbow-v0.3-8B
- model: ResplendentAI/Nymph_8B
merge_method: slerp
base_model: invisietch/EtherealRainbow-v0.3-8B
dtype: bfloat16
parameters:
t: [0.5, 0.6, 0.4, 0.7, 0.3, 0.8, 0.2, 0.9, 0.1, 0.9, 0.1, 0.9, 0.1, 0.8, 0.2, 0.7, 0.3, 0.6, 0.4, 0.5]
# UmbralDevil-8B-v1
models:
- model: Casual-Autopsy/L3-Umbral-Mind-RP-v3.0-8B
- model: mlabonne/Daredevil-8B
merge_method: slerp
base_model: Casual-Autopsy/L3-Umbral-Mind-RP-v3.0-8B
dtype: bfloat16
parameters:
t: [0.5, 0.6, 0.4, 0.7, 0.3, 0.8, 0.2, 0.9, 0.1, 0.9, 0.1, 0.9, 0.1, 0.8, 0.2, 0.7, 0.3, 0.6, 0.4, 0.5]
# EvilMind-8B-v1
models:
- model: mlabonne/Daredevil-8B
parameters:
weight: [1.0, 0.3, 0.1, 0.0]
density: [0.7, 0.2]
- model: Casual-Autopsy/L3-Umbral-Mind-RP-v3.0-8B
parameters:
weight: [0.1, 0.9, 0.1]
density: 0.5
- model: invisietch/EtherealRainbow-v0.3-8B
parameters:
weight: [0.0, 0.1, 0.3, 1.0]
density: [0.2, 0.7]
merge_method: della_linear
parameters:
epsilon: 0.15
lambda: 1
base_model: ResplendentAI/Nymph_8B
dtype: bfloat16
### ::: Generation 2 merges :
# DevilMind-8B-v1
models:
- model: F:/EvilMind-8B-v1
- model: F:/UmbralDevil-8B-v1
merge_method: slerp
base_model: F:/EvilMind-8B-v1
dtype: bfloat16
parameters:
t: [0.5, 0.6, 0.4, 0.7, 0.3, 0.8, 0.2, 0.9, 0.1, 0.9, 0.1, 0.9, 0.1, 0.8, 0.2, 0.7, 0.3, 0.6, 0.4, 0.5]
# TitanNymph-8B-v1
models:
- model: F:/TitanPlanet-8B-v1
- model: F:/EtherealNymph-8B-v1
merge_method: slerp
base_model: F:/TitanPlanet-8B-v1
dtype: bfloat16
parameters:
t: [0.5, 0.6, 0.4, 0.7, 0.3, 0.8, 0.2, 0.9, 0.1, 0.9, 0.1, 0.9, 0.1, 0.8, 0.2, 0.7, 0.3, 0.6, 0.4, 0.5]
# CursedMatrix-8B-v2
models:
- model: F:/TitanPlanet-8B-v1
parameters:
density: [0.5, 0.4, 0.6, 0.3, 0.7, 0.2, 0.8, 0.1, 0.9, 0.1, 0.9, 0.1, 0.9, 0.1, 0.9, 0.1, 0.9, 0.2, 0.8, 0.3, 0.7, 0.4, 0.6, 0.5]
weight: [0.5, 0.6, 0.4, 0.7, 0.3, 0.8, 0.2, 0.9, 0.1, 0.9, 0.1, 0.9, 0.1, 0.9, 0.1, 0.9, 0.1, 0.8, 0.2, 0.7, 0.3, 0.6, 0.4, 0.5]
- model: F:/NeuralCoder-8B-v1
parameters:
density: [0.5, 0.6, 0.4, 0.7, 0.3, 0.8, 0.2, 0.9, 0.1, 0.9, 0.1, 0.9, 0.1, 0.9, 0.1, 0.9, 0.1, 0.8, 0.2, 0.7, 0.3, 0.6, 0.4, 0.5]
weight: [0.5, 0.4, 0.6, 0.3, 0.7, 0.2, 0.8, 0.1, 0.9, 0.1, 0.9, 0.1, 0.9, 0.1, 0.9, 0.1, 0.9, 0.2, 0.8, 0.3, 0.7, 0.4, 0.6, 0.5]
merge_method: dare_ties
base_model: F:/CursedMatrix-8B-v1
dtype: bfloat16
# CursedMatrix-8B-v3
models:
- model: F:/TitanPlanet-8B-v1
- model: F:/CursedMatrix-8B-v1
- model: F:/EtherealNymph-8B-v1
merge_method: model_stock
base_model: F:/CursedMatrix-8B-v2
dtype: bfloat16
### ::: Generation 3 merges :
# CursedMatrix-8B-v4
models:
- model: F:/CursedMatrix-8B-v3
parameters:
weight: 0.8
- model: F:/TitanNymph-8B-v1
parameters:
weight: 0.4
- model: DreadPoor/Trinas_Nectar-8B-model_stock
parameters:
weight: 0.3
- model: F:/DevilMind-8B-v1
parameters:
weight: 0.2
merge_method: task_arithmetic
base_model: F:/CursedMatrix-8B-v3
dtype: bfloat16
# CursedMatrix-8B-v4-rev2
models:
- model: F:/CursedMatrix-8B-v3
parameters:
weight: 0.8
- model: F:/TitanNymph-8B-v1
parameters:
weight: 0.4
- model: DreadPoor/Trinas_Nectar-8B-model_stock
parameters:
weight: 0.3
- model: F:/DevilMind-8B-v1
parameters:
weight: 0.2
merge_method: task_arithmetic
base_model: F:/CursedMatrix-8B-v1
dtype: bfloat16
# CursedMatrix-8B-v5
models:
- model: F:/CursedMatrix-8B-v4-rev2
merge_method: slerp
base_model: F:/CursedMatrix-8B-v4
dtype: bfloat16
parameters:
t: [0.5, 0.6, 0.4, 0.7, 0.3, 0.8, 0.2, 0.8, 0.2, 0.7, 0.3, 0.6, 0.4, 0.5]
### ::: Generation 4 merges :
# CursedMatrix-8B-v6
models:
- model: jeiku/Chaos_RP_l3_8B
- model: ZeroXClem/Llama-3-8B-ProLong-SAO-Roleplay-512k
merge_method: model_stock
base_model: F:/CursedMatrix-8B-v5
dtype: bfloat16
# CursedMatrix-8B-v6-rev2
models:
- model: win10/ArliAI-RPMax-v1.3-merge-8B
- model: v000000/L3.1-Storniitova-8B
- model: d0rj/Llama-3-8B-saiga-suzume-ties
merge_method: model_stock
base_model: F:/CursedMatrix-8B-v5
dtype: bfloat16
# CursedMatrix-8B-v6-rev3
models:
- model: Casual-Autopsy/L3-Super-Nova-RP-8B
- model: IlyaGusev/saiga_llama3_8b
merge_method: model_stock
base_model: F:/CursedMatrix-8B-v5
dtype: bfloat16
# CursedMatrix-8B-v7
models:
- model: F:/CursedMatrix-8B-v6
parameters:
weight: [0.5, 0.4, 0.6, 0.3, 0.7, 0.2, 0.8, 0.2, 0.8, 0.3, 0.7, 0.4, 0.6, 0.5]
density: [0.05, 0.25]
- model: F:/CursedMatrix-8B-v6-rev3
parameters:
weight: [0.5, 0.6, 0.4, 0.7, 0.3, 0.8, 0.2, 0.8, 0.2, 0.7, 0.3, 0.6, 0.4, 0.5]
density: [0.25, 0.05]
merge_method: ties
base_model: F:/CursedMatrix-8B-v6-rev2
dtype: bfloat16
# CursedMatrix-8B-v8
models:
- model: F:/CursedMatrix-8B-v6
- model: F:/CursedMatrix-8B-v6-rev2
- model: F:/CursedMatrix-8B-v6-rev3
merge_method: model_stock
base_model: F:/CursedMatrix-8B-v7
dtype: bfloat16
### ::: Generation 5 merges :
# Cursed-DarkRainbow-8B-v1
models:
- model: invisietch/L3.1-EtherealRainbow-v1.0-rc1-8B
parameters:
weight: [0.5, 0.6, 0.4, 0.7, 0.3, 0.8, 0.2, 0.8, 0.2, 0.7, 0.3, 0.6, 0.4, 0.5]
density: [0.5, 0.4, 0.6, 0.3, 0.7, 0.2, 0.8, 0.2, 0.8, 0.3, 0.7, 0.4, 0.6, 0.5]
- model: rityak/L3.1-DarkStock-8B
parameters:
weight: [0.5, 0.4, 0.6, 0.3, 0.7, 0.2, 0.8, 0.2, 0.8, 0.3, 0.7, 0.4, 0.6, 0.5]
density: [0.5, 0.6, 0.4, 0.7, 0.3, 0.8, 0.2, 0.8, 0.2, 0.7, 0.3, 0.6, 0.4, 0.5]
merge_method: della
parameters:
epsilon: 0.123456789
lambda: 0.987654321
base_model: F:/CursedMatrix-8B-v8
dtype: bfloat16
# Cursed-BlueBaezel-8B-v1
models:
- model: v000000/L3-8B-BlueSerpentine
parameters:
weight: [0.5, 0.6, 0.4, 0.7, 0.3, 0.8, 0.2, 0.8, 0.2, 0.7, 0.3, 0.6, 0.4, 0.5]
density: [0.5, 0.4, 0.6, 0.3, 0.7, 0.2, 0.8, 0.2, 0.8, 0.3, 0.7, 0.4, 0.6, 0.5]
- model: DreadPoor/L3.1-BaeZel-8B-Della
parameters:
weight: [0.5, 0.4, 0.6, 0.3, 0.7, 0.2, 0.8, 0.2, 0.8, 0.3, 0.7, 0.4, 0.6, 0.5]
density: [0.5, 0.6, 0.4, 0.7, 0.3, 0.8, 0.2, 0.8, 0.2, 0.7, 0.3, 0.6, 0.4, 0.5]
merge_method: della
parameters:
epsilon: 0.123456789
lambda: 0.987654321
base_model: F:/CursedMatrix-8B-v8
dtype: bfloat16
# Cursed-SuzumeMaid-8B-v1
models:
- model: d0rj/Llama-3-8B-saiga-suzume-ties
parameters:
weight: [0.5, 0.6, 0.4, 0.7, 0.3, 0.8, 0.2, 0.8, 0.2, 0.7, 0.3, 0.6, 0.4, 0.5]
density: [0.5, 0.4, 0.6, 0.3, 0.7, 0.2, 0.8, 0.2, 0.8, 0.3, 0.7, 0.4, 0.6, 0.5]
- model: bluuwhale/L3-SthenoMaidBlackroot-8B-V1
parameters:
weight: [0.5, 0.4, 0.6, 0.3, 0.7, 0.2, 0.8, 0.2, 0.8, 0.3, 0.7, 0.4, 0.6, 0.5]
density: [0.5, 0.6, 0.4, 0.7, 0.3, 0.8, 0.2, 0.8, 0.2, 0.7, 0.3, 0.6, 0.4, 0.5]
merge_method: della
parameters:
epsilon: 0.123456789
lambda: 0.987654321
base_model: F:/CursedMatrix-8B-v8
dtype: bfloat16
# Cursed-UnalignedSaiga-8B-v1
models:
- model: SicariusSicariiStuff/LLAMA-3_8B_Unaligned_BETA
parameters:
weight: [0.5, 0.6, 0.4, 0.7, 0.3, 0.8, 0.2, 0.8, 0.2, 0.7, 0.3, 0.6, 0.4, 0.5]
density: [0.5, 0.4, 0.6, 0.3, 0.7, 0.2, 0.8, 0.2, 0.8, 0.3, 0.7, 0.4, 0.6, 0.5]
- model: IlyaGusev/saiga_llama3_8b
parameters:
weight: [0.5, 0.4, 0.6, 0.3, 0.7, 0.2, 0.8, 0.2, 0.8, 0.3, 0.7, 0.4, 0.6, 0.5]
density: [0.5, 0.6, 0.4, 0.7, 0.3, 0.8, 0.2, 0.8, 0.2, 0.7, 0.3, 0.6, 0.4, 0.5]
merge_method: della
parameters:
epsilon: 0.123456789
lambda: 0.987654321
base_model: F:/CursedMatrix-8B-v8
dtype: bfloat16
# CursedMatrix-8B-v9
# Final model...
models:
- model: F:/Cursed-UnalignedSaiga-8B-v1
- model: F:/Cursed-DarkRainbow-8B-v1
- model: F:/Cursed-BlueBaezel-8B-v1
- model: F:/Cursed-SuzumeMaid-8B-v1
merge_method: model_stock
base_model: F:/CursedMatrix-8B-v8
dtype: bfloat16
My thanks to the authors of the original models, your work is incredible. Have a good time 🖤