|
--- |
|
base_model: |
|
- 152334H/miqu-1-70b-sf |
|
- NeverSleep/MiquMaid-v1-70B |
|
- Sao10K/WinterGoddess-1.4x-70B-L2 |
|
library_name: transformers |
|
tags: |
|
- mergekit |
|
- merge |
|
--- |
|
# aranea-ancilla-70b-v1.0 |
|
**aka MiquMaid-v1-70B + interleaved WinterGoddess-1.4x-70B-L2** |
|
|
|
![image/png](https://huggingface.co/divinetaco/aranea-ancilla-116b-v1.0/resolve/main/aranea-ancilla.png) |
|
|
|
A [mergekit](https://github.com/arcee-ai/mergekit) frankenmerge based on [MiquMaid-v1-70B](https://huggingface.co/NeverSleep/MiquMaid-v1-70B) with interleaved layers of [Sao10K/WinterGoddess-1.4x-70B-L2](https://huggingface.co/Sao10K/WinterGoddess-1.4x-70B-L2). |
|
This was the top performing model from a series of merge experiments to create a highly coherant creative writing model. |
|
|
|
Tests consisted of a series of private benchmarks and manual comparisons. A number of different base models, interleave models and layer offsets were compared. |
|
|
|
- Usable context ~32768 |
|
- Recommended context ~16384 |
|
|
|
Non frankenstein miqu-1 finetunes generally outperform their frankenstein counterparts at very long contexts due to coherency loss. |
|
As a rough suggestion I might suggest swapping out to either [NeverSleep/MiquMaid-v1-70B](https://huggingface.co/NeverSleep/MiquMaid-v1-70B) or [152334H/miqu-1-70b-sf](https://huggingface.co/152334H/miqu-1-70b-sf) after 16k context. |
|
|
|
Layers: 136 |
|
|
|
### License |
|
|
|
No license. Component models based on the [Mistral AI Miqu-1](https://huggingface.co/miqudev/miqu-1-70b/tree/main) llama2 finetune that was released without license. |
|
|
|
### Interesting observations from benchmarking |
|
|
|
- 10 layer interleave stride with a 20 layer interleave width consistently outperformed alternatives combinations. |
|
- Offsetting the interleaved model's first set of layers generally improved coherency. [14-30] reliably beat the [10-30] mergekit slice configuration for combinations of models. |
|
- Quality of resulting merges can vary wildly. Whilst a merge of two strong models tends to produce a strong frankenstein model, this rule does not always hold true. |
|
|
|
### Quantizations |
|
|
|
Exllamav2 quants will be available when bandwidth permits. |