Edit model card

General Use Sampling:
Mistral-Nemo-12B is very sensitive to the temperature sampler, try values near 0.3 at first or else you will get some weird results. This is mentioned by MistralAI at their Transformers section.

Best Samplers:
I found best success using the following for Hollow-Tail-V1-12B:
Temperature: 1.2
Top K: -1
Min P: 0.05
Rep Penalty: 1.08

Results

Disclaimer: This is a model merge!

Seems to be a bit smarter than I expected from my experience. May need a bit of guidance through system prompts at the beginning but it was quite fun to use. Though one thing I've realized with Mistral-Nemo is that the model does not really seem to end correctly, so instead I use the following format:

Obviously this is just my personal experience but I find this to be a good setup, I strongly recommend you experiment with different system formats and see which is better for you. Note that none of these models are fine-tuned for this specific format, I believe most of them are fine-tuned on Mistral's original [INST] and [/INST] format or ChatML.

<[start_system]>
You are a professional writer.
<[STOP]>
<[start_prompt]>
User prompt here.
<[STOP]>
<[start_model]>
Model response here.
<[STOP]>

Original Models:

Original Model Merge:

Original Model Licenses:

  • Sao10K/MN-12B-Lyra-v2a1 is licensed under Creative Commons Attribution Non Commercial 4.0
  • migtissera/Tess-3-Mistral-Nemo-12B is licensed under apache-2.0
  • TheDrummer/Rocinante-12B-v1.1 license is not specified

Hollow-Tail-V1-12B

This is a merge of pre-trained language models created using mergekit.

Merge Details

Merge Method

This model was merged using the linear merge method using models/Rocinante-12B-v1.1 as a base.

Models Merged

The following models were included in the merge:

  • migtissera/Tess-3-Mistral-Nemo-12B
  • Sao10K/MN-12B-Lyra-v2a1
  • TheDrummer/Rocinante-12B-v1.1

Configuration

The following YAML configuration was used to produce this model:

models:
    - model: Sao10K/MN-12B-Lyra-v2a1
      parameters:
        weight: 0.8
    - model: migtissera/Tess-3-Mistral-Nemo-12B
      parameters:
        weight: 0.2
    - model: TheDrummer/Rocinante-12B-v1.1
      parameters:
        weight: 0.8
merge_method: linear
base_model: TheDrummer/Rocinante-12B-v1.1
parameters:
    normalize: true
    int8_mask: true
dtype: bfloat16
Downloads last month
165
GGUF
Model size
12.2B params
Architecture
llama

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

Inference API
Unable to determine this model’s pipeline type. Check the docs .

Model tree for starble-dev/Hollow-Tail-V1-12B-GGUF

Quantized
(3)
this model

Space using starble-dev/Hollow-Tail-V1-12B-GGUF 1