Edit model card

CollectiveCognition-v1.1-Mistral-7B and airoboros-mistral2.2-7b glued together.

Description

This repo contains quantized files of Mistral-11B-CC-Air.

Model used

Prompt template: Alpaca or default

Below is an instruction that describes a task. Write a response that appropriately completes the request.

### Instruction:
{prompt}

### Response:
USER: <prompt>
ASSISTANT:

The secret sauce

slices:
  - sources:
    - model: teknium/CollectiveCognition-v1.1-Mistral-7B
      layer_range: [0, 24]
  - sources:
    - model: teknium/airoboros-mistral2.2-7b
      layer_range: [8, 32]
merge_method: passthrough
dtype: float16

Special thanks to Sushi.

If you want to support me, you can here.

Downloads last month
7
GGUF
Model size
10.7B params
Architecture
llama

4-bit

5-bit

6-bit

8-bit

Inference API
Unable to determine this model's library. Check the docs .