Edit model card

ExllamaV2 version of the model created by Undi!

Original Model https://huggingface.co/Undi95/Dawn-v2-70B

Requires ExllamaV2, which is being developed by turboderp https://github.com/turboderp/exllamav2 under an MIT license.

Main branch is 4.6bpw 8h (req ??gb)

2.5b8h branch is 2.5bpw 8h (req 24gb and the 8b cache setting) - Add BOS token must be unchecked at this weight or output is nonsense. New quant method applied 12/17/2023

5.0b8h branch is 5.0bpw 8h (req ??gb)

6b8h branch is 6.0bpw 8h requires between 60-72gb


[fp16 - gguf - exl2 : 2.40bpw - 2.55bpw]

Description

This repo contains fp16 files of Dawn-70B, a merge I have done with the new layer shuffle method from mergekit.

UtopiaXL was a huge success for me, I really liked it, so I took the same path to do this 70B: A good base, some psychologic data, some medical data, a little bit of this, of that, and LimaRP at the end as always.

NOTE: This repo contain the file measurement.json needed to do your own exl2 quant (I use wikitext).

Models and loras used

The sauce

!mergekit-layershuffle ./Dawn-v2-70B \
  --model Sao10K/Euryale-1.3-L2-70B --weight 0.3 \
  --model Xwin-LM/Xwin-LM-70B-V0.1 --weight 0.2 \
  --model ehartford/Samantha-1.11-70b --weight 0.1 \
  --model NousResearch/Nous-Hermes-Llama2-70b --weight 0.05 \
  --model augtoma/qCammel-70-x --weight 0.05 \
  --model jondurbin/airoboros-l2-c70b-3.1.2 --weight 0.2 \
  --model fangloveskari/ORCA_LLaMA_70B_QLoRA --weight 0.1 \
  --write-yaml Dawn-v2-70B.yaml

=========================

merge_method: passthrough
slices:
- sources:
  - layer_range:
    - 0
    - 1
    model: fangloveskari/ORCA_LLaMA_70B_QLoRA
- sources:
  - layer_range:
    - 1
    - 2
    model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
  - layer_range:
    - 2
    - 3
    model: Sao10K/Euryale-1.3-L2-70B
- sources:
  - layer_range:
    - 3
    - 4
    model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
  - layer_range:
    - 4
    - 5
    model: fangloveskari/ORCA_LLaMA_70B_QLoRA
- sources:
  - layer_range:
    - 5
    - 6
    model: ehartford/Samantha-1.11-70b
- sources:
  - layer_range:
    - 6
    - 8
    model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
  - layer_range:
    - 8
    - 9
    model: ehartford/Samantha-1.11-70b
- sources:
  - layer_range:
    - 9
    - 10
    model: Sao10K/Euryale-1.3-L2-70B
- sources:
  - layer_range:
    - 10
    - 11
    model: ehartford/Samantha-1.11-70b
- sources:
  - layer_range:
    - 11
    - 12
    model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
  - layer_range:
    - 12
    - 13
    model: fangloveskari/ORCA_LLaMA_70B_QLoRA
- sources:
  - layer_range:
    - 13
    - 14
    model: Sao10K/Euryale-1.3-L2-70B
- sources:
  - layer_range:
    - 14
    - 15
    model: fangloveskari/ORCA_LLaMA_70B_QLoRA
- sources:
  - layer_range:
    - 15
    - 16
    model: Sao10K/Euryale-1.3-L2-70B
- sources:
  - layer_range:
    - 16
    - 17
    model: fangloveskari/ORCA_LLaMA_70B_QLoRA
- sources:
  - layer_range:
    - 17
    - 18
    model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
  - layer_range:
    - 18
    - 19
    model: NousResearch/Nous-Hermes-Llama2-70b
- sources:
  - layer_range:
    - 19
    - 20
    model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
  - layer_range:
    - 20
    - 21
    model: Sao10K/Euryale-1.3-L2-70B
- sources:
  - layer_range:
    - 21
    - 22
    model: ehartford/Samantha-1.11-70b
- sources:
  - layer_range:
    - 22
    - 23
    model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
  - layer_range:
    - 23
    - 24
    model: augtoma/qCammel-70-x
- sources:
  - layer_range:
    - 24
    - 25
    model: Sao10K/Euryale-1.3-L2-70B
- sources:
  - layer_range:
    - 25
    - 27
    model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
  - layer_range:
    - 27
    - 28
    model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
  - layer_range:
    - 28
    - 29
    model: ehartford/Samantha-1.11-70b
- sources:
  - layer_range:
    - 29
    - 30
    model: Sao10K/Euryale-1.3-L2-70B
- sources:
  - layer_range:
    - 30
    - 32
    model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
  - layer_range:
    - 32
    - 33
    model: ehartford/Samantha-1.11-70b
- sources:
  - layer_range:
    - 33
    - 34
    model: augtoma/qCammel-70-x
- sources:
  - layer_range:
    - 34
    - 35
    model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
  - layer_range:
    - 35
    - 37
    model: Sao10K/Euryale-1.3-L2-70B
- sources:
  - layer_range:
    - 37
    - 38
    model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
  - layer_range:
    - 38
    - 39
    model: ehartford/Samantha-1.11-70b
- sources:
  - layer_range:
    - 39
    - 40
    model: augtoma/qCammel-70-x
- sources:
  - layer_range:
    - 40
    - 41
    model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
  - layer_range:
    - 41
    - 42
    model: ehartford/Samantha-1.11-70b
- sources:
  - layer_range:
    - 42
    - 43
    model: Sao10K/Euryale-1.3-L2-70B
- sources:
  - layer_range:
    - 43
    - 44
    model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
  - layer_range:
    - 44
    - 45
    model: NousResearch/Nous-Hermes-Llama2-70b
- sources:
  - layer_range:
    - 45
    - 46
    model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
  - layer_range:
    - 46
    - 48
    model: ehartford/Samantha-1.11-70b
- sources:
  - layer_range:
    - 48
    - 49
    model: Sao10K/Euryale-1.3-L2-70B
- sources:
  - layer_range:
    - 49
    - 50
    model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
  - layer_range:
    - 50
    - 51
    model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
  - layer_range:
    - 51
    - 54
    model: fangloveskari/ORCA_LLaMA_70B_QLoRA
- sources:
  - layer_range:
    - 54
    - 55
    model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
  - layer_range:
    - 55
    - 56
    model: fangloveskari/ORCA_LLaMA_70B_QLoRA
- sources:
  - layer_range:
    - 56
    - 58
    model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
  - layer_range:
    - 58
    - 59
    model: Sao10K/Euryale-1.3-L2-70B
- sources:
  - layer_range:
    - 59
    - 60
    model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
  - layer_range:
    - 60
    - 62
    model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
  - layer_range:
    - 62
    - 63
    model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
  - layer_range:
    - 63
    - 64
    model: fangloveskari/ORCA_LLaMA_70B_QLoRA
- sources:
  - layer_range:
    - 64
    - 65
    model: NousResearch/Nous-Hermes-Llama2-70b
- sources:
  - layer_range:
    - 65
    - 66
    model: Sao10K/Euryale-1.3-L2-70B
- sources:
  - layer_range:
    - 66
    - 67
    model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
  - layer_range:
    - 67
    - 68
    model: augtoma/qCammel-70-x
- sources:
  - layer_range:
    - 68
    - 70
    model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
  - layer_range:
    - 70
    - 71
    model: augtoma/qCammel-70-x
- sources:
  - layer_range:
    - 71
    - 72
    model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
  - layer_range:
    - 72
    - 73
    model: Sao10K/Euryale-1.3-L2-70B
- sources:
  - layer_range:
    - 73
    - 75
    model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
  - layer_range:
    - 75
    - 76
    model: Sao10K/Euryale-1.3-L2-70B
- sources:
  - layer_range:
    - 76
    - 77
    model: augtoma/qCammel-70-x
- sources:
  - layer_range:
    - 77
    - 78
    model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
  - layer_range:
    - 78
    - 79
    model: NousResearch/Nous-Hermes-Llama2-70b
- sources:
  - layer_range:
    - 79
    - 80
    model: Xwin-LM/Xwin-LM-70B-V0.1


=========================

=> Applying Doctor-Shotgun/limarpv3-llama2-70b-qlora x 0.35

Prompt template: Alpaca

Below is an instruction that describes a task. Write a response that appropriately completes the request.

### Instruction:
{prompt}

### Response:

A big thanks to Charles for adding the layer shuffle method to his tool mergekit and Henky/KoboldAI for the machine he let me use.

If you want to support me, you can here.

Downloads last month
12
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.