Edit model card
Configuration Parsing Warning: In config.json: "quantization_config.bits" must be an integer

Magnum-V4-2411-Merge-exl2-4.5bpw

EXL2 Quant of gghfez/Magnum-V4-2411-Merge

A merge of anthracite-org/magnum-v4-123b the new mistralai/Mistral-Large-Instruct-2411 via LoRA extraction.

Model Details

Prompting

A typical input would look like this:


<s>[INST] SYSTEM MESSAGE\nUSER MESSAGE[/INST] ASSISTANT MESSAGE</s>[INST] USER MESSAGE[/INST]

Results

I haven't tested it extensively, but I don't see much difference between this and anthracite-org/magnum-v4-123b.

Downloads last month
5
Inference API
Unable to determine this model's library. Check the docs .

Model tree for gghfez/Magnum-V4-2411-Merge-exl2-4.5bpw

Quantized
(6)
this model