Edit model card

EXL2 Quantization of Putri's Megamix-A1.

GGUF quants from Sao10K here: MegaMix-L2-13B-GGUF

Model details

Quantized at 5.33bpw

Prompt Format

I'm using Alpaca format:

### Instruction:


### Response:
Downloads last month
14
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Collection including R136a1/Megamix-A1-13B-exl2