Edit model card

Miquliz-120b-v2.0-FP8-dynamic

image/jpeg

This quant was made for infermatic.ai

Dynamic FP8 quant of Miquliz 120B v2.0 made with AutoFP8.

Model Details

  • Max Context: 32768 tokens
  • Layers: 140

Prompt template: Mistral

<s>[INST] {prompt} [/INST]
Downloads last month
5
Safetensors
Model size
120B params
Tensor type
FP16
·
F8_E4M3
·
Inference Examples
Inference API (serverless) is not available, repository is disabled.