Edit model card

GGUF quantization for the InternLM-20b-chat model.

Requires a recent version of llama.cpp: https://github.com/ggerganov/llama.cpp/pull/4283

Link to the original model: https://huggingface.co/internlm/internlm-chat-20b

Github repository: https://github.com/InternLM/InternLM

Prompt format:

<|User|>: Your input.<eoh>
<|Bot|>:

License:

Model weights are fully open for academic research and also allow free commercial usage. To apply for a commercial license, please fill in the application form (English)/申请表(中文). For other questions or collaborations, please contact [email protected].

Downloads last month
147
GGUF
Model size
20.1B params
Architecture
llama

2-bit

3-bit

4-bit

5-bit

6-bit

Inference API
Unable to determine this model's library. Check the docs .