Glm-Edge-Chat-4B-GGUF

中文阅读, 点击这里

Inference with Ollama

Installation

The code for adapting this model is actively being integrated into the official llama.cpp. You can test it using the following adapted version:

git clone https://github.com/piDack/llama.cpp -b support_glm_edge_model
cmake -B build -DGGML_CUDA=ON # Or enable other acceleration hardware
cmake --build build -- -j 

Inference

After installation, you can start the GLM-Edge Chat model using the following command:

llama-cli -m <path>/model.gguf -p "<|user|>\nhi<|assistant|>\n" -ngl 999

In the command-line interface, you can interact with the model by entering your requests, and the model will provide the corresponding responses.

License

The usage of this model’s weights is subject to the terms outlined in the LICENSE.

Downloads last month
1,053
GGUF
Model size
4.33B params
Architecture
chatglm

4-bit

5-bit

6-bit

8-bit

16-bit

Inference Examples
Inference API (serverless) has been turned off for this model.

Collection including THUDM/glm-edge-4b-chat-gguf