Spaces:
Runtime error
Runtime error
Update model.py
Browse files
model.py
CHANGED
@@ -5,10 +5,11 @@ from typing import Iterator
|
|
5 |
|
6 |
model_id = 'xuqinyang/baichuan-13b-chat-ggml-int4'
|
7 |
|
8 |
-
from huggingface_hub import snapshot_download
|
9 |
#旧
|
10 |
#snapshot_download(model_id, local_dir="./",revision="7f71a8abefa7b2eede3f74ce0564abe5fbe6874a")
|
11 |
snapshot_download(model_id, local_dir="./",revision="b2414a0ceee68fe09c99ace44446cfc9a1c52b08")
|
|
|
12 |
from llama_cpp import Llama
|
13 |
llm = Llama(model_path="./ggml-model-q4_0.bin", n_ctx=4096,seed=-1)
|
14 |
|
|
|
5 |
|
6 |
model_id = 'xuqinyang/baichuan-13b-chat-ggml-int4'
|
7 |
|
8 |
+
from huggingface_hub import snapshot_download,hf_hub_download
|
9 |
#旧
|
10 |
#snapshot_download(model_id, local_dir="./",revision="7f71a8abefa7b2eede3f74ce0564abe5fbe6874a")
|
11 |
snapshot_download(model_id, local_dir="./",revision="b2414a0ceee68fe09c99ace44446cfc9a1c52b08")
|
12 |
+
hf_hub_download(repo_id="baichuan-inc/Baichuan-13B-Chat",local_dir="./", filename="tokenizer.model")
|
13 |
from llama_cpp import Llama
|
14 |
llm = Llama(model_path="./ggml-model-q4_0.bin", n_ctx=4096,seed=-1)
|
15 |
|