Update app.py
Browse files
app.py
CHANGED
@@ -64,6 +64,8 @@ status = ""
|
|
64 |
|
65 |
agreed = ""
|
66 |
|
|
|
|
|
67 |
from llama_cpp import Llama
|
68 |
llm = Llama(model_path=model_file, model_type="mistral",n_gpu_layers=-1,n_ctx = 2048)
|
69 |
|
@@ -170,8 +172,7 @@ def generate(prompt, history):
|
|
170 |
"""
|
171 |
|
172 |
response = ""
|
173 |
-
|
174 |
-
|
175 |
while(len(response) < 1):
|
176 |
output = llm(context, max_tokens=400, stop=["Nurse:"], echo=False)
|
177 |
response = output["choices"][0]["text"]
|
|
|
64 |
|
65 |
agreed = ""
|
66 |
|
67 |
+
history=[]
|
68 |
+
|
69 |
from llama_cpp import Llama
|
70 |
llm = Llama(model_path=model_file, model_type="mistral",n_gpu_layers=-1,n_ctx = 2048)
|
71 |
|
|
|
172 |
"""
|
173 |
|
174 |
response = ""
|
175 |
+
|
|
|
176 |
while(len(response) < 1):
|
177 |
output = llm(context, max_tokens=400, stop=["Nurse:"], echo=False)
|
178 |
response = output["choices"][0]["text"]
|