Spaces:
Paused
Paused
llama adapter doesn't support use_cache yet
Browse files
llama_lora/ui/inference_ui.py
CHANGED
@@ -111,6 +111,7 @@ def do_inference(
|
|
111 |
top_k=top_k,
|
112 |
repetition_penalty=repetition_penalty,
|
113 |
num_beams=num_beams,
|
|
|
114 |
)
|
115 |
|
116 |
generate_params = {
|
|
|
111 |
top_k=top_k,
|
112 |
repetition_penalty=repetition_penalty,
|
113 |
num_beams=num_beams,
|
114 |
+
use_cache=False,
|
115 |
)
|
116 |
|
117 |
generate_params = {
|