Spaces:
Paused
Paused
llama adapter doesn't support use_cache yet
Browse files
llama_lora/ui/inference_ui.py
CHANGED
|
@@ -111,6 +111,7 @@ def do_inference(
|
|
| 111 |
top_k=top_k,
|
| 112 |
repetition_penalty=repetition_penalty,
|
| 113 |
num_beams=num_beams,
|
|
|
|
| 114 |
)
|
| 115 |
|
| 116 |
generate_params = {
|
|
|
|
| 111 |
top_k=top_k,
|
| 112 |
repetition_penalty=repetition_penalty,
|
| 113 |
num_beams=num_beams,
|
| 114 |
+
use_cache=False,
|
| 115 |
)
|
| 116 |
|
| 117 |
generate_params = {
|