Spaces:
Running
Running
Update modules/local_llm.py
Browse files- modules/local_llm.py +9 -6
modules/local_llm.py
CHANGED
|
@@ -21,7 +21,7 @@ def _get_llm():
|
|
| 21 |
|
| 22 |
logger.info("Carregando OpenHermes-2.5-Mistral-7B Q4_K_M (4.8 GB RAM)...")
|
| 23 |
if not os.path.exists(MODEL_PATH):
|
| 24 |
-
logger.error("GGUF não encontrado!
|
| 25 |
return None
|
| 26 |
|
| 27 |
try:
|
|
@@ -29,20 +29,23 @@ def _get_llm():
|
|
| 29 |
model_path=MODEL_PATH,
|
| 30 |
n_ctx=4096,
|
| 31 |
n_threads=4,
|
| 32 |
-
n_gpu_layers=0, # CPU only
|
| 33 |
n_batch=512,
|
|
|
|
| 34 |
verbose=False
|
| 35 |
)
|
| 36 |
|
| 37 |
-
|
| 38 |
-
|
| 39 |
-
|
|
|
|
|
|
|
|
|
|
| 40 |
|
| 41 |
_llm_global = llm
|
| 42 |
logger.info("OpenHermes 2.5 Q4_K_M + LoRA → ONLINE EM 4.8 GB!")
|
| 43 |
return llm
|
| 44 |
except Exception as e:
|
| 45 |
-
logger.error(f"ERRO GGUF: {e}")
|
| 46 |
return None
|
| 47 |
|
| 48 |
class LocalLLM:
|
|
|
|
| 21 |
|
| 22 |
logger.info("Carregando OpenHermes-2.5-Mistral-7B Q4_K_M (4.8 GB RAM)...")
|
| 23 |
if not os.path.exists(MODEL_PATH):
|
| 24 |
+
logger.error("GGUF não encontrado! Verifica o download.")
|
| 25 |
return None
|
| 26 |
|
| 27 |
try:
|
|
|
|
| 29 |
model_path=MODEL_PATH,
|
| 30 |
n_ctx=4096,
|
| 31 |
n_threads=4,
|
|
|
|
| 32 |
n_batch=512,
|
| 33 |
+
n_gpu_layers=0,
|
| 34 |
verbose=False
|
| 35 |
)
|
| 36 |
|
| 37 |
+
# CARREGA LORA ANGOLANO
|
| 38 |
+
lora_path = f"{FINETUNED_PATH}/adapter_model.bin"
|
| 39 |
+
if os.path.exists(lora_path):
|
| 40 |
+
logger.info("LORA ANGOLANO DETECTADO → CARREGANDO...")
|
| 41 |
+
llm.load_lora(lora_path)
|
| 42 |
+
logger.info("AKIRA COM SOTAQUE DE LUANDA 100% ATIVA!")
|
| 43 |
|
| 44 |
_llm_global = llm
|
| 45 |
logger.info("OpenHermes 2.5 Q4_K_M + LoRA → ONLINE EM 4.8 GB!")
|
| 46 |
return llm
|
| 47 |
except Exception as e:
|
| 48 |
+
logger.error(f"ERRO AO CARREGAR GGUF: {e}")
|
| 49 |
return None
|
| 50 |
|
| 51 |
class LocalLLM:
|