Spaces:
Running
on
Zero
Running
on
Zero
Remove unavailable Llama fallback
Browse files- __pycache__/app.cpython-313.pyc +0 -0
- app.py +0 -1
__pycache__/app.cpython-313.pyc
CHANGED
|
Binary files a/__pycache__/app.cpython-313.pyc and b/__pycache__/app.cpython-313.pyc differ
|
|
|
app.py
CHANGED
|
@@ -32,7 +32,6 @@ DEFAULT_TOP_P = float(os.environ.get("DEFAULT_TOP_P", "0.9"))
|
|
| 32 |
USE_4BIT = os.environ.get("LOAD_IN_4BIT", "1") not in {"0", "false", "False"}
|
| 33 |
|
| 34 |
MODEL_FALLBACKS = [
|
| 35 |
-
"Alovestocode/router-llama31-merged",
|
| 36 |
"Alovestocode/router-qwen3-32b-merged",
|
| 37 |
"Alovestocode/router-gemma3-merged",
|
| 38 |
]
|
|
|
|
| 32 |
USE_4BIT = os.environ.get("LOAD_IN_4BIT", "1") not in {"0", "false", "False"}
|
| 33 |
|
| 34 |
MODEL_FALLBACKS = [
|
|
|
|
| 35 |
"Alovestocode/router-qwen3-32b-merged",
|
| 36 |
"Alovestocode/router-gemma3-merged",
|
| 37 |
]
|