Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -19,7 +19,7 @@ FESTIVE = True # Set to False for production-only mode
|
|
| 19 |
|
| 20 |
print("π Loading SAM-Z-1 Model...")
|
| 21 |
|
| 22 |
-
MODEL_REPO = "Smilyai-labs/Sam-
|
| 23 |
CACHE_DIR = "./model_cache"
|
| 24 |
|
| 25 |
# ============================================================================
|
|
@@ -242,7 +242,7 @@ from transformers import AutoTokenizer
|
|
| 242 |
hf_tokenizer = AutoTokenizer.from_pretrained("gpt2")
|
| 243 |
|
| 244 |
# Add custom tokens to match model's vocab size
|
| 245 |
-
custom_tokens = ["<|im_start|>", "<|im_end|>", "<think>", "<think/>"]
|
| 246 |
hf_tokenizer.add_special_tokens({"additional_special_tokens": custom_tokens})
|
| 247 |
|
| 248 |
# Save and reload as tokenizers format
|
|
|
|
| 19 |
|
| 20 |
print("π Loading SAM-Z-1 Model...")
|
| 21 |
|
| 22 |
+
MODEL_REPO = "Smilyai-labs/Sam-large-2"
|
| 23 |
CACHE_DIR = "./model_cache"
|
| 24 |
|
| 25 |
# ============================================================================
|
|
|
|
| 242 |
hf_tokenizer = AutoTokenizer.from_pretrained("gpt2")
|
| 243 |
|
| 244 |
# Add custom tokens to match model's vocab size
|
| 245 |
+
custom_tokens = ["<|im_start|>", "<|im_end|>", "<think>", "<think/>", "<CONTINUE>"]
|
| 246 |
hf_tokenizer.add_special_tokens({"additional_special_tokens": custom_tokens})
|
| 247 |
|
| 248 |
# Save and reload as tokenizers format
|