Adapt max tokens
Browse files- e2bqwen.py +2 -0
e2bqwen.py
CHANGED
|
@@ -443,11 +443,13 @@ class QwenVLAPIModel(Model):
|
|
| 443 |
self.base_model = HfApiModel(
|
| 444 |
model_id="https://n5wr7lfx6wp94tvl.us-east-1.aws.endpoints.huggingface.cloud",
|
| 445 |
token=hf_token,
|
|
|
|
| 446 |
)
|
| 447 |
self.fallback_model = HfApiModel(
|
| 448 |
model_id,
|
| 449 |
provider="hyperbolic",
|
| 450 |
token=hf_token,
|
|
|
|
| 451 |
)
|
| 452 |
|
| 453 |
def __call__(
|
|
|
|
| 443 |
self.base_model = HfApiModel(
|
| 444 |
model_id="https://n5wr7lfx6wp94tvl.us-east-1.aws.endpoints.huggingface.cloud",
|
| 445 |
token=hf_token,
|
| 446 |
+
max_tokens=4096,
|
| 447 |
)
|
| 448 |
self.fallback_model = HfApiModel(
|
| 449 |
model_id,
|
| 450 |
provider="hyperbolic",
|
| 451 |
token=hf_token,
|
| 452 |
+
max_tokens=4096,
|
| 453 |
)
|
| 454 |
|
| 455 |
def __call__(
|