Spaces:
Sleeping
Sleeping
Commit
·
1e66988
1
Parent(s):
437e3cd
Added max_tokens
Browse files- api/llm.py +2 -1
api/llm.py
CHANGED
|
@@ -59,7 +59,7 @@ class LLMManager:
|
|
| 59 |
|
| 60 |
def get_text(self, messages):
|
| 61 |
try:
|
| 62 |
-
response = self.client.chat.completions.create(model=self.config.llm.name, messages=messages, temperature=1)
|
| 63 |
if not response.choices:
|
| 64 |
raise APIError("LLM Get Text Error", details="No choices in response")
|
| 65 |
return response.choices[0].message.content.strip()
|
|
@@ -73,6 +73,7 @@ class LLMManager:
|
|
| 73 |
messages=messages,
|
| 74 |
temperature=1,
|
| 75 |
stream=True,
|
|
|
|
| 76 |
)
|
| 77 |
except Exception as e:
|
| 78 |
raise APIError(f"LLM End Interview Error: Unexpected error: {e}")
|
|
|
|
| 59 |
|
| 60 |
def get_text(self, messages):
|
| 61 |
try:
|
| 62 |
+
response = self.client.chat.completions.create(model=self.config.llm.name, messages=messages, temperature=1, max_tokens=2000)
|
| 63 |
if not response.choices:
|
| 64 |
raise APIError("LLM Get Text Error", details="No choices in response")
|
| 65 |
return response.choices[0].message.content.strip()
|
|
|
|
| 73 |
messages=messages,
|
| 74 |
temperature=1,
|
| 75 |
stream=True,
|
| 76 |
+
max_tokens=2000,
|
| 77 |
)
|
| 78 |
except Exception as e:
|
| 79 |
raise APIError(f"LLM End Interview Error: Unexpected error: {e}")
|