Update app.py
Browse files
app.py
CHANGED
|
@@ -38,6 +38,7 @@ model = AutoModelForCausalLM.from_pretrained(
|
|
| 38 |
device_map="auto",
|
| 39 |
ignore_mismatched_sizes=True)
|
| 40 |
|
|
|
|
| 41 |
|
| 42 |
@spaces.GPU()
|
| 43 |
def stream_chat(
|
|
@@ -75,7 +76,8 @@ def stream_chat(
|
|
| 75 |
top_k=top_k,
|
| 76 |
temperature=temperature,
|
| 77 |
streamer=streamer,
|
| 78 |
-
pad_token_id=
|
|
|
|
| 79 |
)
|
| 80 |
|
| 81 |
with torch.no_grad():
|
|
@@ -86,6 +88,8 @@ def stream_chat(
|
|
| 86 |
for new_text in streamer:
|
| 87 |
buffer += new_text
|
| 88 |
yield buffer
|
|
|
|
|
|
|
| 89 |
|
| 90 |
|
| 91 |
chatbot = gr.Chatbot(height=600, placeholder=PLACEHOLDER)
|
|
|
|
| 38 |
device_map="auto",
|
| 39 |
ignore_mismatched_sizes=True)
|
| 40 |
|
| 41 |
+
eos_token_id = tokenizer.eos_token_id
|
| 42 |
|
| 43 |
@spaces.GPU()
|
| 44 |
def stream_chat(
|
|
|
|
| 76 |
top_k=top_k,
|
| 77 |
temperature=temperature,
|
| 78 |
streamer=streamer,
|
| 79 |
+
pad_token_id=eos_token_id,
|
| 80 |
+
eos_token_id=eos_token_id,
|
| 81 |
)
|
| 82 |
|
| 83 |
with torch.no_grad():
|
|
|
|
| 88 |
for new_text in streamer:
|
| 89 |
buffer += new_text
|
| 90 |
yield buffer
|
| 91 |
+
if eos_token_id in tokenizer.encode(new_text):
|
| 92 |
+
break
|
| 93 |
|
| 94 |
|
| 95 |
chatbot = gr.Chatbot(height=600, placeholder=PLACEHOLDER)
|