Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
|
@@ -49,7 +49,9 @@ def run_generation(user_text, use_assistant, temperature, max_new_tokens):
|
|
| 49 |
model_output = ""
|
| 50 |
for new_text in streamer:
|
| 51 |
model_output += new_text
|
| 52 |
-
|
|
|
|
|
|
|
| 53 |
|
| 54 |
|
| 55 |
def reset_textbox():
|
|
@@ -82,11 +84,11 @@ with gr.Blocks() as demo:
|
|
| 82 |
temperature = gr.Slider(
|
| 83 |
minimum=0.0, maximum=2.0, value=0.6, step=0.05, interactive=True, label="Temperature (0.0 = Greedy)",
|
| 84 |
)
|
| 85 |
-
gr.Markdown("###
|
| 86 |
-
|
| 87 |
|
| 88 |
generate_inputs = [user_text, use_assistant, temperature, max_new_tokens]
|
| 89 |
-
generate_outputs = [model_output,
|
| 90 |
user_text.submit(run_generation, generate_inputs, generate_outputs)
|
| 91 |
button_submit.click(run_generation, generate_inputs, generate_outputs)
|
| 92 |
|
|
|
|
| 49 |
model_output = ""
|
| 50 |
for new_text in streamer:
|
| 51 |
model_output += new_text
|
| 52 |
+
time_so_far = round(time.time() - start, 3)
|
| 53 |
+
tokens_so_far = tokenizer(model_output, return_tensors="pt").input_ids.shape[1]
|
| 54 |
+
yield [model_output, tokens_so_far/time_so_far]
|
| 55 |
|
| 56 |
|
| 57 |
def reset_textbox():
|
|
|
|
| 84 |
temperature = gr.Slider(
|
| 85 |
minimum=0.0, maximum=2.0, value=0.6, step=0.05, interactive=True, label="Temperature (0.0 = Greedy)",
|
| 86 |
)
|
| 87 |
+
gr.Markdown("### Tokens per secon")
|
| 88 |
+
tokens_per_second = gr.Textbox(lines=1, interactive=False, show_label=False)
|
| 89 |
|
| 90 |
generate_inputs = [user_text, use_assistant, temperature, max_new_tokens]
|
| 91 |
+
generate_outputs = [model_output, tokens_per_second]
|
| 92 |
user_text.submit(run_generation, generate_inputs, generate_outputs)
|
| 93 |
button_submit.click(run_generation, generate_inputs, generate_outputs)
|
| 94 |
|