Update app.py
Browse files
app.py
CHANGED
|
@@ -391,7 +391,7 @@ def generate_auswahl(prompt, file, chatbot, history, rag_option, model_option, o
|
|
| 391 |
image = Image.open(io.BytesIO(result))
|
| 392 |
image_64 = umwandeln_fuer_anzeige(image)
|
| 393 |
chatbot[-1][1] = "<img src='data:image/png;base64,{0}'/>".format(b64encode(image_64).decode('utf-8'))
|
| 394 |
-
history[
|
| 395 |
print("history zeichnen......................")
|
| 396 |
print(chatbot)
|
| 397 |
return chatbot, history, "Success"
|
|
@@ -399,7 +399,11 @@ def generate_auswahl(prompt, file, chatbot, history, rag_option, model_option, o
|
|
| 399 |
result = generate_text(prompt, file, chatbot, history, rag_option, model_option, openai_api_key, k=3, top_p=0.6, temperature=0.5, max_new_tokens=4048, max_context_length_tokens=2048, repetition_penalty=1.3,)
|
| 400 |
#Antwort als Stream ausgeben... wenn Textantwort gefordert
|
| 401 |
chatbot[-1][1] = result
|
| 402 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 403 |
print("history nach Zusatz und mit KI Antwort...........")
|
| 404 |
print(chatbot)
|
| 405 |
return chatbot, history, "Success"
|
|
@@ -439,7 +443,9 @@ def generate_text (prompt, file, chatbot, history, rag_option, model_option, ope
|
|
| 439 |
if (file == None):
|
| 440 |
history_text_und_prompt = generate_prompt_with_history_openai(prompt, history)
|
| 441 |
else:
|
| 442 |
-
|
|
|
|
|
|
|
| 443 |
print("prompt_neu............................")
|
| 444 |
print(prompt_neu)
|
| 445 |
history_text_und_prompt = generate_prompt_with_history_openai(prompt_neu, history)
|
|
|
|
| 391 |
image = Image.open(io.BytesIO(result))
|
| 392 |
image_64 = umwandeln_fuer_anzeige(image)
|
| 393 |
chatbot[-1][1] = "<img src='data:image/png;base64,{0}'/>".format(b64encode(image_64).decode('utf-8'))
|
| 394 |
+
history = history + [[prompt, b64encode(image_64).decode('utf-8')]]
|
| 395 |
print("history zeichnen......................")
|
| 396 |
print(chatbot)
|
| 397 |
return chatbot, history, "Success"
|
|
|
|
| 399 |
result = generate_text(prompt, file, chatbot, history, rag_option, model_option, openai_api_key, k=3, top_p=0.6, temperature=0.5, max_new_tokens=4048, max_context_length_tokens=2048, repetition_penalty=1.3,)
|
| 400 |
#Antwort als Stream ausgeben... wenn Textantwort gefordert
|
| 401 |
chatbot[-1][1] = result
|
| 402 |
+
if (file == None):
|
| 403 |
+
history = history + [[prompt, result]]
|
| 404 |
+
else:
|
| 405 |
+
history = history + [[prompt+"hier noch die URL zum File", result]]
|
| 406 |
+
|
| 407 |
print("history nach Zusatz und mit KI Antwort...........")
|
| 408 |
print(chatbot)
|
| 409 |
return chatbot, history, "Success"
|
|
|
|
| 443 |
if (file == None):
|
| 444 |
history_text_und_prompt = generate_prompt_with_history_openai(prompt, history)
|
| 445 |
else:
|
| 446 |
+
print("file....................")
|
| 447 |
+
print(file)
|
| 448 |
+
prompt_neu = prompt + b64encode(file).decode("utf-8").path()
|
| 449 |
print("prompt_neu............................")
|
| 450 |
print(prompt_neu)
|
| 451 |
history_text_und_prompt = generate_prompt_with_history_openai(prompt_neu, history)
|