Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -338,14 +338,15 @@ def generate(text, history, rag_option, model_option, k=3, top_p=0.6, temperatu
|
|
338 |
result = response.json()
|
339 |
print("result:------------------")
|
340 |
chatbot_response = result[0]['generated_text']
|
341 |
-
print(chatbot_response)
|
342 |
print("anzahl tokens gesamt antwort:------------------")
|
343 |
print (len(chatbot_response.split()))
|
344 |
except Exception as e:
|
345 |
raise gr.Error(e)
|
346 |
|
347 |
chatbot_message = chatbot_response[len(prompt):].strip()
|
348 |
-
|
|
|
|
|
349 |
|
350 |
#Antwort als Stream ausgeben...
|
351 |
for i in range(len(chatbot_message)):
|
|
|
338 |
result = response.json()
|
339 |
print("result:------------------")
|
340 |
chatbot_response = result[0]['generated_text']
|
|
|
341 |
print("anzahl tokens gesamt antwort:------------------")
|
342 |
print (len(chatbot_response.split()))
|
343 |
except Exception as e:
|
344 |
raise gr.Error(e)
|
345 |
|
346 |
chatbot_message = chatbot_response[len(prompt):].strip()
|
347 |
+
print("history/chatbot_rsponse:--------------------------------")
|
348 |
+
print(history)
|
349 |
+
print(chatbot_message)
|
350 |
|
351 |
#Antwort als Stream ausgeben...
|
352 |
for i in range(len(chatbot_message)):
|