alexkueck commited on
Commit
8f3283d
1 Parent(s): abe3764

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +13 -5
app.py CHANGED
@@ -292,15 +292,23 @@ def generate(text, history, rag_option, model_option, temperature=0.5, max_new_
292
  #payload = tokenizer.apply_chat_template(prompt,tokenize=False)
293
  #result = client.text_generation(payload, do_sample=True,return_full_text=False, max_new_tokens=2048,top_p=0.9,temperature=0.6,)
294
  #inference allg:
295
- response= requests.post(API_URL, headers=HEADERS, json={"inputs": prompt})
296
- result = response.json()
297
- print("result:------------------")
298
- print(result[0]['generated_text'])
 
 
 
 
 
 
 
 
299
 
300
  except Exception as e:
301
  raise gr.Error(e)
302
 
303
- return result[0]['generated_text']
304
 
305
  #Antwort als Stream ausgeben...
306
  #for i in range(len(result)):
 
292
  #payload = tokenizer.apply_chat_template(prompt,tokenize=False)
293
  #result = client.text_generation(payload, do_sample=True,return_full_text=False, max_new_tokens=2048,top_p=0.9,temperature=0.6,)
294
  #inference allg:
295
+ data = {
296
+ "inputs": prompt,
297
+ "options": {"max_new_tokens": 1024},
298
+ }
299
+ response= requests.post(API_URL, headers=HEADERS, data=json.dumps(data))
300
+ #result = response.json()
301
+ #print("result:------------------")
302
+ #print(result[0]['generated_text'])
303
+ if response.status_code == 200:
304
+ result = json.loads(response.text)["generated_text"]
305
+ else:
306
+ raise Exception(f"Request failed with status code {response.status_code}: {response.text}")
307
 
308
  except Exception as e:
309
  raise gr.Error(e)
310
 
311
+ return result #result[0]['generated_text']
312
 
313
  #Antwort als Stream ausgeben...
314
  #for i in range(len(result)):