mishig HF staff commited on
Commit
9383422
1 Parent(s): 1710a0f

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +1 -7
app.py CHANGED
@@ -37,15 +37,9 @@ Here are relevant extract from docs that you can use to generate the answer:
37
 
38
  messages.append({"role": "user", "content": processed_prompt})
39
 
40
- generate_kwargs = dict(
41
- temperature=0.6,
42
- max_tokens=8192,
43
- top_p=0.95,
44
- )
45
-
46
  output = ""
47
 
48
- for token in inference_client.chat_completion(messages, stream=True, **generate_kwargs):
49
  new_content = token.choices[0].delta.content
50
  output += new_content
51
  yield output + f"\n\nsources: {sourced_md}"
 
37
 
38
  messages.append({"role": "user", "content": processed_prompt})
39
 
 
 
 
 
 
 
40
  output = ""
41
 
42
+ for token in inference_client.chat_completion(messages, stream=True):
43
  new_content = token.choices[0].delta.content
44
  output += new_content
45
  yield output + f"\n\nsources: {sourced_md}"