derek-thomas HF staff commited on
Commit
37df3b9
1 Parent(s): f1ecb2c

Updating to 4 min.

Browse files
Files changed (1) hide show
  1. app.py +2 -2
app.py CHANGED
@@ -46,7 +46,7 @@ def bot(history, hyde=False):
46
  document_start = perf_counter()
47
  if hyde:
48
  hyde_document = generate(f"Write a wikipedia article intro paragraph to answer this query: {query}").split(
49
- '### Response: [|AI|]')[-1]
50
 
51
  logger.warning(hyde_document)
52
  documents = retriever(hyde_document, top_k=top_k)
@@ -90,7 +90,7 @@ You can see the prompt clearly displayed below the chatbot to understand what is
90
 
91
  # Read this if you get an error
92
  I'm using Inference Endpoint's Scale to Zero to save money on GPUs. If the staus shows its not "Running" send a
93
- chat to wake it up. You will get a `500 error` and it will take ~7 min to wake up.
94
  """
95
 
96
 
 
46
  document_start = perf_counter()
47
  if hyde:
48
  hyde_document = generate(f"Write a wikipedia article intro paragraph to answer this query: {query}").split(
49
+ '### Response: [|AI|]')[-1]
50
 
51
  logger.warning(hyde_document)
52
  documents = retriever(hyde_document, top_k=top_k)
 
90
 
91
  # Read this if you get an error
92
  I'm using Inference Endpoint's Scale to Zero to save money on GPUs. If the staus shows its not "Running" send a
93
+ chat to wake it up. You will get a `500 error` and it will take ~7 minutes to wake up.
94
  """
95
 
96