Spaces:
Sleeping
Sleeping
Zwea Htet
commited on
Commit
•
38bc9e2
1
Parent(s):
9839b9f
added chat history
Browse files- .gitignore +2 -1
- app.py +5 -2
- models/bloom.py +3 -1
.gitignore
CHANGED
@@ -2,4 +2,5 @@ venv
|
|
2 |
data/__pycache__
|
3 |
models/__pycache__
|
4 |
.env
|
5 |
-
__pycache__
|
|
|
|
2 |
data/__pycache__
|
3 |
models/__pycache__
|
4 |
.env
|
5 |
+
__pycache__
|
6 |
+
vectorStores
|
app.py
CHANGED
@@ -31,9 +31,12 @@ if input_text is not None:
|
|
31 |
st.session_state.messages.append(('User', input_text))
|
32 |
with st.spinner("Processing your query..."):
|
33 |
bot_response = get_response(index, input_text)
|
34 |
-
|
35 |
st.session_state.messages.append(('Bot', bot_response))
|
36 |
|
37 |
# Display previous messages
|
|
|
38 |
for sender, msg in st.session_state.messages[::-1]:
|
39 |
-
|
|
|
|
|
|
31 |
st.session_state.messages.append(('User', input_text))
|
32 |
with st.spinner("Processing your query..."):
|
33 |
bot_response = get_response(index, input_text)
|
34 |
+
print("bot: ", bot_response)
|
35 |
st.session_state.messages.append(('Bot', bot_response))
|
36 |
|
37 |
# Display previous messages
|
38 |
+
msg_key = 0
|
39 |
for sender, msg in st.session_state.messages[::-1]:
|
40 |
+
is_user = sender == "User"
|
41 |
+
message(str(msg), is_user, key=str(msg_key)+f'_{sender}')
|
42 |
+
msg_key += 1
|
models/bloom.py
CHANGED
@@ -2,6 +2,7 @@ import os
|
|
2 |
from json import dumps, loads
|
3 |
|
4 |
import numpy as np
|
|
|
5 |
import pandas as pd
|
6 |
from dotenv import load_dotenv
|
7 |
from llama_index import (Document, GPTVectorStoreIndex, LLMPredictor,
|
@@ -12,6 +13,7 @@ from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
|
|
12 |
from utils.customLLM import CustomLLM
|
13 |
|
14 |
load_dotenv()
|
|
|
15 |
|
16 |
# get model
|
17 |
# model_name = "bigscience/bloom-560m"
|
@@ -68,7 +70,7 @@ def initialize_index(index_name):
|
|
68 |
storage_context = StorageContext.from_defaults(persist_dir=file_path)
|
69 |
# load index
|
70 |
index = load_index_from_storage(storage_context)
|
71 |
-
return
|
72 |
else:
|
73 |
documents = prepare_data(r"./assets/regItems.json")
|
74 |
index = GPTVectorStoreIndex.from_documents(documents, service_context=service_context)
|
|
|
2 |
from json import dumps, loads
|
3 |
|
4 |
import numpy as np
|
5 |
+
import openai
|
6 |
import pandas as pd
|
7 |
from dotenv import load_dotenv
|
8 |
from llama_index import (Document, GPTVectorStoreIndex, LLMPredictor,
|
|
|
13 |
from utils.customLLM import CustomLLM
|
14 |
|
15 |
load_dotenv()
|
16 |
+
openai.api_key = os.getenv("OPENAI_API_KEY")
|
17 |
|
18 |
# get model
|
19 |
# model_name = "bigscience/bloom-560m"
|
|
|
70 |
storage_context = StorageContext.from_defaults(persist_dir=file_path)
|
71 |
# load index
|
72 |
index = load_index_from_storage(storage_context)
|
73 |
+
return index
|
74 |
else:
|
75 |
documents = prepare_data(r"./assets/regItems.json")
|
76 |
index = GPTVectorStoreIndex.from_documents(documents, service_context=service_context)
|