model-test / app.py
Abrar20's picture
Update app.py
1ca4c81 verified
raw
history blame
6.47 kB
import gradio as gr
from langchain.document_loaders import PyPDFLoader
from langchain.text_splitter import CharacterTextSplitter
from langchain.llms import HuggingFaceHub
from langchain.embeddings import HuggingFaceHubEmbeddings
from langchain.vectorstores import Chroma
from langchain.chains import RetrievalQA
from langchain.prompts import PromptTemplate
def get_chain(llm, retriever):
prompt_template = """
Instructions:
You are a knowledgeable assistant focused on providing safety guidelines for coastal areas during cyclones. Your goal is to generate personalized, clear, and actionable advice based on the specific details provided about the user's infrastructure, proximity to the cyclone pathway, cyclone speed, and proximity to the nearest shelter, person's location.
Please:
- Carefully analyze the provided context from the PDF.
- Offer tailored guidance that addresses the user's unique situation.
- Calculate the nearest shelter by location of the person (lat, lon) and shelter coordinates (lat, lon).
- Calculate proximity to cyclone by location of the person (lat, lon) and predicted cyclone coordinates (lat, lon).
- Ensure that your advice is practical and directly applicable.
- If information is missing or unclear, use logical assumptions based on the context to provide the best possible recommendations.
- Be concise but thorough, offering detailed steps when necessary to enhance safety and preparedness.
Context:\n{context}\n
Question: \n{question}\n
Personalized Guideline:
"""
PROMPT = PromptTemplate(template=prompt_template, input_variables=["context", "question"])
chain_type_kwargs = {"prompt": PROMPT}
qa_chain = RetrievalQA.from_chain_type(
llm=llm,
chain_type="stuff",
retriever=retriever,
chain_type_kwargs=chain_type_kwargs,
return_source_documents=True
)
return qa_chain
def load_pdf_to_langchain(pdf_path, repo_id):
# Load the PDF using PyPDFLoader
loader = PyPDFLoader(pdf_path)
documents = loader.load()
text_splitter = CharacterTextSplitter(chunk_size=2096, chunk_overlap=0)
texts = text_splitter.split_documents(documents)
embeddings = HuggingFaceHubEmbeddings()
db = Chroma.from_documents(texts, embeddings)
retriever = db.as_retriever()
llm = HuggingFaceHub(
repo_id=repo_id,
model_kwargs={'temperature': 0.3}
)
qa_chain = get_chain(llm, retriever)
return qa_chain
def generate_guideline(infrastructure, location, cyclone_predicted_coordinates, cyclone_speed):
if infrastructure and location and cyclone_predicted_coordinates and cyclone_speed:
user_question = f"""{infrastructure} Infrastructure, location of the person (lat, lon) is {location}, Cyclone Speed in knots is {cyclone_speed}, Predicted Cyclone Coordinates (lat, lon) is {cyclone_predicted_coordinates}. Please give guidelines on what will be best in this context. Give precise instructions by calculating proximity to cyclone by location of the person (lat, lon) and predicted cyclone coordinates (lat, lon). Also, give the location of the nearest shelter by calculating location of the person (lat, lon) and shelter coordinates (lat, lon) (from the text chunk given). Don't give proximity to cyclone and proximity to shelter though (only use this to generate the guideline). Also, give the helpline number at the end: 333."""
result = qa({'query': user_question})
return result['result']
else:
return "Please provide all inputs."
css = """
#col-container {max-width: 700px; margin-left: auto; margin-right: auto;}
"""
title = """
<h1>🌪️ Generate Tailored Cyclone Safety Guidelines</h1>
"""
with gr.Blocks(css=css, theme='Taithrah/Minimal') as demo:
with gr.Column(elem_id='col-container'):
gr.HTML(title)
# LLM selection
repo_id = gr.Dropdown(
label='Select Language Model (LLM)',
choices=[
'mistralai/Mistral-7B-Instruct-v0.1',
'HuggingFaceH4/zephyr-7b-beta',
'meta-llama/Llama-2-7b-chat-hf',
'01-ai/Yi-6B-200K',
'cognitivecomputations/dolphin-2.5-mixtral-8x7b',
'tiiuae/falcon-7b-instruct',
'tiiuae/falcon-7b',
'google/flan-t5-base',
'google/flan-t5-large',
'google/flan-t5-xl',
'bigscience/bloom-7b1',
'EleutherAI/gpt-neo-2.7B',
'EleutherAI/gpt-j-6B',
'facebook/opt-6.7b',
'facebook/galactica-6.7b',
'cerebras/Cerebras-GPT-6.7B',
'RWKV/rwkv-4-7b-pile',
# Add more models here as desired
],
value='mistralai/Mistral-7B-Instruct-v0.1'
)
# Status display
langchain_status = gr.Textbox(
label='Status', placeholder='', interactive=False, value="Loading guideline1.pdf..."
)
# Input fields for user information
infrastructure = gr.Textbox(label='Infrastructure')
location = gr.Textbox(label='Location Coordinates (lat, lon)')
cyclone_predicted_coordinates = gr.Textbox(label='Predicted Cyclone Coordinates (lat, lon)')
cyclone_speed = gr.Textbox(label='Cyclone Speed in Knots')
submit_btn = gr.Button('Generate Guideline')
output = gr.Textbox(label='Personalized Guideline', lines=10)
# Global variable to store the QA chain
qa = None
# Function to initialize the QA chain
def initialize_qa(repo_id_value):
global qa
pdf_path = 'guideline1.pdf' # Ensure this PDF is in the same directory
qa = load_pdf_to_langchain(pdf_path, repo_id_value)
return f"Loaded guideline1.pdf with LLM: {repo_id_value}"
# Initialize QA chain with default LLM
initial_status = initialize_qa(repo_id.value)
langchain_status.value = initial_status
# Update QA chain when LLM selection changes
def on_repo_id_change(repo_id_value):
status = initialize_qa(repo_id_value)
return status
repo_id.change(
on_repo_id_change,
inputs=repo_id,
outputs=langchain_status
)
submit_btn.click(
generate_guideline,
inputs=[infrastructure, location, cyclone_predicted_coordinates, cyclone_speed],
outputs=output
)
demo.launch()