Shariar00's picture
update tokenizer
ec75ddf verified
raw
history blame
1.62 kB
import gradio as gr
from transformers import pipeline, AutoModelForSeq2SeqLM, AutoTokenizer
# Assuming you have loaded your model and tokenizer
# Replace this with your actual model and tokenizer
# Define the model function for Gradio
def generate_summary(input_text):
# # Tokenize the input text
# inputs = tokenizer(input_text, return_tensors="pt", max_length=512, truncation=True)
# # Generate summary using the model
# outputs = model.generate(**inputs)
# # Decode the generated summary
# summary = tokenizer.decode(outputs[0], skip_special_tokens=True)
# return summary
# Create a text generation pipeline
# text_generation_pipeline = pipeline("Falconsai/medical_summarization", model=model, tokenizer=tokenizer)
tokenizer = AutoTokenizer.from_pretrained("Falconsai/medical_summarization")
model = AutoModelForSeq2SeqLM.from_pretrained("Shariar00/medical_summarization_finetune_medical_qa")
text_generation_pipeline = pipeline("summarization", model=model, tokenizer=tokenizer)
# Generate text using the pipeline
prompt = "Hello, I am feeling very pain on my leg, I can not walk properly. I have some knee pain also. what can I do now?"
output = text_generation_pipeline(input_text, max_length=512, num_return_sequences=1)
# Print the generated text
generated_text = output[0]
return generated_text
# Create a Gradio interface
iface = gr.Interface(
fn=generate_summary,
inputs="text",
outputs="text",
# Set to True for live updates without restarting the server
)
# Launch the Gradio interface
iface.launch()