Spaces:
Sleeping
Sleeping
import gradio as gr | |
from transformers import AutoModelForCausalLM | |
model = AutoModelForCausalLM.from_pretrained("automaise/quokka-7b") | |
def create_gradio_app(model, tokenizer): | |
def gradio_fn(question): | |
inputs = tokenizer(question, return_tensors="pt") | |
outputs = model(inputs)[0] | |
generated_text = tokenizer.decode(outputs[0], skip_special_tokens=True) | |
return generated_text | |
iface = gr.Interface( | |
fn=gradio_fn, | |
inputs=gr.inputs.Textbox(), | |
outputs=gr.outputs.Textbox(), | |
live=True, | |
title="Gradio App", | |
description="Create a gradio app using the Quokka-7b model.", | |
) | |
return iface | |
iface = create_gradio_app(model, tokenizer) | |
if __name__ == "__main__": | |
iface.launch() | |