#!/usr/bin/env python | |
import os | |
import gradio as gr | |
import torch | |
from app_image_to_3d import create_demo as create_demo_image_to_3d | |
from app_text_to_3d import create_demo as create_demo_text_to_3d | |
from model import Model | |
DESCRIPTION = '# [Shap-E](https://github.com/openai/shap-e)' | |
if (SPACE_ID := os.getenv('SPACE_ID')) is not None: | |
DESCRIPTION += f'\n<p>For faster inference without waiting in queue, you may duplicate the space and upgrade to GPU in settings. <a href="https://huggingface.co/spaces/{SPACE_ID}?duplicate=true"><img style="display: inline; margin-top: 0em; margin-bottom: 0em" src="https://bit.ly/3gLdBN6" alt="Duplicate Space" /></a></p>' | |
if not torch.cuda.is_available(): | |
DESCRIPTION += '\n<p>Running on CPU 🥶 This demo does not work on CPU.</p>' | |
model = Model() | |
with gr.Blocks(css='style.css') as demo: | |
gr.Markdown(DESCRIPTION) | |
with gr.Tabs(): | |
with gr.Tab(label='Text to 3D'): | |
create_demo_text_to_3d(model) | |
with gr.Tab(label='Image to 3D'): | |
create_demo_image_to_3d(model) | |
demo.queue(api_open=False, max_size=10).launch() | |