import gradio as gr from multit2i import ( load_models, find_model_list, infer_multi, infer_multi_random, save_gallery_images, change_model, get_model_info_md, loaded_models, get_positive_prefix, get_positive_suffix, get_negative_prefix, get_negative_suffix, get_recom_prompt_type, set_recom_prompt_preset, ) models = [ 'yodayo-ai/kivotos-xl-2.0', 'yodayo-ai/holodayo-xl-2.1', 'cagliostrolab/animagine-xl-3.1', 'votepurchase/ponyDiffusionV6XL', 'eienmojiki/Anything-XL', 'eienmojiki/Starry-XL-v5.2', 'digiplay/majicMIX_sombre_v2', 'digiplay/majicMIX_realistic_v7', 'votepurchase/counterfeitV30_v30', 'Meina/MeinaMix_V11', 'KBlueLeaf/Kohaku-XL-Epsilon-rev3', 'kayfahaarukku/UrangDiffusion-1.1', 'Raelina/Rae-Diffusion-XL-V2', 'Raelina/Raemu-XL-V4', ] # Examples: #models = ['yodayo-ai/kivotos-xl-2.0', 'yodayo-ai/holodayo-xl-2.1'] # specific models #models = find_model_list("John6666", [], "", "last_modified", 20) # John6666's latest 20 models #models = find_model_list("John6666", ["anime"], "", "last_modified", 20) # John6666's latest 20 models with 'anime' tag #models = find_model_list("John6666", [], "anime", "last_modified", 20) # John6666's latest 20 models without 'anime' tag #models = find_model_list("", [], "", "last_modified", 20) # latest 20 text-to-image models of huggingface #models = find_model_list("", [], "", "downloads", 20) # monthly most downloaded 20 text-to-image models of huggingface load_models(models, 10) #load_models(models, 20) # Fetching 20 models at the same time. default: 5 css = """ #model_info { text-align: center; display:block; } """ with gr.Blocks(theme="NoCrypt/miku@>=1.2.2", css=css) as demo: with gr.Column(): with gr.Accordion("Advanced settings", open=False): with gr.Accordion("Recommended Prompt"): recom_prompt_preset = gr.Radio(label="Set Presets", choices=get_recom_prompt_type(), value="Common") positive_prefix = gr.CheckboxGroup(label="Use Positive Prefix", choices=get_positive_prefix(), value=[]) positive_suffix = gr.CheckboxGroup(label="Use Positive Suffix", choices=get_positive_suffix(), value=["Common"]) negative_prefix = gr.CheckboxGroup(label="Use Negative Prefix", choices=get_negative_prefix(), value=[], visible=False) negative_suffix = gr.CheckboxGroup(label="Use Negative Suffix", choices=get_negative_suffix(), value=["Common"], visible=False) with gr.Group(): model_name = gr.Dropdown(label="Select Model", choices=list(loaded_models.keys()), value=list(loaded_models.keys())[0]) model_info = gr.Markdown(value=get_model_info_md(list(loaded_models.keys())[0]), elem_id="model_info") prompt = gr.Text(label="Prompt", lines=1, max_lines=8, placeholder="1girl, solo, ...") neg_prompt = gr.Text(label="Negative Prompt", lines=1, max_lines=8, placeholder="", visible=False) with gr.Row(): run_button = gr.Button("Generate Image", scale=6) random_button = gr.Button("Random Model 🎲", scale=3) image_num = gr.Number(label="Count", minimum=1, maximum=16, value=1, step=1, interactive=True, scale=1) results = gr.Gallery(label="Gallery", interactive=False, show_download_button=True, show_share_button=False, container=True, format="png", object_fit="contain") image_files = gr.Files(label="Download", interactive=False) clear_results = gr.Button("Clear Gallery / Download") examples = gr.Examples( examples = [ ["souryuu asuka langley, 1girl, neon genesis evangelion, plugsuit, pilot suit, red bodysuit, sitting, crossing legs, black eye patch, cat hat, throne, symmetrical, looking down, from bottom, looking at viewer, outdoors"], ["sailor moon, magical girl transformation, sparkles and ribbons, soft pastel colors, crescent moon motif, starry night sky background, shoujo manga style"], ["kafuu chino, 1girl, solo"], ["1girl"], ["beautiful sunset"], ], inputs=[prompt], ) gr.Markdown( f"""This demo was created in reference to the following demos. - [Nymbo/Flood](https://huggingface.co/spaces/Nymbo/Flood). - [Yntec/ToyWorldXL](https://huggingface.co/spaces/Yntec/ToyWorldXL).
The first startup takes a mind-boggling amount of time, but not so much after the second. This is due to the time it takes for Gradio to generate an example image to cache. """ ) gr.DuplicateButton(value="Duplicate Space") model_name.change(change_model, [model_name], [model_info], queue=False, show_api=False) gr.on( triggers=[run_button.click, prompt.submit], fn=infer_multi, inputs=[prompt, neg_prompt, results, image_num, model_name, positive_prefix, positive_suffix, negative_prefix, negative_suffix], outputs=[results], queue=True, show_progress="full", show_api=True, ).success(save_gallery_images, [results], [results, image_files], queue=False, show_api=False) gr.on( triggers=[random_button.click], fn=infer_multi_random, inputs=[prompt, neg_prompt, results, image_num, positive_prefix, positive_suffix, negative_prefix, negative_suffix], outputs=[results], queue=True, show_progress="full", show_api=True, ).success(save_gallery_images, [results], [results, image_files], queue=False, show_api=False) clear_results.click(lambda: (None, None), None, [results, image_files], queue=False, show_api=False) recom_prompt_preset.change(set_recom_prompt_preset, [recom_prompt_preset], [positive_prefix, positive_suffix, negative_prefix, negative_suffix], queue=False, show_api=False) demo.queue() demo.launch()