Spaces:
Running
Running
update duration
Browse files
app.py
CHANGED
@@ -57,14 +57,14 @@ def get_server_status():
|
|
57 |
return {"cpu": f"{cpu_percent}%", "memory": f"{memory.percent}%", "disk": f"{disk.percent}%", "gpu": gpu_info}
|
58 |
|
59 |
|
60 |
-
@spaces.GPU
|
61 |
def generate_vanilla(prompt, num_inference_steps, guidance_scale, progress=gr.Progress(track_tqdm=True)):
|
62 |
engine = load_model()
|
63 |
video_path = generate(engine, prompt, num_inference_steps, guidance_scale)
|
64 |
return video_path
|
65 |
|
66 |
|
67 |
-
@spaces.GPU
|
68 |
def generate_vs(
|
69 |
prompt,
|
70 |
num_inference_steps,
|
|
|
57 |
return {"cpu": f"{cpu_percent}%", "memory": f"{memory.percent}%", "disk": f"{disk.percent}%", "gpu": gpu_info}
|
58 |
|
59 |
|
60 |
+
@spaces.GPU(duration=120)
|
61 |
def generate_vanilla(prompt, num_inference_steps, guidance_scale, progress=gr.Progress(track_tqdm=True)):
|
62 |
engine = load_model()
|
63 |
video_path = generate(engine, prompt, num_inference_steps, guidance_scale)
|
64 |
return video_path
|
65 |
|
66 |
|
67 |
+
@spaces.GPU(duration=120)
|
68 |
def generate_vs(
|
69 |
prompt,
|
70 |
num_inference_steps,
|