Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -176,7 +176,7 @@ class GuiSD:
|
|
176 |
|
177 |
if self.model.base_model_id != model_name:
|
178 |
load_now_time = datetime.now()
|
179 |
-
elapsed_time = (load_now_time - self.last_load).total_seconds()
|
180 |
|
181 |
if elapsed_time <= 8:
|
182 |
print("Waiting for the previous model's time ops...")
|
@@ -205,7 +205,7 @@ class GuiSD:
|
|
205 |
yield f"Model loaded: {model_name}"
|
206 |
|
207 |
# @spaces.GPU(duration=59)
|
208 |
-
|
209 |
def generate_pipeline(
|
210 |
self,
|
211 |
prompt,
|
@@ -519,7 +519,7 @@ class GuiSD:
|
|
519 |
|
520 |
def dynamic_gpu_duration(func, duration, *args):
|
521 |
|
522 |
-
@torch.inference_mode()
|
523 |
@spaces.GPU(duration=duration)
|
524 |
def wrapped_func():
|
525 |
yield from func(*args)
|
@@ -1201,4 +1201,4 @@ app.launch(
|
|
1201 |
show_error=True,
|
1202 |
debug=True,
|
1203 |
allowed_paths=["./images/"],
|
1204 |
-
)
|
|
|
176 |
|
177 |
if self.model.base_model_id != model_name:
|
178 |
load_now_time = datetime.now()
|
179 |
+
elapsed_time = max((load_now_time - self.last_load).total_seconds(), 0)
|
180 |
|
181 |
if elapsed_time <= 8:
|
182 |
print("Waiting for the previous model's time ops...")
|
|
|
205 |
yield f"Model loaded: {model_name}"
|
206 |
|
207 |
# @spaces.GPU(duration=59)
|
208 |
+
@torch.inference_mode()
|
209 |
def generate_pipeline(
|
210 |
self,
|
211 |
prompt,
|
|
|
519 |
|
520 |
def dynamic_gpu_duration(func, duration, *args):
|
521 |
|
522 |
+
# @torch.inference_mode()
|
523 |
@spaces.GPU(duration=duration)
|
524 |
def wrapped_func():
|
525 |
yield from func(*args)
|
|
|
1201 |
show_error=True,
|
1202 |
debug=True,
|
1203 |
allowed_paths=["./images/"],
|
1204 |
+
)
|