Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -265,7 +265,7 @@ def uploadNote(prompt,num_inference_steps,guidance_scale,timestamp,denoise):
|
|
265 |
|
266 |
@spaces.GPU(duration=40)
|
267 |
def generate_30(
|
268 |
-
prompt: str,
|
269 |
negative_prompt: str = "",
|
270 |
use_negative_prompt: bool = False,
|
271 |
style_selection: str = "",
|
@@ -276,7 +276,7 @@ def generate_30(
|
|
276 |
num_inference_steps: int = 125,
|
277 |
randomize_seed: bool = False,
|
278 |
use_resolution_binning: bool = True,
|
279 |
-
latent_file, # Add latents file input
|
280 |
progress=gr.Progress(track_tqdm=True) # Add progress as a keyword argument
|
281 |
):
|
282 |
#torch.backends.cudnn.benchmark = False
|
@@ -331,7 +331,7 @@ def generate_60(
|
|
331 |
num_inference_steps: int = 250,
|
332 |
randomize_seed: bool = False,
|
333 |
use_resolution_binning: bool = True,
|
334 |
-
latent_file, # Add latents file input
|
335 |
progress=gr.Progress(track_tqdm=True) # Add progress as a keyword argument
|
336 |
):
|
337 |
#torch.backends.cudnn.benchmark = True
|
@@ -386,7 +386,7 @@ def generate_90(
|
|
386 |
num_inference_steps: int = 250,
|
387 |
randomize_seed: bool = False,
|
388 |
use_resolution_binning: bool = True,
|
389 |
-
latent_file, # Add latents file input
|
390 |
progress=gr.Progress(track_tqdm=True) # Add progress as a keyword argument
|
391 |
):
|
392 |
#torch.backends.cudnn.benchmark = True
|
|
|
265 |
|
266 |
@spaces.GPU(duration=40)
|
267 |
def generate_30(
|
268 |
+
prompt: str = "",
|
269 |
negative_prompt: str = "",
|
270 |
use_negative_prompt: bool = False,
|
271 |
style_selection: str = "",
|
|
|
276 |
num_inference_steps: int = 125,
|
277 |
randomize_seed: bool = False,
|
278 |
use_resolution_binning: bool = True,
|
279 |
+
latent_file = None, # Add latents file input
|
280 |
progress=gr.Progress(track_tqdm=True) # Add progress as a keyword argument
|
281 |
):
|
282 |
#torch.backends.cudnn.benchmark = False
|
|
|
331 |
num_inference_steps: int = 250,
|
332 |
randomize_seed: bool = False,
|
333 |
use_resolution_binning: bool = True,
|
334 |
+
latent_file = None, # Add latents file input
|
335 |
progress=gr.Progress(track_tqdm=True) # Add progress as a keyword argument
|
336 |
):
|
337 |
#torch.backends.cudnn.benchmark = True
|
|
|
386 |
num_inference_steps: int = 250,
|
387 |
randomize_seed: bool = False,
|
388 |
use_resolution_binning: bool = True,
|
389 |
+
latent_file = None, # Add latents file input
|
390 |
progress=gr.Progress(track_tqdm=True) # Add progress as a keyword argument
|
391 |
):
|
392 |
#torch.backends.cudnn.benchmark = True
|