Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -162,7 +162,7 @@ def generate(
|
|
162 |
width: int = 768,
|
163 |
height: int = 768,
|
164 |
guidance_scale: float = 5,
|
165 |
-
num_inference_steps: int =
|
166 |
randomize_seed: bool = False,
|
167 |
use_resolution_binning: bool = True,
|
168 |
num_images: int = 1,
|
@@ -171,7 +171,7 @@ def generate(
|
|
171 |
global models
|
172 |
pipe = models[model_choice]
|
173 |
seed = int(randomize_seed_fn(seed, randomize_seed))
|
174 |
-
generator = torch.Generator(device='
|
175 |
|
176 |
prompt, negative_prompt = apply_style(style_selection, prompt, negative_prompt)
|
177 |
|
@@ -366,7 +366,7 @@ with gr.Blocks(css=css, theme="bethecloud/storj_theme") as demo:
|
|
366 |
minimum=10,
|
367 |
maximum=1000,
|
368 |
step=10,
|
369 |
-
value=
|
370 |
)
|
371 |
|
372 |
gr.Examples(
|
|
|
162 |
width: int = 768,
|
163 |
height: int = 768,
|
164 |
guidance_scale: float = 5,
|
165 |
+
num_inference_steps: int = 300,
|
166 |
randomize_seed: bool = False,
|
167 |
use_resolution_binning: bool = True,
|
168 |
num_images: int = 1,
|
|
|
171 |
global models
|
172 |
pipe = models[model_choice]
|
173 |
seed = int(randomize_seed_fn(seed, randomize_seed))
|
174 |
+
generator = torch.Generator(device='cuda').manual_seed(seed)
|
175 |
|
176 |
prompt, negative_prompt = apply_style(style_selection, prompt, negative_prompt)
|
177 |
|
|
|
366 |
minimum=10,
|
367 |
maximum=1000,
|
368 |
step=10,
|
369 |
+
value=300,
|
370 |
)
|
371 |
|
372 |
gr.Examples(
|