Spaces:
Running
on
T4
Running
on
T4
Update app.py
Browse files
app.py
CHANGED
@@ -11,11 +11,11 @@ upscaler = StableDiffusionLatentUpscalePipeline.from_pretrained("stabilityai/sd-
|
|
11 |
upscaler = upscaler.to(device)
|
12 |
pipe = pipe.to(device)
|
13 |
|
14 |
-
def genie (
|
15 |
generator = torch.Generator(device=device).manual_seed(Seed)
|
16 |
#images = pipe(prompt, num_inference_steps=steps, guidance_scale=scale, generator=generator).images[0]
|
17 |
-
low_res_latents = pipe(
|
18 |
-
upscaled_image = upscaler(prompt, image=low_res_latents, num_inference_steps=20, guidance_scale=0, generator=generator).images[0]
|
19 |
return upscaled_image
|
20 |
|
21 |
gr.Interface(fn=genie, inputs=[gr.Textbox(label='What you want the AI to generate. 77 Token Limit.'),
|
|
|
11 |
upscaler = upscaler.to(device)
|
12 |
pipe = pipe.to(device)
|
13 |
|
14 |
+
def genie (Prompt, scale, steps, Seed):
|
15 |
generator = torch.Generator(device=device).manual_seed(Seed)
|
16 |
#images = pipe(prompt, num_inference_steps=steps, guidance_scale=scale, generator=generator).images[0]
|
17 |
+
low_res_latents = pipe(Prompt, num_inference_steps=steps, guidance_scale=scale, generator=generator, output_type="latent").images
|
18 |
+
upscaled_image = upscaler(prompt=' ', image=low_res_latents, num_inference_steps=20, guidance_scale=0, generator=generator).images[0]
|
19 |
return upscaled_image
|
20 |
|
21 |
gr.Interface(fn=genie, inputs=[gr.Textbox(label='What you want the AI to generate. 77 Token Limit.'),
|