ford442 commited on
Commit
a6d47ca
·
1 Parent(s): 735e9bb

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +8 -5
app.py CHANGED
@@ -22,9 +22,10 @@ import paramiko
22
  import gc
23
  import time
24
  import datetime
25
- from diffusers.schedulers import AysSchedules
26
 
27
  from gradio import themes
 
28
 
29
  torch.backends.cuda.matmul.allow_tf32 = False
30
  torch.backends.cuda.matmul.allow_bf16_reduced_precision_reduction = False
@@ -91,7 +92,7 @@ DEFAULT_STYLE_NAME = "Style Zero"
91
  STYLE_NAMES = list(styles.keys())
92
  HF_TOKEN = os.getenv("HF_TOKEN")
93
 
94
- sampling_schedule = AysSchedules["StableDiffusionXLTimesteps"]
95
 
96
  def apply_style(style_name: str, positive: str, negative: str = "") -> Tuple[str, str]:
97
  if style_name in styles:
@@ -149,6 +150,8 @@ def load_and_prepare_model(model_id):
149
  pipe.to(device)
150
  pipe.to(torch.bfloat16)
151
 
 
 
152
  pipe.unet.set_default_attn_processor()
153
  pipe.vae.set_default_attn_processor()
154
 
@@ -229,7 +232,7 @@ def generate_30(
229
  "guidance_scale": guidance_scale,
230
  "num_inference_steps": num_inference_steps,
231
  "generator": generator,
232
- "timesteps": sampling_schedule,
233
  "output_type": "pil",
234
  }
235
  if use_resolution_binning:
@@ -296,7 +299,7 @@ def generate_60(
296
  "guidance_scale": guidance_scale,
297
  "num_inference_steps": num_inference_steps,
298
  "generator": generator,
299
- "timesteps": sampling_schedule,
300
  "output_type": "pil",
301
  }
302
  if use_resolution_binning:
@@ -363,7 +366,7 @@ def generate_90(
363
  "guidance_scale": guidance_scale,
364
  "num_inference_steps": num_inference_steps,
365
  "generator": generator,
366
- "timesteps": sampling_schedule,
367
  "output_type": "pil",
368
  }
369
  if use_resolution_binning:
 
22
  import gc
23
  import time
24
  import datetime
25
+ #from diffusers.schedulers import AysSchedules
26
 
27
  from gradio import themes
28
+ from hidiffusion import apply_hidiffusion, remove_hidiffusion
29
 
30
  torch.backends.cuda.matmul.allow_tf32 = False
31
  torch.backends.cuda.matmul.allow_bf16_reduced_precision_reduction = False
 
92
  STYLE_NAMES = list(styles.keys())
93
  HF_TOKEN = os.getenv("HF_TOKEN")
94
 
95
+ #sampling_schedule = AysSchedules["StableDiffusionXLTimesteps"]
96
 
97
  def apply_style(style_name: str, positive: str, negative: str = "") -> Tuple[str, str]:
98
  if style_name in styles:
 
150
  pipe.to(device)
151
  pipe.to(torch.bfloat16)
152
 
153
+ apply_hidiffusion(pipe)
154
+
155
  pipe.unet.set_default_attn_processor()
156
  pipe.vae.set_default_attn_processor()
157
 
 
232
  "guidance_scale": guidance_scale,
233
  "num_inference_steps": num_inference_steps,
234
  "generator": generator,
235
+ # "timesteps": sampling_schedule,
236
  "output_type": "pil",
237
  }
238
  if use_resolution_binning:
 
299
  "guidance_scale": guidance_scale,
300
  "num_inference_steps": num_inference_steps,
301
  "generator": generator,
302
+ # "timesteps": sampling_schedule,
303
  "output_type": "pil",
304
  }
305
  if use_resolution_binning:
 
366
  "guidance_scale": guidance_scale,
367
  "num_inference_steps": num_inference_steps,
368
  "generator": generator,
369
+ # "timesteps": sampling_schedule,
370
  "output_type": "pil",
371
  }
372
  if use_resolution_binning: