Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -91,14 +91,14 @@ class ModelWrapper:
|
|
91 |
|
92 |
DTYPE = prompt_embed.dtype
|
93 |
print(DTYPE)
|
94 |
-
print(
|
95 |
-
print(
|
96 |
-
print(
|
97 |
|
98 |
for constant in all_timesteps:
|
99 |
current_timesteps = torch.ones(len(prompt_embed), device="cuda", dtype=torch.long) * constant
|
100 |
eval_images = self.model(noise, current_timesteps, prompt_embed, added_cond_kwargs=unet_added_conditions).sample
|
101 |
-
print(
|
102 |
|
103 |
eval_images = get_x0_from_noise(noise, eval_images, alphas_cumprod, current_timesteps).to(self.DTYPE)
|
104 |
|
|
|
91 |
|
92 |
DTYPE = prompt_embed.dtype
|
93 |
print(DTYPE)
|
94 |
+
print(type(noise))
|
95 |
+
print(type(current_timesteps))
|
96 |
+
print(type(unet_added_conditions))
|
97 |
|
98 |
for constant in all_timesteps:
|
99 |
current_timesteps = torch.ones(len(prompt_embed), device="cuda", dtype=torch.long) * constant
|
100 |
eval_images = self.model(noise, current_timesteps, prompt_embed, added_cond_kwargs=unet_added_conditions).sample
|
101 |
+
print(type(eval_images))
|
102 |
|
103 |
eval_images = get_x0_from_noise(noise, eval_images, alphas_cumprod, current_timesteps).to(self.DTYPE)
|
104 |
|