-
Notifications
You must be signed in to change notification settings - Fork 5.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Cog] some minor fixes and nits #9466
Conversation
The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Oops 😬 So, we don't have a test that passes both prompt embeds and negative prompt embeds? I believe that would have caught this, no? Surprising and very lucky that it was working for the normal prompt inputs so far.
Apologies for missing this on my part
I believe that could be added but the test would be about checking if an error is thrown if |
Sorry, I have additional updates to make. Will let you all know here after that. |
src/diffusers/pipelines/cogvideo/pipeline_cogvideox_video2video.py
Outdated
Show resolved
Hide resolved
|
||
height = height or self.transformer.config.sample_size * self.vae_scale_factor_spatial | ||
width = width or self.transformer.config.sample_size * self.vae_scale_factor_spatial |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
We don't need it because height
and width
are already at their default values.
self.vae_scaling_factor_image = ( | ||
self.vae.config.scaling_factor if hasattr(self, "vae") and self.vae is not None else 0.7 | ||
) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Is beneficial for scenarios where we want to run the pipeline without the VAE.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks good! Thanks for the fixes
@@ -188,6 +188,9 @@ def __init__( | |||
self.vae_scale_factor_temporal = ( | |||
self.vae.config.temporal_compression_ratio if hasattr(self, "vae") and self.vae is not None else 4 | |||
) | |||
self.vae_scaling_factor_image = ( |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Adding this so that a pipeline can operate with the vae
.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
thanks!
thank you for amazing work i am using the code that is running here : https://huggingface.co/spaces/THUDM/CogVideoX-5B-Space however I am getting below errors both in text to video and also video to video could there have been something broken the pipe?
|
Open a separate issue with a minimal fully reproducible code snippet. https://github.com/user-attachments/files/17161626/app.txt is quite long and is not minimal and the snippet you provided is incomplete. |
@sayakpaul current diffusers doesn't have this? it can't find it installed latest one on Windows 0.30.3 |
v0.30.3 only released the image-to-video and video-to-video pipelines. LoRA support will be added in a future release. For now, you can use the lora loader mixin via the main branch |
* fix positional arguments in check_inputs(). * add video and latetns to check_inputs(). * prep latents_in_channels. * quality * multiple fixes. * fix
* fix positional arguments in check_inputs(). * add video and latetns to check_inputs(). * prep latents_in_channels. * quality * multiple fixes. * fix
What does this PR do?
check_inputs()
prepare_latents()
at the beginning of the method so that we can error as early as possible.