-
Notifications
You must be signed in to change notification settings - Fork 10.9k
Support video tiny VAEs #10884
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Support video tiny VAEs #10884
Conversation
|
Great pr,very useful. how about also lightvaew2_1? lighttaew2_1 sometimes can have very bad results, i think lightvaew2_1 is an alternative |
Only first frame for now as live preview playback is currently only available through VHS custom nodes.
Yeah sure, that was simple enough. |
comfy/sd.py
Outdated
| elif "taesd_decoder.1.weight" in sd: | ||
| self.latent_channels = sd["taesd_decoder.1.weight"].shape[1] | ||
| self.first_stage_model = comfy.taesd.taesd.TAESD(latent_channels=self.latent_channels) | ||
| elif "decoder.22.bias" in sd: # taehv, taew and lighttae |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
You should put this whole elif block at the end of the elif blocks so that we can be sure it doesn't accidentally get picked up by any other VAEs.
…g rate for lightvae
|
Thanks so much 👍,I tried hunyuan, very fast(using decode instead of tiled,about 30s to 3s), having some quality loss but good enought.Do they have encoder function, can not encode right now. A little confusing is that i have to copy vae to 'vae_approx' foler to let it work |
It's the folder for tiny VAEs yeah, as they are picked from there for the sampler preview when TAESD preview mode is enabled. The light TAEs generally doesn't seem to support encoding, the original code only had decode function for them as well, or at least I don't know how to use them for that, but Light VAE works for encoding, as does taehv. |
|
I see. By the way , can you support SteadyDancer using native comfyui workflow, very similar to wan animate. character consistency looks great |
Is the tiny VAE for Hunyuan Video 1.5 and Wan Video 2.1 the one from the link below? https://huggingface.co/lightx2v/Autoencoders/blob/main/lighttaehy1_5.safetensors https://huggingface.co/lightx2v/Autoencoders/blob/main/lighttaew2_1.safetensors |
|
yes |
Thanks. What are the differences among the three types of VAEs: lighttae, lightvae, and tae? Which of them can be used in the current ‘load vae’ node? |
All of the video Tiny VAEs can now be used with that node, some of those do not support encoding though, just decoding. The differences are explained in that link. |
Thank you, I see it. The quality ranking is official VAE > lightVAE > lightTAE > TAE. |
Adds support to use these VAEs like normal VAEs (though some of them do not support encoding):
Also added support for sampler (first frame) preview with these TAEs.
https://github.com/madebyollin/taehv/blob/main/taehv.pth
Will upload .safetensors later
Lightx2v light TAEs:
Lightx2v pruned VAE:
Memory estimations aren't that relevant, but I did my best estimations as huge frame counts still use lots of VRAM.
Some of the VAEs need scaling, some don't... this was determined by trial and error.