Skip to content

Conversation

@kijai
Copy link
Contributor

@kijai kijai commented Nov 25, 2025

Adds support to use these VAEs like normal VAEs (though some of them do not support encoding):
Also added support for sampler (first frame) preview with these TAEs.

Lightx2v light TAEs:

  • lighttaew: Wan 2.1 and 2.2
  • lighttaehy1_5: HunyuanVideo 1.5

Lightx2v pruned VAE:

  • lightvaew2_1

Memory estimations aren't that relevant, but I did my best estimations as huge frame counts still use lots of VRAM.
Some of the VAEs need scaling, some don't... this was determined by trial and error.

@kijai kijai requested a review from Kosinkadink as a code owner November 25, 2025 18:52
@zwukong
Copy link

zwukong commented Nov 26, 2025

Great pr,very useful. how about also lightvaew2_1? lighttaew2_1 sometimes can have very bad results, i think lightvaew2_1 is an alternative

Only first frame for now as live preview playback is currently only available through VHS custom nodes.
@kijai
Copy link
Contributor Author

kijai commented Nov 26, 2025

Great pr,very useful. how about also lightvaew2_1? lighttaew2_1 sometimes can have very bad results, i think lightvaew2_1 is an alternative

Yeah sure, that was simple enough.

comfy/sd.py Outdated
elif "taesd_decoder.1.weight" in sd:
self.latent_channels = sd["taesd_decoder.1.weight"].shape[1]
self.first_stage_model = comfy.taesd.taesd.TAESD(latent_channels=self.latent_channels)
elif "decoder.22.bias" in sd: # taehv, taew and lighttae
Copy link
Owner

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

You should put this whole elif block at the end of the elif blocks so that we can be sure it doesn't accidentally get picked up by any other VAEs.

@zwukong
Copy link

zwukong commented Nov 27, 2025

Thanks so much 👍,I tried hunyuan, very fast(using decode instead of tiled,about 30s to 3s), having some quality loss but good enought.Do they have encoder function, can not encode right now. A little confusing is that i have to copy vae to 'vae_approx' foler to let it work

@kijai
Copy link
Contributor Author

kijai commented Nov 27, 2025

Thanks so much 👍,I tried hunyuan, very fast(using decode instead of tiled,about 30s to 3s), having some quality loss but good enought.Do they have encoder function, can not encode right now. A little confusing is that i have to copy vae to 'vae_approx' foler to let it work

It's the folder for tiny VAEs yeah, as they are picked from there for the sampler preview when TAESD preview mode is enabled. The light TAEs generally doesn't seem to support encoding, the original code only had decode function for them as well, or at least I don't know how to use them for that, but Light VAE works for encoding, as does taehv.

@zwukong
Copy link

zwukong commented Nov 27, 2025

I see. By the way , can you support SteadyDancer using native comfyui workflow, very similar to wan animate. character consistency looks great

@comfyanonymous comfyanonymous merged commit b907085 into comfyanonymous:master Nov 29, 2025
10 checks passed
@Amazon90
Copy link

Amazon90 commented Nov 29, 2025

Thanks so much 👍,I tried hunyuan, very fast(using decode instead of tiled,about 30s to 3s), having some quality loss but good enought.Do they have encoder function, can not encode right now. A little confusing is that i have to copy vae to 'vae_approx' foler to let it work

Is the tiny VAE for Hunyuan Video 1.5 and Wan Video 2.1 the one from the link below?

https://huggingface.co/lightx2v/Autoencoders/blob/main/lighttaehy1_5.safetensors

https://huggingface.co/lightx2v/Autoencoders/blob/main/lighttaew2_1.safetensors

@zwukong
Copy link

zwukong commented Nov 29, 2025

yes

@Amazon90
Copy link

Amazon90 commented Nov 29, 2025

yes

Thanks.

What are the differences among the three types of VAEs: lighttae, lightvae, and tae? Which of them can be used in the current ‘load vae’ node?

https://huggingface.co/lightx2v/Autoencoders/

@kijai
Copy link
Contributor Author

kijai commented Nov 29, 2025

yes

Thanks.

What are the differences among the three types of VAEs: lighttae, lightvae, and tae? Which of them can be used in the current ‘load vae’ node?

https://huggingface.co/lightx2v/Autoencoders/

All of the video Tiny VAEs can now be used with that node, some of those do not support encoding though, just decoding.

The differences are explained in that link.

@Amazon90
Copy link

yes

Thanks.
What are the differences among the three types of VAEs: lighttae, lightvae, and tae? Which of them can be used in the current ‘load vae’ node?
https://huggingface.co/lightx2v/Autoencoders/

All of the video Tiny VAEs can now be used with that node, some of those do not support encoding though, just decoding.

The differences are explained in that link.

Thank you, I see it. The quality ranking is official VAE > lightVAE > lightTAE > TAE.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants