8000 GitHub - CRCODE22/ComfyUI-Sana: Unofficial custom-node for SANA: https://github.com/NVlabs/Sana.
[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Skip to content

CRCODE22/ComfyUI-Sana

Repository files navigation

Warning:

  • A init node with lots of bugs, do not try unless interested.
  • ram will not released (tried but failed).
  • 4gb vram cuda device will oom in text_encoder with non-4bit, but retry works.
  • test on rtx cuda-device with win10+py311+torch 2.5.1+cu126.

Known issue:

  • Batch_size not work, it's a loop which i don't think is a good idea, so i left it empty.

Hardware requirements:

  • ram & vram: 16+gb ram, init model needs lots of ram. 4gb vram at least.
  • text_encoder: gemma-2-2b-it ~ 5gb vram, gemma-2-2b-it-bnb-4bit ~ 2.3gb vram.
  • dit: ~ 3.5gb vram.
  • vae: ~ 3.1gb vram for 4k.

workflow in assets dir

Instruction:

dit: download from below links into ComfyUI\models\unet, only 1.6B-1204px works for now.

vae: autodownload or manual download or git from below links into ComfyUI\models\vae, rename folder_name to models--mit-han-lab--dc-ae-f32c32-sana-1.0.

text_encoder: autodownload or manual download or git from below links into ComfyUI\models\text_encoders, rename folder_name to models--unsloth--gemma-2-2b-it.

4bit text_encoder: autodownload or manual download or git from below links into ComfyUI\models\text_encoders, rename folder_name to models--unsloth--gemma-2-2b-it-bnb-4bit. It will remain 1.5gb in vram after text_encode, which can be freed by ComfyUI-Manager Free model and node cache.

About requirements:

  • I only installed 2 modules, so i do not know what are really needed.

About

Unofficial custom-node for SANA: https://github.com/NVlabs/Sana.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

0