Update readme.md

This commit is contained in:
Jukka Seppänen 2024-10-23 22:38:47 +03:00 committed by GitHub
parent 426c35d6b0
commit 754085eaf5
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194

View File

@ -11,6 +11,14 @@ Not sure if this can be worked around, I compiled a wheel for my Windows setup (
https://huggingface.co/Kijai/Mochi_preview_comfy/blob/main/flash_attn-2.6.3-cp312-torch250cu125-win_amd64.whl
Python 3.10 / CUDA 12.4 / Torch 2.4.1:
https://huggingface.co/Kijai/Mochi_preview_comfy/blob/main/flash_attn-2.6.3-cp310-cp310-win_amd64.whl
Other sources for pre-compiled wheels:
https://github.com/oobabooga/flash-attention/releases
Depending on frame count can fit under 20GB, VAE decoding is heavy and there is experimental tiled decoder (taken from CogVideoX -diffusers code) which allows higher frame counts, so far highest I've done is 97 with the default tile size 2x2 grid.
Models: