-
Notifications
You must be signed in to change notification settings - Fork 94
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Even 80 GB is not sufficient for SANA 4K model VAE decoding - something feels wrong? #139
Comments
even A100 - 80 GB fails
|
try this, runs on an A40 |
@geronimi73 thanks but it is not even merged i can't use it - non-technical people installing via my installers @lawrence-cj please consider fixing the pipeline here |
@FurkanGozukara try to use patch_conv on the vae decoder :) https://github.com/mit-han-lab/patch_conv |
interesting gonna test now so it is model = convert_model(model, splits=4) # The only modification you need to make how many splits should i make? why 4? |
@xieenze i tried and not working no errors but 0 difference
|
@FurkanGozukara try only convert vae? such as below |
thanks i tested here and didnt work but gonna test yours now
|
@xieenze not working either tried 4 / 8 / 16 / 32 / 64 / 128 :( i have rtx 3090 - 24 gb
|
@xieenze i think will be necessary implement a tiledvae module for this, I think SUPIR implement this but i dont know if the arch is similar. |
We are working on it. Will fix when this PR is merged. |
so you are not going to update SANA pipeline here? we have to move to diffusers i guess? @lawrence-cj can you add a diffusers demo with all features here? https://github.com/NVlabs/Sana/tree/main/app |
Tried to allocate 36.00 GiB. GPU 0 has a total capacity of 44.52 GiB of which 30.29 GiB is free
Testing on L40S and still fails
Using official pipeline : https://github.com/NVlabs/Sana/blob/main/app/sana_pipeline.py
During inference it uses around 18 GB VRAM but VAE decode is causing issue
SANA 2K and 1K works great
The text was updated successfully, but these errors were encountered: