Replies: 1 comment
-
Hi @Can-Zhao, could you please help share some comments here? Thanks. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi!
The MAISI paper describes how feature maps are split into segments, and each segment is allocated to a different device to perform convolutions. And in
autoencoderkl_maisi.py
, I do see that tensor splitting is done on Line 247, and convolutions are performed on each segment on Line 253. However, correct me if I'm wrong, but it seems that convolutions are all performed on the same device. I don't see an option to specify multiple devices, according to how TSP is described in the paper. If there doesn't exist this option, will it be added in a future release?I see it's reported in the MAISI README that for 256x256x256 images with latent size of 4x64x64x64, peak training memory only hits 8G. I'm working with smaller images and a smaller latent size, but memory exceeds 30G (on 40G A100), so I'm wondering if TSP is the differentiating factor here.
Thanks!
Update: The memory issue was mainly due to incorrect use of transformations, but I'm still curious about the TSP question!
Beta Was this translation helpful? Give feedback.
All reactions