Hello! I tried using the sota-16z model to reconstruct a 1024x1024x24 frame video with chunk=8, but it requires 94GB of GPU memory. Does the model support multi-GPU inference or tiled-vae inference to reduce memory usage?
Additionally, I tried using chunk=4 for reconstruction, but the results have some artifacts.