William a3b9b3c1c3 Memory check before inference to avoid VAE Decode using exceeded VRAM.
Check if free memory is not less than expected before doing actual decoding,
and if it fails, switch to tiled VAE decoding directly.

It seems PyTorch may continue occupying memory until the model is destroyed
after OOM occurs. This commit tries to avoid OOM from happening in the first
place for VAE Decode.

This is for VAE Decode ran with exceeded VRAM from #5737.
2024-11-24 18:47:01 +08:00
..
2023-01-16 22:37:14 -05:00
2024-06-27 18:43:11 -04:00
2024-11-21 08:38:23 -05:00
2024-06-10 16:00:03 -04:00
2024-02-18 02:20:23 -05:00
2024-11-22 10:49:15 -05:00
2024-10-22 16:34:50 -04:00
2024-10-20 22:27:00 -04:00
2024-11-13 07:18:30 -05:00