mirror of
https://github.com/comfyanonymous/ComfyUI.git
synced 2025-03-13 22:31:08 +00:00
Make lowvram less aggressive when there are large amounts of free memory.
This commit is contained in:
parent
c24f897352
commit
e2382b6adb
@ -446,7 +446,7 @@ def load_models_gpu(models, memory_required=0, force_patch_weights=False):
|
||||
if lowvram_available and (vram_set_state == VRAMState.LOW_VRAM or vram_set_state == VRAMState.NORMAL_VRAM):
|
||||
model_size = loaded_model.model_memory_required(torch_dev)
|
||||
current_free_mem = get_free_memory(torch_dev)
|
||||
lowvram_model_memory = int(max(64 * (1024 * 1024), (current_free_mem - 1024 * (1024 * 1024)) / 1.3 ))
|
||||
lowvram_model_memory = int(max(64 * (1024 * 1024), (current_free_mem - extra_mem)))
|
||||
if model_size <= (current_free_mem - inference_memory): #only switch to lowvram if really necessary
|
||||
lowvram_model_memory = 0
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user