mirror of
https://github.com/comfyanonymous/ComfyUI.git
synced 2025-04-18 01:53:31 +00:00
fix the memory leakage issue
This commit is contained in:
parent
8558803f44
commit
bc9eb9dfdb
@ -499,12 +499,19 @@ class ModelPatcher:
|
|||||||
if len(k) > 2:
|
if len(k) > 2:
|
||||||
function = k[2]
|
function = k[2]
|
||||||
org_key=key.replace("diffusion_model", "diffusion_model._orig_mod")
|
org_key=key.replace("diffusion_model", "diffusion_model._orig_mod")
|
||||||
if key in model_sd or org_key in model_sd:
|
if key in model_sd:
|
||||||
p.add(k)
|
p.add(k)
|
||||||
current_patches = self.patches.get(key, [])
|
current_patches = self.patches.get(key, [])
|
||||||
current_patches.append((strength_patch, patches[k], strength_model, offset, function))
|
current_patches.append((strength_patch, patches[k], strength_model, offset, function))
|
||||||
self.patches[key] = current_patches
|
self.patches[key] = current_patches
|
||||||
self.patches[org_key] = current_patches
|
self.patches[org_key] = current_patches
|
||||||
|
elif org_key in model_sd:
|
||||||
|
if key in self.patches:
|
||||||
|
self.patches.pop(key)
|
||||||
|
p.add(k)
|
||||||
|
current_patches = self.patches.get(org_key, [])
|
||||||
|
current_patches.append((strength_patch, patches[k], strength_model, offset, function))
|
||||||
|
self.patches[org_key] = current_patches
|
||||||
|
|
||||||
self.patches_uuid = uuid.uuid4()
|
self.patches_uuid = uuid.uuid4()
|
||||||
return list(p)
|
return list(p)
|
||||||
|
@ -19,7 +19,7 @@ class TorchCompileModel:
|
|||||||
"backend": (["inductor", "cudagraphs", "openvino"],),
|
"backend": (["inductor", "cudagraphs", "openvino"],),
|
||||||
},
|
},
|
||||||
"optional": {
|
"optional": {
|
||||||
"openvino_device": (available_devices,),
|
"openvino device": (available_devices,),
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -30,6 +30,7 @@ class TorchCompileModel:
|
|||||||
EXPERIMENTAL = True
|
EXPERIMENTAL = True
|
||||||
|
|
||||||
def patch(self, model, backend, openvino_device):
|
def patch(self, model, backend, openvino_device):
|
||||||
|
print(model.__class__.__name__)
|
||||||
if backend == "openvino":
|
if backend == "openvino":
|
||||||
options = {"device": openvino_device}
|
options = {"device": openvino_device}
|
||||||
try:
|
try:
|
||||||
@ -39,6 +40,12 @@ class TorchCompileModel:
|
|||||||
"Could not import openvino python package. "
|
"Could not import openvino python package. "
|
||||||
"Please install it with `pip install openvino`."
|
"Please install it with `pip install openvino`."
|
||||||
)
|
)
|
||||||
|
import openvino.frontend.pytorch.torchdynamo.execute as ov_ex
|
||||||
|
|
||||||
|
torch._dynamo.reset()
|
||||||
|
ov_ex.compiled_cache.clear()
|
||||||
|
ov_ex.req_cache.clear()
|
||||||
|
ov_ex.partitioned_modules.clear()
|
||||||
else:
|
else:
|
||||||
options = None
|
options = None
|
||||||
m = model.clone()
|
m = model.clone()
|
||||||
|
Loading…
Reference in New Issue
Block a user