Commit Graph

545 Commits

Author SHA1 Message Date
comfyanonymous
f10b8948c3 768-v support for uni_pc sampler. 2023-02-11 04:34:58 -05:00
comfyanonymous
ce0aeb109e Remove print. 2023-02-11 03:41:40 -05:00
comfyanonymous
5489d5af04 Add uni_pc sampler to KSampler* nodes. 2023-02-11 03:34:09 -05:00
comfyanonymous
1a4edd19cd Fix overflow issue with inplace softmax. 2023-02-10 11:47:41 -05:00
comfyanonymous
509c7dfc6d Use real softmax in split op to fix issue with some images. 2023-02-10 03:13:49 -05:00
comfyanonymous
7e1e193f39 Automatically enable lowvram mode if vram is less than 4GB.
Use: --normalvram to disable it.
2023-02-10 00:47:56 -05:00
comfyanonymous
e9d3ac2ba0 Typo. 2023-02-09 16:31:54 -05:00
comfyanonymous
9ae4e1b0ff Add a list of features to the Readme. 2023-02-09 16:27:52 -05:00
comfyanonymous
324273fff2 Fix embedding not working when on new line. 2023-02-09 14:12:02 -05:00
comfyanonymous
1f6a467e92 Update ldm dir with latest upstream stable diffusion changes. 2023-02-09 13:47:36 -05:00
BazettFraga
642516a3a6 create output dir if none is present 2023-02-09 12:49:31 -05:00
comfyanonymous
773cdabfce Same thing but for the other places where it's used. 2023-02-09 12:43:29 -05:00
comfyanonymous
df40d4f3bf torch.cuda.OutOfMemoryError is not present on older pytorch versions. 2023-02-09 12:33:27 -05:00
comfyanonymous
1d9ec62cfb Use absolute output directory path. 2023-02-09 09:59:43 -05:00
comfyanonymous
05d571fe7f Merge branch 'master' of https://github.com/bazettfraga/ComfyUI into merge_pr2 2023-02-09 00:44:38 -05:00
comfyanonymous
e8c499ddd4 Split optimization for VAE attention block. 2023-02-08 22:04:20 -05:00
comfyanonymous
5b4e312749 Use inplace operations for less OOM issues. 2023-02-08 22:04:13 -05:00
BazettFraga
e58887dfa7 forgot windows does double backslashes for paths due to its use as escape char. 2023-02-09 01:30:06 +01:00
BazettFraga
81082045c2 add recursive_search, swap relevant os.listdirs 2023-02-09 01:22:33 +01:00
comfyanonymous
3fd87cbd21 Slightly smarter batching behaviour.
Try to keep batch sizes more consistent which seems to improve things on
AMD GPUs.
2023-02-08 17:28:43 -05:00
comfyanonymous
bbdcf0b737 Use relative imports for k_diffusion. 2023-02-08 16:51:19 -05:00
comfyanonymous
3e22815a9a Fix k_diffusion not getting imported from the folder. 2023-02-08 16:29:22 -05:00
comfyanonymous
708138c77d Remove print. 2023-02-08 14:51:18 -05:00
comfyanonymous
047775615b Lower the chances of an OOM. 2023-02-08 14:24:27 -05:00
comfyanonymous
853e96ada3 Increase it/s by batching together some stuff sent to unet. 2023-02-08 14:24:00 -05:00
comfyanonymous
c92633eaa2 Auto calculate amount of memory to use for --lowvram 2023-02-08 11:42:37 -05:00
comfyanonymous
534736b924 Add some low vram modes: --lowvram and --novram 2023-02-08 11:37:10 -05:00
comfyanonymous
a84cd0d1ad Don't unload/reload model from CPU uselessly. 2023-02-08 03:40:43 -05:00
comfyanonymous
e3e65947f2 Add a --help to main.py 2023-02-07 22:13:42 -05:00
comfyanonymous
1f18221e17 Add --port to set custom port. 2023-02-07 21:57:17 -05:00
comfyanonymous
6e40393b6b Fix delete sometimes not properly refreshing queue state. 2023-02-07 00:07:31 -05:00
comfyanonymous
d71d0c88e5 Add some simple queue management to the GUI. 2023-02-06 23:40:38 -05:00
comfyanonymous
b1a7c9ebf6 Embeddings/textual inversion support for SD2.x 2023-02-05 15:49:03 -05:00
comfyanonymous
1de5aa6a59 Add a CLIPLoader node to load standalone clip weights.
Put them in models/clip
2023-02-05 15:20:18 -05:00
comfyanonymous
56d802e1f3 Use transformers CLIP instead of open_clip for SD2.x
This should make things a bit cleaner.
2023-02-05 14:36:28 -05:00
comfyanonymous
bf9ccffb17 Small fix for SD2.x loras. 2023-02-05 11:38:25 -05:00
comfyanonymous
678105fade SD2.x CLIP support for Loras. 2023-02-05 01:54:09 -05:00
comfyanonymous
3f3d77a324 Fix image node always executing instead of only when the image changed. 2023-02-04 16:08:29 -05:00
comfyanonymous
4225d1cb9f Add a basic ImageScale node.
It's pretty much the same as the LatentUpscale node for now but for images
in pixel space.
2023-02-04 16:01:01 -05:00
comfyanonymous
bff0e11941 Add a LatentCrop node. 2023-02-04 15:21:46 -05:00
comfyanonymous
43c795f462 Add a --listen argument to listen on 0.0.0.0 2023-02-04 12:01:53 -05:00
comfyanonymous
41a7532c15 A bit bigger. 2023-02-03 13:56:00 -05:00
comfyanonymous
7bc3f91bd6 Add some instructions how to use the venv from another SD install. 2023-02-03 13:54:45 -05:00
comfyanonymous
149a4de3f2 Fix potential issue if exception happens when patching model. 2023-02-03 03:55:50 -05:00
comfyanonymous
ef90e9c376 Add a LoraLoader node to apply loras to models and clip.
The models are modified in place before being used and unpatched after.
I think this is better than monkeypatching since it might make it easier
to use faster non pytorch unet inference in the future.
2023-02-03 02:46:24 -05:00
comfyanonymous
96664f5d5e Web interface bug fix for multiple inputs from the same node. 2023-02-03 00:39:28 -05:00
comfyanonymous
1d84a44b08 Fix some small annoyances with the UI. 2023-02-02 14:36:11 -05:00
comfyanonymous
e65a20e62a Add a button to queue prompts to the front of the queue. 2023-02-01 22:34:59 -05:00
comfyanonymous
4b08314257 Add more features to the backend queue code.
The queue can now be queried, entries can be deleted and prompts easily
queued to the front of the queue.

Just need to expose it in the UI next.
2023-02-01 22:33:10 -05:00
comfyanonymous
9d611a90e8 Small web interface fixes. 2023-01-31 03:37:34 -05:00