765 Commits (1a0486bb96fb1ff10f4ea3c0d62eb815e9630585)

Author SHA1 Message Date
comfyanonymous 1e6b67101c Support diffusers format t2i adapters. 1 year ago
comfyanonymous 326577d04c Allow cancelling of everything with a progress bar. 1 year ago
comfyanonymous f88f7f413a Add a ConditioningSetAreaPercentage node. 1 year ago
comfyanonymous 1938f5c5fe Add a force argument to soft_empty_cache to force a cache empty. 1 year ago
Simon Lui 2da73b7073 Revert changes in comfy/ldm/modules/diffusionmodules/util.py, which is unused. 1 year ago
comfyanonymous a74c5dbf37 Move some functions to utils.py 1 year ago
Simon Lui 4a0c4ce4ef Some fixes to generalize CUDA specific functionality to Intel or other GPUs. 1 year ago
comfyanonymous 77a176f9e0 Use common function to reshape batch to. 1 year ago
comfyanonymous 7931ff0fd9 Support SDXL inpaint models. 1 year ago
comfyanonymous 0e3b641172 Remove xformers related print. 1 year ago
comfyanonymous 5c363a9d86 Fix controlnet bug. 1 year ago
comfyanonymous cfe1c54de8 Fix controlnet issue. 1 year ago
comfyanonymous 1c012d69af It doesn't make sense for c_crossattn and c_concat to be lists. 1 year ago
comfyanonymous 7e941f9f24 Clean up DiffusersLoader node. 1 year ago
Simon Lui 18617967e5
Fix error message in model_patcher.py 1 year ago
comfyanonymous fe4c07400c Fix "Load Checkpoint with config" node. 1 year ago
comfyanonymous f2f5e5dcbb Support SDXL t2i adapters with 3 channel input. 1 year ago
comfyanonymous 15adc3699f Move beta_schedule to model_config and allow disabling unet creation. 1 year ago
comfyanonymous bed116a1f9 Remove optimization that caused border. 1 year ago
comfyanonymous 65cae62c71 No need to check filename extensions to detect shuffle controlnet. 1 year ago
comfyanonymous 4e89b2c25a Put clip vision outputs on the CPU. 1 year ago
comfyanonymous a094b45c93 Load clipvision model to GPU for faster performance. 1 year ago
comfyanonymous 1300a1bb4c Text encoder should initially load on the offload_device not the regular. 1 year ago
comfyanonymous f92074b84f Move ModelPatcher to model_patcher.py 1 year ago
comfyanonymous 4798cf5a62 Implement loras with norm keys. 1 year ago
comfyanonymous b8c7c770d3 Enable bf16-vae by default on ampere and up. 1 year ago
comfyanonymous 1c794a2161 Fallback to slice attention if xformers doesn't support the operation. 1 year ago
comfyanonymous d935ba50c4 Make --bf16-vae work on torch 2.0 1 year ago
comfyanonymous a57b0c797b Fix lowvram model merging. 1 year ago
comfyanonymous f72780a7e3 The new smart memory management makes this unnecessary. 1 year ago
comfyanonymous c77f02e1c6 Move controlnet code to comfy/controlnet.py 1 year ago
comfyanonymous 15a7716fa6 Move lora code to comfy/lora.py 1 year ago
comfyanonymous ec96f6d03a Move text_projection to base clip model. 1 year ago
comfyanonymous 30eb92c3cb Code cleanups. 1 year ago
comfyanonymous 51dde87e97 Try to free enough vram for control lora inference. 1 year ago
comfyanonymous e3d0a9a490 Fix potential issue with text projection matrix multiplication. 1 year ago
comfyanonymous cc44ade79e Always shift text encoder to GPU when the device supports fp16. 1 year ago
comfyanonymous a6ef08a46a Even with forced fp16 the cpu device should never use it. 1 year ago
comfyanonymous 00c0b2c507 Initialize text encoder to target dtype. 1 year ago
comfyanonymous f081017c1a Save memory by storing text encoder weights in fp16 in most situations. 1 year ago
comfyanonymous afcb9cb1df All resolutions now work with t2i adapter for SDXL. 1 year ago
comfyanonymous 85fde89d7f T2I adapter SDXL. 1 year ago
comfyanonymous cf5ae46928 Controlnet/t2iadapter cleanup. 1 year ago
comfyanonymous 763b0cf024 Fix control lora not working in fp32. 1 year ago
comfyanonymous 199d73364a Fix ControlLora on lowvram. 1 year ago
comfyanonymous d08e53de2e Remove autocast from controlnet code. 1 year ago
comfyanonymous 0d7b0a4dc7 Small cleanups. 1 year ago
Simon Lui 9225465975 Further tuning and fix mem_free_total. 1 year ago
Simon Lui 2c096e4260 Add ipex optimize and other enhancements for Intel GPUs based on recent memory changes. 1 year ago
comfyanonymous e9469e732d --disable-smart-memory now disables loading model directly to vram. 1 year ago