228 Commits (679bd2845af8e22b2802cf326b99b40a26ba7811)

Author SHA1 Message Date
BlenderNeko d9b1595f85 made sample functions more explicit 2 years ago
BlenderNeko 5818539743 add docstrings 2 years ago
BlenderNeko 2a09e2aa27 refactor/split various bits of code for sampling 2 years ago
comfyanonymous 5282f56434 Implement Linear hypernetworks. 2 years ago
comfyanonymous 6908f9c949 This makes pytorch2.0 attention perform a bit faster. 2 years ago
comfyanonymous 907010e082 Remove some useless code. 2 years ago
comfyanonymous 96b57a9ad6 Don't pass adm to model when it doesn't support it. 2 years ago
comfyanonymous 3696d1699a Add support for GLIGEN textbox model. 2 years ago
comfyanonymous 884ea653c8 Add a way for nodes to set a custom CFG function. 2 years ago
comfyanonymous 73c3e11e83 Fix model_management import so it doesn't get executed twice. 2 years ago
comfyanonymous 81d1f00df3 Some refactoring: from_tokens -> encode_from_tokens 2 years ago
BlenderNeko d0b1b6c6bf fixed improper padding 2 years ago
comfyanonymous deb2b93e79 Move code to empty gpu cache to model_management.py 2 years ago
comfyanonymous 04d9bc13af Safely load pickled embeds that don't load with weights_only=True. 2 years ago
BlenderNeko da115bd78d ensure backwards compat with optional args 2 years ago
BlenderNeko 752f7a162b align behavior with old tokenize function 2 years ago
comfyanonymous 334aab05e5 Don't stop workflow if loading embedding fails. 2 years ago
BlenderNeko 73175cf58c split tokenizer from encoder 2 years ago
BlenderNeko 8489cba140 add unique ID per word/embedding for tokenizer 2 years ago
comfyanonymous 92eca60ec9 Fix for new transformers version. 2 years ago
comfyanonymous 1e1875f674 Print xformers version and warning about 0.0.18 2 years ago
comfyanonymous 7e254d2f69 Clarify what --windows-standalone-build does. 2 years ago
comfyanonymous 44fea05064 Cleanup. 2 years ago
comfyanonymous 58ed0f2da4 Fix loading SD1.5 diffusers checkpoint. 2 years ago
comfyanonymous 64557d6781 Add a --force-fp32 argument to force fp32 for debugging. 2 years ago
comfyanonymous bceccca0e5 Small refactor. 2 years ago
EllangoK 28fff5d1db fixes lack of support for multi configs 2 years ago
comfyanonymous f84f2508cc Rename the cors parameter to something more verbose. 2 years ago
EllangoK 48efae1608 makes cors a cli parameter 2 years ago
EllangoK 01c1fc669f set listen flag to listen on all if specifed 2 years ago
藍+85CD 3e2608e12b Fix auto lowvram detection on CUDA 2 years ago
sALTaccount 60127a8304 diffusers loader 2 years ago
藍+85CD 7cb924f684 Use separate variables instead of `vram_state` 2 years ago
藍+85CD 84b9c0ac2f Import intel_extension_for_pytorch as ipex 2 years ago
EllangoK e5e587b1c0 seperates out arg parser and imports args 2 years ago
藍+85CD 37713e3b0a Add basic XPU device support 2 years ago
comfyanonymous e46b1c3034 Disable xformers in VAE when xformers == 0.0.18 2 years ago
comfyanonymous 1718730e80 Ignore embeddings when sizes don't match and print a WARNING. 2 years ago
comfyanonymous 23524ad8c5 Remove print. 2 years ago
comfyanonymous 539ff487a8 Pull latest tomesd code from upstream. 2 years ago
comfyanonymous f50b1fec69 Add noise augmentation setting to unCLIPConditioning. 2 years ago
comfyanonymous 809bcc8ceb Add support for unCLIP SD2.x models. 2 years ago
comfyanonymous 0d972b85e6 This seems to give better quality in tome. 2 years ago
comfyanonymous 18a6c1db33 Add a TomePatchModel node to the _for_testing section. 2 years ago
comfyanonymous 61ec3c9d5d Add a way to pass options to the transformers blocks. 2 years ago
comfyanonymous afd65d3819 Fix noise mask not working with > 1 batch size on ksamplers. 2 years ago
comfyanonymous b2554bc4dd Split VAE decode batches depending on free memory. 2 years ago
comfyanonymous 0d65cb17b7 Fix ddim_uniform crashing with 37 steps. 2 years ago
Francesco Yoshi Gobbo f55755f0d2
code cleanup 2 years ago
Francesco Yoshi Gobbo cf0098d539
no lowvram state if cpu only 2 years ago