Browse Source

Always use fp16 for the text encoders.

pull/2724/head
comfyanonymous 10 months ago
parent
commit
4b0239066d
  1. 4
      comfy/model_management.py

4
comfy/model_management.py

@ -546,10 +546,8 @@ def text_encoder_dtype(device=None):
if is_device_cpu(device):
return torch.float16
if should_use_fp16(device, prioritize_performance=False):
return torch.float16
else:
return torch.float32
def intermediate_device():
if args.gpu_only:

Loading…
Cancel
Save