From f72780a7e3634de1400bc3dd13207c463884dcb9 Mon Sep 17 00:00:00 2001 From: comfyanonymous Date: Fri, 25 Aug 2023 18:02:15 -0400 Subject: [PATCH] The new smart memory management makes this unnecessary. --- comfy/model_management.py | 3 --- 1 file changed, 3 deletions(-) diff --git a/comfy/model_management.py b/comfy/model_management.py index 0e86df41..01643449 100644 --- a/comfy/model_management.py +++ b/comfy/model_management.py @@ -111,9 +111,6 @@ if not args.normalvram and not args.cpu: if lowvram_available and total_vram <= 4096: print("Trying to enable lowvram mode because your GPU seems to have 4GB or less. If you don't want this use: --normalvram") set_vram_to = VRAMState.LOW_VRAM - elif total_vram > total_ram * 1.1 and total_vram > 14336: - print("Enabling highvram mode because your GPU has more vram than your computer has ram. If you don't want this use: --normalvram") - vram_state = VRAMState.HIGH_VRAM try: OOM_EXCEPTION = torch.cuda.OutOfMemoryError