|
|
@ -120,6 +120,11 @@ total_vram = get_total_memory(get_torch_device()) / (1024 * 1024) |
|
|
|
total_ram = psutil.virtual_memory().total / (1024 * 1024) |
|
|
|
total_ram = psutil.virtual_memory().total / (1024 * 1024) |
|
|
|
logging.info("Total VRAM {:0.0f} MB, total RAM {:0.0f} MB".format(total_vram, total_ram)) |
|
|
|
logging.info("Total VRAM {:0.0f} MB, total RAM {:0.0f} MB".format(total_vram, total_ram)) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
try: |
|
|
|
|
|
|
|
logging.info("pytorch version: {}".format(torch.version.__version__)) |
|
|
|
|
|
|
|
except: |
|
|
|
|
|
|
|
pass |
|
|
|
|
|
|
|
|
|
|
|
try: |
|
|
|
try: |
|
|
|
OOM_EXCEPTION = torch.cuda.OutOfMemoryError |
|
|
|
OOM_EXCEPTION = torch.cuda.OutOfMemoryError |
|
|
|
except: |
|
|
|
except: |
|
|
|