This commit is contained in:
layerdiffusion
2024-08-19 18:54:54 -07:00
parent 8eeeace725
commit 475524496d
2 changed files with 4 additions and 4 deletions

View File

@@ -189,7 +189,7 @@ def calc_cond_uncond_batch(model, cond, uncond, x_in, timestep, model_options):
free_memory = memory_management.get_free_memory(x_in.device)
if not args.disable_gpu_warning:
if (not args.disable_gpu_warning) and x_in.device.type == 'cuda':
free_memory_mb = free_memory / (1024.0 * 1024.0)
safe_memory_mb = 1536.0
if free_memory_mb < safe_memory_mb:
@@ -197,8 +197,8 @@ def calc_cond_uncond_batch(model, cond, uncond, x_in, timestep, model_options):
print(f"[Low GPU VRAM Warning] Your current GPU free memory is {free_memory_mb:.2f} MB for this diffusion iteration.")
print(f"[Low GPU VRAM Warning] This number is lower than the safe value of {safe_memory_mb:.2f} MB.")
print(f"[Low GPU VRAM Warning] If you continue the diffusion process, you may cause NVIDIA GPU degradation, and the speed may be extremely slow (about 10x slower).")
print(f"[Low GPU VRAM Warning] To solve the problem, you can set the 'GPU Weight' (on the top of page) to a lower value.")
print(f"[Low GPU VRAM Warning] If you cannot find 'GPU Weight', you can click the 'all' option in the 'UI' area on the left-top corner of the webpage.")
print(f"[Low GPU VRAM Warning] To solve the problem, you can set the 'GPU Weights' (on the top of page) to a lower value.")
print(f"[Low GPU VRAM Warning] If you cannot find 'GPU Weights', you can click the 'all' option in the 'UI' area on the left-top corner of the webpage.")
print(f"[Low GPU VRAM Warning] If you want to take the risk of NVIDIA GPU fallback and test the 10x slower speed, you can (but are highly not recommended to) add '--disable-gpu-warning' to CMD flags to remove this warning.")
print(f"----------------------\n\n")