|
@ -10,7 +10,7 @@ def empty_cuda(): |
|
|
while True: |
|
|
while True: |
|
|
gc.collect() |
|
|
gc.collect() |
|
|
torch.cuda.empty_cache() |
|
|
torch.cuda.empty_cache() |
|
|
time.sleep(0.5) |
|
|
time.sleep(0.5) |
|
|
vram = nvidia_smi.get_gpu_stats()["memory_used"] |
|
|
vram = nvidia_smi.get_gpu_stats()["memory_used"] |
|
|
print("vram: %d MB" % vram) |
|
|
print("vram: %d MB" % vram) |
|
|
if vram < 200: |
|
|
if vram < 200: |
|
|