From 3a3174ad5aa3e9367f608bb810161e6803818b8e Mon Sep 17 00:00:00 2001 From: bfloat16 <38366253+bfloat16@users.noreply.github.com> Date: Thu, 25 Jan 2024 23:10:27 +0800 Subject: [PATCH] Remove redundant judgments --- webui.py | 10 +++------- 1 file changed, 3 insertions(+), 7 deletions(-) diff --git a/webui.py b/webui.py index 4461056..4b896ff 100644 --- a/webui.py +++ b/webui.py @@ -51,17 +51,13 @@ n_cpu=cpu_count() ngpu = torch.cuda.device_count() gpu_infos = [] mem = [] -if_gpu_ok = False if torch.cuda.is_available() or ngpu != 0: for i in range(ngpu): gpu_name = torch.cuda.get_device_name(i) - if any(value in gpu_name.upper()for value in ["10","16","20","30","40","A2","A3","A4","P4","A50","500","A60","70","80","90","M4","T4","TITAN","L"]): - # A10#A100#V100#A40#P40#M40#K80#A4500 - if_gpu_ok = True # 至少有一张能用的N卡 - gpu_infos.append("%s\t%s" % (i, gpu_name)) - mem.append(int(torch.cuda.get_device_properties(i).total_memory/ 1024/ 1024/ 1024+ 0.4)) -if if_gpu_ok and len(gpu_infos) > 0: + gpu_infos.append("%s\t%s" % (i, gpu_name)) + mem.append(int(torch.cuda.get_device_properties(i).total_memory/ 1024/ 1024/ 1024+ 0.4)) +if len(gpu_infos) > 0: gpu_info = "\n".join(gpu_infos) default_batch_size = min(mem) // 2 else: