From 00cbadcaaddd891505737a78c98e1368439e242e Mon Sep 17 00:00:00 2001 From: RVC-Boss <129054828+RVC-Boss@users.noreply.github.com> Date: Thu, 27 Feb 2025 14:38:55 +0800 Subject: [PATCH] =?UTF-8?q?=E5=B1=8F=E8=94=BDlora=20strict=3DFalse?= =?UTF-8?q?=E7=9A=84print=E5=81=87=E5=91=8A=E8=AD=A6=EF=BC=8C=E5=85=B6?= =?UTF-8?q?=E5=AE=9E=E5=8A=A0=E8=BD=BD=E9=80=BB=E8=BE=91=E6=B2=A1=E9=97=AE?= =?UTF-8?q?=E9=A2=98=E3=80=82?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit 屏蔽lora strict=False的print假告警,其实加载逻辑没问题。 --- GPT_SoVITS/inference_webui.py | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/GPT_SoVITS/inference_webui.py b/GPT_SoVITS/inference_webui.py index 9595835..74e0cab 100644 --- a/GPT_SoVITS/inference_webui.py +++ b/GPT_SoVITS/inference_webui.py @@ -288,7 +288,8 @@ def change_sovits_weights(sovits_path,prompt_language=None,text_language=None): init_lora_weights=True, ) vq_model.cfm = get_peft_model(vq_model.cfm, lora_config) - print("loading sovits_v3_lora%s"%(lora_rank),vq_model.load_state_dict(dict_s2["weight"], strict=False)) + print("loading sovits_v3_lora%s"%(lora_rank)) + vq_model.load_state_dict(dict_s2["weight"], strict=False)## vq_model.cfm = vq_model.cfm.merge_and_unload() # torch.save(vq_model.state_dict(),"merge_win.pth") vq_model.eval()