Merge 485cb85a552c5bd5b6387569da9b1727332e940e into 11aa78bd9bda8b53047cfcae03abf7ca94d27391

This commit is contained in:
Jacky He 2025-09-18 09:48:39 +08:00 committed by GitHub
commit b9ec6b9ae4
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
2 changed files with 138 additions and 52 deletions

View File

@ -7,79 +7,163 @@ from GPT_SoVITS.inference_webui import change_gpt_weights, change_sovits_weights
i18n = I18nAuto() i18n = I18nAuto()
LANGUAGE_CHOICES = ["中文", "英文", "日文", "韩文", "粤语"]
MIXED_LANGUAGE_CHOICES = ["中英混合", "日英混合", "粤英混合", "韩英混合", "多语种混合"]
def synthesize( SLICE_METHOD_CHOICES = ["凑四句一切", "凑50字一切", "按标点符号切", "按中文句号。切", "按英文句号.切"]
GPT_model_path,
SoVITS_model_path,
ref_audio_path,
ref_text_path,
ref_language,
target_text_path,
target_language,
output_path,
):
# Read reference text
with open(ref_text_path, "r", encoding="utf-8") as file:
ref_text = file.read()
# Read target text def synthesize(args: argparse.Namespace):
with open(target_text_path, "r", encoding="utf-8") as file:
target_text = file.read()
# Change model weights # Change model weights
change_gpt_weights(gpt_path=GPT_model_path) change_gpt_weights(gpt_path=args.gpt_path)
change_sovits_weights(sovits_path=SoVITS_model_path) change_sovits_weights(sovits_path=args.sovits_path)
params = {
"ref_wav_path": args.ref_audio,
"prompt_text": args.ref_text,
"prompt_language": i18n(args.ref_language),
"text": args.target_text,
"text_language": i18n(args.target_language),
}
# region - optional params
if args.slicer: params["how_to_cut"] = i18n(args.slicer)
if args.top_k: params["top_k"] = args.top_k
if args.top_p: params["top_p"] = args.top_p
if args.temperature: params["temperature"] = args.temperature
if args.ref_free: params["ref_free"] = args.ref_free
if args.speed: params["speed"] = args.speed
if args.if_freeze: params["if_freeze"] = args.if_freeze
if args.inp_refs: params["inp_refs"] = args.inp_refs
if args.sample_steps: params["sample_steps"] = args.sample_steps
if args.if_sr: params["if_sr"] = args.if_sr
if args.pause_second: params["pause_second"] = args.pause_second
# endregion - optional params
# Synthesize audio # Synthesize audio
synthesis_result = get_tts_wav( synthesis_result = get_tts_wav(**params)
ref_wav_path=ref_audio_path,
prompt_text=ref_text,
prompt_language=i18n(ref_language),
text=target_text,
text_language=i18n(target_language),
top_p=1,
temperature=1,
)
result_list = list(synthesis_result) result_list = list(synthesis_result)
if result_list: if result_list:
os.makedirs(args.output_path, exist_ok=True) # Create output directory if it doesn't exist
if args.output_path.endswith(".wav"):
output_wav_path = args.output_path
else:
output_wav_path = os.path.join(args.output_path, "output.wav")
last_sampling_rate, last_audio_data = result_list[-1] last_sampling_rate, last_audio_data = result_list[-1]
output_wav_path = os.path.join(output_path, "output.wav")
sf.write(output_wav_path, last_audio_data, last_sampling_rate) sf.write(output_wav_path, last_audio_data, last_sampling_rate)
print(f"Audio saved to {output_wav_path}") print(f"Audio saved to {output_wav_path}")
def main(): def build_parser():
parser = argparse.ArgumentParser(description="GPT-SoVITS Command Line Tool") parser = argparse.ArgumentParser(description="GPT-SoVITS Command Line Tool")
parser.add_argument("--gpt_model", required=True, help="Path to the GPT model file")
parser.add_argument("--sovits_model", required=True, help="Path to the SoVITS model file") # reference settings
parser.add_argument("--ref_audio", required=True, help="Path to the reference audio file") parser.add_argument("--ref_audio", required=True, help="Path to the reference audio file")
parser.add_argument("--ref_text", required=True, help="Path to the reference text file") parser.add_argument("--ref_text", required=True, help="Transcript of the reference audio")
parser.add_argument( parser.add_argument("--ref_language", required=True,
"--ref_language", required=True, choices=["中文", "英文", "日文"], help="Language of the reference audio" choices=LANGUAGE_CHOICES, help="Language of the reference audio")
)
parser.add_argument("--target_text", required=True, help="Path to the target text file") # output settings
parser.add_argument( parser.add_argument("--target_text", required=True, help="Text to be synthesized")
"--target_language", parser.add_argument("--target_language", required=True,
required=True, choices=LANGUAGE_CHOICES+MIXED_LANGUAGE_CHOICES,
choices=["中文", "英文", "日文", "中英混合", "日英混合", "多语种混合"], help="Language of the target text")
help="Language of the target text", parser.add_argument("--slicer", required=False,
) choices=SLICE_METHOD_CHOICES, help="Slicer method")
parser.add_argument("--output_path", required=True, help="Path to the output directory") parser.add_argument("--output_path", required=True, help="Path to the output directory")
# region - inference settings
parser.add_argument("--top_k", required=False, type=int, help="Top-k value")
parser.add_argument("--top_p", required=False, type=float, help="Top-p value")
parser.add_argument("--temperature", required=False, type=float, help="Temperature value")
parser.add_argument("--ref_free", required=False, type=bool, help="Reference free value")
parser.add_argument("--speed", required=False, type=float, help="Speed value")
parser.add_argument("--if_freeze", required=False, type=bool, help="If freeze value")
parser.add_argument("--inp_refs", required=False, type=str, help="Input references")
parser.add_argument("--sample_steps", required=False, type=int, help="Sample steps value")
parser.add_argument("--if_sr", required=False, type=bool, help="If super resolution value")
parser.add_argument("--pause_second", required=False, type=float, help="Pause second value")
# endregion - inference settings
# region - model selection
sub = parser.add_subparsers(dest="mode", required=True)
# Mode 1: provide model paths directly
p_paths = sub.add_parser("paths", help="Use explicit model file paths")
p_paths.add_argument("--gpt_path", required=True, help="Path to the GPT model file")
p_paths.add_argument("--sovits_path", required=True, help="Path to the SoVITS model file")
# Mode 2: select by experiment/version
p_sel = sub.add_parser("select", help="Select models by experiment/version")
p_sel.add_argument("--exp_name", required=True, help="Experiment name")
available_gpt_versions = ["v1", "v2", "v2Pro", "v2ProPlus", "v3", "v4"]
p_sel.add_argument("--gpt_version", required=True, choices=available_gpt_versions, help="Version of the GPT model")
available_sovits_versions = ["v1", "v2", "v2Pro", "v2ProPlus", "v3", "v4"]
p_sel.add_argument("--sovits_version", required=True, choices=available_sovits_versions, help="Version of the SoVITS model")
p_sel.add_argument("--gpt_epoch", type=int, help="Epoch of the GPT model")
p_sel.add_argument("--sovits_epoch", type=int, help="Epoch of the SoVITS model")
# endregion - model selection
return parser
def get_model_path(args)->argparse.Namespace:
"""
Get the model path from exp_name, version and epoch
Args:
args: argparse.Namespace
Returns:
args: argparse.Namespace
"""
exist_gpt_path = []
exist_sovits_path = []
def _get_model_dir(model_type, version):
if version == "v1":
return f"{model_type}_weights"
else:
return f"{model_type}_weights_{version}"
# get all the model paths with the same exp_name
for files in os.listdir(_get_model_dir("GPT", args.gpt_version)):
if args.exp_name in files and files.endswith(".ckpt"):
exist_gpt_path.append(os.path.join(_get_model_dir("GPT", args.gpt_version), files))
for files in os.listdir(_get_model_dir("SoVITS", args.sovits_version)):
if args.exp_name in files and files.endswith(".pth"):
exist_sovits_path.append(os.path.join(_get_model_dir("SoVITS", args.sovits_version), files))
# get the largest epoch if not specified
if args.gpt_epoch:
args.gpt_path = [i for i in exist_gpt_path if f"e{args.gpt_epoch}" in i]
else:
args.gpt_path = sorted(exist_gpt_path)[-1]
if args.sovits_epoch:
args.sovits_path = [i for i in exist_sovits_path if f"e{args.sovits_epoch}" in i]
else:
args.sovits_path = sorted(exist_sovits_path)[-1]
if not args.gpt_path or not args.sovits_path:
raise ValueError("No model found")
return args
def main():
parser = build_parser()
args = parser.parse_args() args = parser.parse_args()
synthesize( print(args)
args.gpt_model, if args.mode == "select":
args.sovits_model, args = get_model_path(args)
args.ref_audio,
args.ref_text, args.target_text = args.target_text.replace("'", "").replace('"', "")
args.ref_language,
args.target_text,
args.target_language, synthesize(args)
args.output_path,
)
if __name__ == "__main__": if __name__ == "__main__":

View File

@ -180,6 +180,8 @@ def get_bert_feature(text, word2ph):
for i in range(len(word2ph)): for i in range(len(word2ph)):
repeat_feature = res[i].repeat(word2ph[i], 1) repeat_feature = res[i].repeat(word2ph[i], 1)
phone_level_feature.append(repeat_feature) phone_level_feature.append(repeat_feature)
if len(phone_level_feature) == 0:
return torch.empty((res.shape[1], 0), dtype=res.dtype, device=res.device)
phone_level_feature = torch.cat(phone_level_feature, dim=0) phone_level_feature = torch.cat(phone_level_feature, dim=0)
return phone_level_feature.T return phone_level_feature.T