mirror of
https://github.com/RVC-Boss/GPT-SoVITS.git
synced 2025-10-08 16:00:01 +08:00
添加with torch.no_grad(),速度快一大截
This commit is contained in:
parent
ec7647e08d
commit
adb7f71b64
4
.gitignore
vendored
4
.gitignore
vendored
@ -10,6 +10,8 @@ reference
|
|||||||
GPT_weights
|
GPT_weights
|
||||||
SoVITS_weights
|
SoVITS_weights
|
||||||
TEMP
|
TEMP
|
||||||
|
PortableGit
|
||||||
ffmpeg.exe
|
ffmpeg.exe
|
||||||
ffprobe.exe
|
ffprobe.exe
|
||||||
|
tmp_audio
|
||||||
|
trained
|
||||||
|
@ -249,8 +249,6 @@ class TTS:
|
|||||||
if self.configs.is_half and str(self.configs.device)!="cpu":
|
if self.configs.is_half and str(self.configs.device)!="cpu":
|
||||||
self.bert_model = self.bert_model.half()
|
self.bert_model = self.bert_model.half()
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
def init_vits_weights(self, weights_path: str):
|
def init_vits_weights(self, weights_path: str):
|
||||||
print(f"Loading VITS weights from {weights_path}")
|
print(f"Loading VITS weights from {weights_path}")
|
||||||
self.configs.vits_weights_path = weights_path
|
self.configs.vits_weights_path = weights_path
|
||||||
@ -282,7 +280,6 @@ class TTS:
|
|||||||
if self.configs.is_half and str(self.configs.device)!="cpu":
|
if self.configs.is_half and str(self.configs.device)!="cpu":
|
||||||
self.vits_model = self.vits_model.half()
|
self.vits_model = self.vits_model.half()
|
||||||
|
|
||||||
|
|
||||||
def init_t2s_weights(self, weights_path: str):
|
def init_t2s_weights(self, weights_path: str):
|
||||||
print(f"Loading Text2Semantic weights from {weights_path}")
|
print(f"Loading Text2Semantic weights from {weights_path}")
|
||||||
self.configs.t2s_weights_path = weights_path
|
self.configs.t2s_weights_path = weights_path
|
||||||
@ -379,7 +376,6 @@ class TTS:
|
|||||||
# self.refer_spec = spec
|
# self.refer_spec = spec
|
||||||
self.prompt_cache["refer_spec"] = spec
|
self.prompt_cache["refer_spec"] = spec
|
||||||
|
|
||||||
|
|
||||||
def _set_prompt_semantic(self, ref_wav_path:str):
|
def _set_prompt_semantic(self, ref_wav_path:str):
|
||||||
zero_wav = np.zeros(
|
zero_wav = np.zeros(
|
||||||
int(self.configs.sampling_rate * 0.3),
|
int(self.configs.sampling_rate * 0.3),
|
||||||
@ -420,7 +416,8 @@ class TTS:
|
|||||||
max_length = max(seq_lengths)
|
max_length = max(seq_lengths)
|
||||||
else:
|
else:
|
||||||
max_length = max(seq_lengths) if max_length < max(seq_lengths) else max_length
|
max_length = max(seq_lengths) if max_length < max(seq_lengths) else max_length
|
||||||
|
# 我爱套 torch.no_grad()
|
||||||
|
# with torch.no_grad():
|
||||||
padded_sequences = []
|
padded_sequences = []
|
||||||
for seq, length in zip(sequences, seq_lengths):
|
for seq, length in zip(sequences, seq_lengths):
|
||||||
padding = [0] * axis + [0, max_length - length] + [0] * (ndim - axis - 1)
|
padding = [0] * axis + [0, max_length - length] + [0] * (ndim - axis - 1)
|
||||||
@ -438,111 +435,113 @@ class TTS:
|
|||||||
precision:torch.dtype=torch.float32,
|
precision:torch.dtype=torch.float32,
|
||||||
):
|
):
|
||||||
|
|
||||||
_data:list = []
|
# 但是这里不能套,反而会负优化
|
||||||
index_and_len_list = []
|
# with torch.no_grad():
|
||||||
for idx, item in enumerate(data):
|
_data:list = []
|
||||||
norm_text_len = len(item["norm_text"])
|
index_and_len_list = []
|
||||||
index_and_len_list.append([idx, norm_text_len])
|
for idx, item in enumerate(data):
|
||||||
|
norm_text_len = len(item["norm_text"])
|
||||||
|
index_and_len_list.append([idx, norm_text_len])
|
||||||
|
|
||||||
batch_index_list = []
|
batch_index_list = []
|
||||||
if split_bucket:
|
if split_bucket:
|
||||||
index_and_len_list.sort(key=lambda x: x[1])
|
index_and_len_list.sort(key=lambda x: x[1])
|
||||||
index_and_len_list = np.array(index_and_len_list, dtype=np.int64)
|
index_and_len_list = np.array(index_and_len_list, dtype=np.int64)
|
||||||
|
|
||||||
batch_index_list_len = 0
|
batch_index_list_len = 0
|
||||||
pos = 0
|
pos = 0
|
||||||
while pos <index_and_len_list.shape[0]:
|
while pos <index_and_len_list.shape[0]:
|
||||||
# batch_index_list.append(index_and_len_list[pos:min(pos+batch_size,len(index_and_len_list))])
|
# batch_index_list.append(index_and_len_list[pos:min(pos+batch_size,len(index_and_len_list))])
|
||||||
pos_end = min(pos+batch_size,index_and_len_list.shape[0])
|
pos_end = min(pos+batch_size,index_and_len_list.shape[0])
|
||||||
while pos < pos_end:
|
while pos < pos_end:
|
||||||
batch=index_and_len_list[pos:pos_end, 1].astype(np.float32)
|
batch=index_and_len_list[pos:pos_end, 1].astype(np.float32)
|
||||||
score=batch[(pos_end-pos)//2]/(batch.mean()+1e-8)
|
score=batch[(pos_end-pos)//2]/(batch.mean()+1e-8)
|
||||||
if (score>=threshold) or (pos_end-pos==1):
|
if (score>=threshold) or (pos_end-pos==1):
|
||||||
batch_index=index_and_len_list[pos:pos_end, 0].tolist()
|
batch_index=index_and_len_list[pos:pos_end, 0].tolist()
|
||||||
batch_index_list_len += len(batch_index)
|
batch_index_list_len += len(batch_index)
|
||||||
batch_index_list.append(batch_index)
|
batch_index_list.append(batch_index)
|
||||||
pos = pos_end
|
pos = pos_end
|
||||||
break
|
break
|
||||||
pos_end=pos_end-1
|
pos_end=pos_end-1
|
||||||
|
|
||||||
assert batch_index_list_len == len(data)
|
assert batch_index_list_len == len(data)
|
||||||
|
|
||||||
else:
|
else:
|
||||||
for i in range(len(data)):
|
for i in range(len(data)):
|
||||||
if i%batch_size == 0:
|
if i%batch_size == 0:
|
||||||
batch_index_list.append([])
|
batch_index_list.append([])
|
||||||
batch_index_list[-1].append(i)
|
batch_index_list[-1].append(i)
|
||||||
|
|
||||||
|
for batch_idx, index_list in enumerate(batch_index_list):
|
||||||
for batch_idx, index_list in enumerate(batch_index_list):
|
item_list = [data[idx] for idx in index_list]
|
||||||
item_list = [data[idx] for idx in index_list]
|
phones_list = []
|
||||||
phones_list = []
|
phones_len_list = []
|
||||||
phones_len_list = []
|
# bert_features_list = []
|
||||||
# bert_features_list = []
|
all_phones_list = []
|
||||||
all_phones_list = []
|
all_phones_len_list = []
|
||||||
all_phones_len_list = []
|
all_bert_features_list = []
|
||||||
all_bert_features_list = []
|
norm_text_batch = []
|
||||||
norm_text_batch = []
|
bert_max_len = 0
|
||||||
bert_max_len = 0
|
phones_max_len = 0
|
||||||
phones_max_len = 0
|
# 但是这里也不能套,反而会负优化
|
||||||
for item in item_list:
|
# with torch.no_grad():
|
||||||
if prompt_data is not None:
|
for item in item_list:
|
||||||
all_bert_features = torch.cat([prompt_data["bert_features"], item["bert_features"]], 1)\
|
if prompt_data is not None:
|
||||||
|
all_bert_features = torch.cat([prompt_data["bert_features"], item["bert_features"]], 1)\
|
||||||
|
.to(dtype=precision, device=device)
|
||||||
|
all_phones = torch.LongTensor(prompt_data["phones"]+item["phones"]).to(device)
|
||||||
|
phones = torch.LongTensor(item["phones"]).to(device)
|
||||||
|
# norm_text = prompt_data["norm_text"]+item["norm_text"]
|
||||||
|
else:
|
||||||
|
all_bert_features = item["bert_features"]\
|
||||||
.to(dtype=precision, device=device)
|
.to(dtype=precision, device=device)
|
||||||
all_phones = torch.LongTensor(prompt_data["phones"]+item["phones"]).to(device)
|
phones = torch.LongTensor(item["phones"]).to(device)
|
||||||
phones = torch.LongTensor(item["phones"]).to(device)
|
all_phones = phones
|
||||||
# norm_text = prompt_data["norm_text"]+item["norm_text"]
|
# norm_text = item["norm_text"]
|
||||||
else:
|
|
||||||
all_bert_features = item["bert_features"]\
|
|
||||||
.to(dtype=precision, device=device)
|
|
||||||
phones = torch.LongTensor(item["phones"]).to(device)
|
|
||||||
all_phones = phones
|
|
||||||
# norm_text = item["norm_text"]
|
|
||||||
|
|
||||||
bert_max_len = max(bert_max_len, all_bert_features.shape[-1])
|
bert_max_len = max(bert_max_len, all_bert_features.shape[-1])
|
||||||
phones_max_len = max(phones_max_len, phones.shape[-1])
|
phones_max_len = max(phones_max_len, phones.shape[-1])
|
||||||
|
|
||||||
phones_list.append(phones)
|
phones_list.append(phones)
|
||||||
phones_len_list.append(phones.shape[-1])
|
phones_len_list.append(phones.shape[-1])
|
||||||
all_phones_list.append(all_phones)
|
all_phones_list.append(all_phones)
|
||||||
all_phones_len_list.append(all_phones.shape[-1])
|
all_phones_len_list.append(all_phones.shape[-1])
|
||||||
all_bert_features_list.append(all_bert_features)
|
all_bert_features_list.append(all_bert_features)
|
||||||
norm_text_batch.append(item["norm_text"])
|
norm_text_batch.append(item["norm_text"])
|
||||||
|
|
||||||
phones_batch = phones_list
|
phones_batch = phones_list
|
||||||
all_phones_batch = all_phones_list
|
all_phones_batch = all_phones_list
|
||||||
all_bert_features_batch = all_bert_features_list
|
all_bert_features_batch = all_bert_features_list
|
||||||
|
|
||||||
|
# max_len = max(bert_max_len, phones_max_len)
|
||||||
|
# phones_batch = self.batch_sequences(phones_list, axis=0, pad_value=0, max_length=max_len)
|
||||||
|
#### 直接对phones和bert_features进行pad,会增大复读概率。
|
||||||
|
# all_phones_batch = self.batch_sequences(all_phones_list, axis=0, pad_value=0, max_length=max_len)
|
||||||
|
# all_bert_features_batch = all_bert_features_list
|
||||||
|
# all_bert_features_batch = torch.zeros(len(item_list), 1024, max_len, dtype=precision, device=device)
|
||||||
|
# for idx, item in enumerate(all_bert_features_list):
|
||||||
|
# all_bert_features_batch[idx, :, : item.shape[-1]] = item
|
||||||
|
|
||||||
# max_len = max(bert_max_len, phones_max_len)
|
# #### 先对phones进行embedding、对bert_features进行project,再pad到相同长度,以缓解复读问题。(可能还有其他因素导致复读)
|
||||||
# phones_batch = self.batch_sequences(phones_list, axis=0, pad_value=0, max_length=max_len)
|
# all_phones_list = [self.t2s_model.model.ar_text_embedding(item.to(self.t2s_model.device)) for item in all_phones_list]
|
||||||
#### 直接对phones和bert_features进行pad,会增大复读概率。
|
# all_phones_list = [F.pad(item,(0,0,0,max_len-item.shape[0]),value=0) for item in all_phones_list]
|
||||||
# all_phones_batch = self.batch_sequences(all_phones_list, axis=0, pad_value=0, max_length=max_len)
|
# all_phones_batch = torch.stack(all_phones_list, dim=0)
|
||||||
# all_bert_features_batch = all_bert_features_list
|
|
||||||
# all_bert_features_batch = torch.zeros(len(item_list), 1024, max_len, dtype=precision, device=device)
|
|
||||||
# for idx, item in enumerate(all_bert_features_list):
|
|
||||||
# all_bert_features_batch[idx, :, : item.shape[-1]] = item
|
|
||||||
|
|
||||||
# #### 先对phones进行embedding、对bert_features进行project,再pad到相同长度,以缓解复读问题。(可能还有其他因素导致复读)
|
# all_bert_features_list = [self.t2s_model.model.bert_proj(item.to(self.t2s_model.device).transpose(0, 1)) for item in all_bert_features_list]
|
||||||
# all_phones_list = [self.t2s_model.model.ar_text_embedding(item.to(self.t2s_model.device)) for item in all_phones_list]
|
# all_bert_features_list = [F.pad(item,(0,0,0,max_len-item.shape[0]), value=0) for item in all_bert_features_list]
|
||||||
# all_phones_list = [F.pad(item,(0,0,0,max_len-item.shape[0]),value=0) for item in all_phones_list]
|
# all_bert_features_batch = torch.stack(all_bert_features_list, dim=0)
|
||||||
# all_phones_batch = torch.stack(all_phones_list, dim=0)
|
|
||||||
|
|
||||||
# all_bert_features_list = [self.t2s_model.model.bert_proj(item.to(self.t2s_model.device).transpose(0, 1)) for item in all_bert_features_list]
|
batch = {
|
||||||
# all_bert_features_list = [F.pad(item,(0,0,0,max_len-item.shape[0]), value=0) for item in all_bert_features_list]
|
"phones": phones_batch,
|
||||||
# all_bert_features_batch = torch.stack(all_bert_features_list, dim=0)
|
"phones_len": torch.LongTensor(phones_len_list).to(device),
|
||||||
|
"all_phones": all_phones_batch,
|
||||||
|
"all_phones_len": torch.LongTensor(all_phones_len_list).to(device),
|
||||||
|
"all_bert_features": all_bert_features_batch,
|
||||||
|
"norm_text": norm_text_batch
|
||||||
|
}
|
||||||
|
_data.append(batch)
|
||||||
|
|
||||||
batch = {
|
return _data, batch_index_list
|
||||||
"phones": phones_batch,
|
|
||||||
"phones_len": torch.LongTensor(phones_len_list).to(device),
|
|
||||||
"all_phones": all_phones_batch,
|
|
||||||
"all_phones_len": torch.LongTensor(all_phones_len_list).to(device),
|
|
||||||
"all_bert_features": all_bert_features_batch,
|
|
||||||
"norm_text": norm_text_batch
|
|
||||||
}
|
|
||||||
_data.append(batch)
|
|
||||||
|
|
||||||
return _data, batch_index_list
|
|
||||||
|
|
||||||
def recovery_order(self, data:list, batch_index_list:list)->list:
|
def recovery_order(self, data:list, batch_index_list:list)->list:
|
||||||
'''
|
'''
|
||||||
@ -568,7 +567,6 @@ class TTS:
|
|||||||
'''
|
'''
|
||||||
self.stop_flag = True
|
self.stop_flag = True
|
||||||
|
|
||||||
|
|
||||||
def run(self, inputs:dict):
|
def run(self, inputs:dict):
|
||||||
"""
|
"""
|
||||||
Text to speech inference.
|
Text to speech inference.
|
||||||
@ -596,156 +594,159 @@ class TTS:
|
|||||||
returns:
|
returns:
|
||||||
tuple[int, np.ndarray]: sampling rate and audio data.
|
tuple[int, np.ndarray]: sampling rate and audio data.
|
||||||
"""
|
"""
|
||||||
########## variables initialization ###########
|
|
||||||
self.stop_flag:bool = False
|
|
||||||
text:str = inputs.get("text", "")
|
|
||||||
text_lang:str = inputs.get("text_lang", "")
|
|
||||||
ref_audio_path:str = inputs.get("ref_audio_path", "")
|
|
||||||
prompt_text:str = inputs.get("prompt_text", "")
|
|
||||||
prompt_lang:str = inputs.get("prompt_lang", "")
|
|
||||||
top_k:int = inputs.get("top_k", 5)
|
|
||||||
top_p:float = inputs.get("top_p", 1)
|
|
||||||
temperature:float = inputs.get("temperature", 1)
|
|
||||||
text_split_method:str = inputs.get("text_split_method", "cut0")
|
|
||||||
batch_size = inputs.get("batch_size", 1)
|
|
||||||
batch_threshold = inputs.get("batch_threshold", 0.75)
|
|
||||||
speed_factor = inputs.get("speed_factor", 1.0)
|
|
||||||
split_bucket = inputs.get("split_bucket", True)
|
|
||||||
return_fragment = inputs.get("return_fragment", False)
|
|
||||||
fragment_interval = inputs.get("fragment_interval", 0.3)
|
|
||||||
seed = inputs.get("seed", -1)
|
|
||||||
seed = -1 if seed in ["", None] else seed
|
|
||||||
actual_seed = set_seed(seed)
|
|
||||||
|
|
||||||
if return_fragment:
|
def make_batch(batch_texts):
|
||||||
# split_bucket = False
|
batch_data = []
|
||||||
print(i18n("分段返回模式已开启"))
|
print(i18n("############ 提取文本Bert特征 ############"))
|
||||||
|
for text in tqdm(batch_texts):
|
||||||
|
phones, bert_features, norm_text = self.text_preprocessor.segment_and_extract_feature_for_text(text, text_lang)
|
||||||
|
if phones is None:
|
||||||
|
continue
|
||||||
|
res={
|
||||||
|
"phones": phones,
|
||||||
|
"bert_features": bert_features,
|
||||||
|
"norm_text": norm_text,
|
||||||
|
}
|
||||||
|
batch_data.append(res)
|
||||||
|
if len(batch_data) == 0:
|
||||||
|
return None
|
||||||
|
batch, _ = self.to_batch(batch_data,
|
||||||
|
prompt_data=self.prompt_cache if not no_prompt_text else None,
|
||||||
|
batch_size=batch_size,
|
||||||
|
threshold=batch_threshold,
|
||||||
|
split_bucket=False,
|
||||||
|
device=self.configs.device,
|
||||||
|
precision=self.precision
|
||||||
|
)
|
||||||
|
return batch[0]
|
||||||
|
|
||||||
|
# 直接给全体套一个torch.no_grad()
|
||||||
|
with torch.no_grad():
|
||||||
|
########## variables initialization ###########
|
||||||
|
self.stop_flag:bool = False
|
||||||
|
text:str = inputs.get("text", "")
|
||||||
|
text_lang:str = inputs.get("text_lang", "")
|
||||||
|
ref_audio_path:str = inputs.get("ref_audio_path", "")
|
||||||
|
prompt_text:str = inputs.get("prompt_text", "")
|
||||||
|
prompt_lang:str = inputs.get("prompt_lang", "")
|
||||||
|
top_k:int = inputs.get("top_k", 5)
|
||||||
|
top_p:float = inputs.get("top_p", 1)
|
||||||
|
temperature:float = inputs.get("temperature", 1)
|
||||||
|
text_split_method:str = inputs.get("text_split_method", "cut0")
|
||||||
|
batch_size = inputs.get("batch_size", 1)
|
||||||
|
batch_threshold = inputs.get("batch_threshold", 0.75)
|
||||||
|
speed_factor = inputs.get("speed_factor", 1.0)
|
||||||
|
split_bucket = inputs.get("split_bucket", True)
|
||||||
|
return_fragment = inputs.get("return_fragment", False)
|
||||||
|
fragment_interval = inputs.get("fragment_interval", 0.3)
|
||||||
|
seed = inputs.get("seed", -1)
|
||||||
|
seed = -1 if seed in ["", None] else seed
|
||||||
|
actual_seed = set_seed(seed)
|
||||||
|
|
||||||
|
if return_fragment:
|
||||||
|
# split_bucket = False
|
||||||
|
print(i18n("分段返回模式已开启"))
|
||||||
|
if split_bucket:
|
||||||
|
split_bucket = False
|
||||||
|
print(i18n("分段返回模式不支持分桶处理,已自动关闭分桶处理"))
|
||||||
|
|
||||||
if split_bucket:
|
if split_bucket:
|
||||||
split_bucket = False
|
print(i18n("分桶处理模式已开启"))
|
||||||
print(i18n("分段返回模式不支持分桶处理,已自动关闭分桶处理"))
|
|
||||||
|
|
||||||
if split_bucket:
|
if fragment_interval<0.01:
|
||||||
print(i18n("分桶处理模式已开启"))
|
fragment_interval = 0.01
|
||||||
|
print(i18n("分段间隔过小,已自动设置为0.01"))
|
||||||
|
|
||||||
if fragment_interval<0.01:
|
no_prompt_text = False
|
||||||
fragment_interval = 0.01
|
if prompt_text in [None, ""]:
|
||||||
print(i18n("分段间隔过小,已自动设置为0.01"))
|
no_prompt_text = True
|
||||||
|
|
||||||
no_prompt_text = False
|
assert text_lang in self.configs.languages
|
||||||
if prompt_text in [None, ""]:
|
if not no_prompt_text:
|
||||||
no_prompt_text = True
|
assert prompt_lang in self.configs.languages
|
||||||
|
|
||||||
assert text_lang in self.configs.languages
|
if ref_audio_path in [None, ""] and \
|
||||||
if not no_prompt_text:
|
((self.prompt_cache["prompt_semantic"] is None) or (self.prompt_cache["refer_spec"] is None)):
|
||||||
assert prompt_lang in self.configs.languages
|
raise ValueError("ref_audio_path cannot be empty, when the reference audio is not set using set_ref_audio()")
|
||||||
|
|
||||||
if ref_audio_path in [None, ""] and \
|
###### setting reference audio and prompt text preprocessing ########
|
||||||
((self.prompt_cache["prompt_semantic"] is None) or (self.prompt_cache["refer_spec"] is None)):
|
t0 = ttime()
|
||||||
raise ValueError("ref_audio_path cannot be empty, when the reference audio is not set using set_ref_audio()")
|
if (ref_audio_path is not None) and (ref_audio_path != self.prompt_cache["ref_audio_path"]):
|
||||||
|
|
||||||
|
|
||||||
###### setting reference audio and prompt text preprocessing ########
|
|
||||||
t0 = ttime()
|
|
||||||
if (ref_audio_path is not None) and (ref_audio_path != self.prompt_cache["ref_audio_path"]):
|
|
||||||
self.set_ref_audio(ref_audio_path)
|
self.set_ref_audio(ref_audio_path)
|
||||||
|
|
||||||
if not no_prompt_text:
|
if not no_prompt_text:
|
||||||
prompt_text = prompt_text.strip("\n")
|
prompt_text = prompt_text.strip("\n")
|
||||||
if (prompt_text[-1] not in splits): prompt_text += "。" if prompt_lang != "en" else "."
|
if (prompt_text[-1] not in splits): prompt_text += "。" if prompt_lang != "en" else "."
|
||||||
print(i18n("实际输入的参考文本:"), prompt_text)
|
print(i18n("实际输入的参考文本:"), prompt_text)
|
||||||
if self.prompt_cache["prompt_text"] != prompt_text:
|
if self.prompt_cache["prompt_text"] != prompt_text:
|
||||||
self.prompt_cache["prompt_text"] = prompt_text
|
self.prompt_cache["prompt_text"] = prompt_text
|
||||||
self.prompt_cache["prompt_lang"] = prompt_lang
|
self.prompt_cache["prompt_lang"] = prompt_lang
|
||||||
phones, bert_features, norm_text = \
|
phones, bert_features, norm_text = \
|
||||||
self.text_preprocessor.segment_and_extract_feature_for_text(
|
self.text_preprocessor.segment_and_extract_feature_for_text(
|
||||||
prompt_text,
|
prompt_text,
|
||||||
prompt_lang)
|
prompt_lang)
|
||||||
self.prompt_cache["phones"] = phones
|
self.prompt_cache["phones"] = phones
|
||||||
self.prompt_cache["bert_features"] = bert_features
|
self.prompt_cache["bert_features"] = bert_features
|
||||||
self.prompt_cache["norm_text"] = norm_text
|
self.prompt_cache["norm_text"] = norm_text
|
||||||
|
|
||||||
|
###### text preprocessing ########
|
||||||
|
t1 = ttime()
|
||||||
|
data:list = None
|
||||||
|
if not return_fragment:
|
||||||
|
data = self.text_preprocessor.preprocess(text, text_lang, text_split_method)
|
||||||
|
if len(data) == 0:
|
||||||
|
yield self.configs.sampling_rate, np.zeros(int(self.configs.sampling_rate),
|
||||||
|
dtype=np.int16)
|
||||||
|
return
|
||||||
|
|
||||||
|
batch_index_list:list = None
|
||||||
|
data, batch_index_list = self.to_batch(data,
|
||||||
|
prompt_data=self.prompt_cache if not no_prompt_text else None,
|
||||||
|
batch_size=batch_size,
|
||||||
|
threshold=batch_threshold,
|
||||||
|
split_bucket=split_bucket,
|
||||||
|
device=self.configs.device,
|
||||||
|
precision=self.precision
|
||||||
|
)
|
||||||
|
else:
|
||||||
|
print(i18n("############ 切分文本 ############"))
|
||||||
|
texts = self.text_preprocessor.pre_seg_text(text, text_lang, text_split_method)
|
||||||
|
data = []
|
||||||
|
for i in range(len(texts)):
|
||||||
|
if i%batch_size == 0:
|
||||||
|
data.append([])
|
||||||
|
data[-1].append(texts[i])
|
||||||
|
|
||||||
|
|
||||||
###### text preprocessing ########
|
|
||||||
t1 = ttime()
|
|
||||||
data:list = None
|
|
||||||
if not return_fragment:
|
|
||||||
data = self.text_preprocessor.preprocess(text, text_lang, text_split_method)
|
|
||||||
if len(data) == 0:
|
|
||||||
yield self.configs.sampling_rate, np.zeros(int(self.configs.sampling_rate),
|
|
||||||
dtype=np.int16)
|
|
||||||
return
|
|
||||||
|
|
||||||
batch_index_list:list = None
|
t2 = ttime()
|
||||||
data, batch_index_list = self.to_batch(data,
|
try:
|
||||||
prompt_data=self.prompt_cache if not no_prompt_text else None,
|
print("############ 推理 ############")
|
||||||
batch_size=batch_size,
|
###### inference ######
|
||||||
threshold=batch_threshold,
|
t_34 = 0.0
|
||||||
split_bucket=split_bucket,
|
t_45 = 0.0
|
||||||
device=self.configs.device,
|
audio = []
|
||||||
precision=self.precision
|
for item in data:
|
||||||
)
|
t3 = ttime()
|
||||||
else:
|
if return_fragment:
|
||||||
print(i18n("############ 切分文本 ############"))
|
item = make_batch(item)
|
||||||
texts = self.text_preprocessor.pre_seg_text(text, text_lang, text_split_method)
|
if item is None:
|
||||||
data = []
|
continue
|
||||||
for i in range(len(texts)):
|
|
||||||
if i%batch_size == 0:
|
|
||||||
data.append([])
|
|
||||||
data[-1].append(texts[i])
|
|
||||||
|
|
||||||
def make_batch(batch_texts):
|
batch_phones:List[torch.LongTensor] = item["phones"]
|
||||||
batch_data = []
|
batch_phones_len:torch.LongTensor = item["phones_len"]
|
||||||
print(i18n("############ 提取文本Bert特征 ############"))
|
all_phoneme_ids:List[torch.LongTensor] = item["all_phones"]
|
||||||
for text in tqdm(batch_texts):
|
all_phoneme_lens:torch.LongTensor = item["all_phones_len"]
|
||||||
phones, bert_features, norm_text = self.text_preprocessor.segment_and_extract_feature_for_text(text, text_lang)
|
all_bert_features:List[torch.LongTensor] = item["all_bert_features"]
|
||||||
if phones is None:
|
norm_text:str = item["norm_text"]
|
||||||
continue
|
|
||||||
res={
|
|
||||||
"phones": phones,
|
|
||||||
"bert_features": bert_features,
|
|
||||||
"norm_text": norm_text,
|
|
||||||
}
|
|
||||||
batch_data.append(res)
|
|
||||||
if len(batch_data) == 0:
|
|
||||||
return None
|
|
||||||
batch, _ = self.to_batch(batch_data,
|
|
||||||
prompt_data=self.prompt_cache if not no_prompt_text else None,
|
|
||||||
batch_size=batch_size,
|
|
||||||
threshold=batch_threshold,
|
|
||||||
split_bucket=False,
|
|
||||||
device=self.configs.device,
|
|
||||||
precision=self.precision
|
|
||||||
)
|
|
||||||
return batch[0]
|
|
||||||
|
|
||||||
t2 = ttime()
|
print(i18n("前端处理后的文本(每句):"), norm_text)
|
||||||
try:
|
if no_prompt_text :
|
||||||
print("############ 推理 ############")
|
prompt = None
|
||||||
###### inference ######
|
else:
|
||||||
t_34 = 0.0
|
prompt = self.prompt_cache["prompt_semantic"].expand(len(all_phoneme_ids), -1).to(self.configs.device)
|
||||||
t_45 = 0.0
|
|
||||||
audio = []
|
|
||||||
for item in data:
|
|
||||||
t3 = ttime()
|
|
||||||
if return_fragment:
|
|
||||||
item = make_batch(item)
|
|
||||||
if item is None:
|
|
||||||
continue
|
|
||||||
|
|
||||||
batch_phones:List[torch.LongTensor] = item["phones"]
|
|
||||||
batch_phones_len:torch.LongTensor = item["phones_len"]
|
|
||||||
all_phoneme_ids:List[torch.LongTensor] = item["all_phones"]
|
|
||||||
all_phoneme_lens:torch.LongTensor = item["all_phones_len"]
|
|
||||||
all_bert_features:List[torch.LongTensor] = item["all_bert_features"]
|
|
||||||
norm_text:str = item["norm_text"]
|
|
||||||
|
|
||||||
print(i18n("前端处理后的文本(每句):"), norm_text)
|
|
||||||
if no_prompt_text :
|
|
||||||
prompt = None
|
|
||||||
else:
|
|
||||||
prompt = self.prompt_cache["prompt_semantic"].expand(len(all_phoneme_ids), -1).to(self.configs.device)
|
|
||||||
|
|
||||||
with torch.no_grad():
|
|
||||||
pred_semantic_list, idx_list = self.t2s_model.model.infer_panel(
|
pred_semantic_list, idx_list = self.t2s_model.model.infer_panel(
|
||||||
all_phoneme_ids,
|
all_phoneme_ids,
|
||||||
all_phoneme_lens,
|
all_phoneme_lens,
|
||||||
@ -757,96 +758,98 @@ class TTS:
|
|||||||
temperature=temperature,
|
temperature=temperature,
|
||||||
early_stop_num=self.configs.hz * self.configs.max_sec,
|
early_stop_num=self.configs.hz * self.configs.max_sec,
|
||||||
)
|
)
|
||||||
t4 = ttime()
|
t4 = ttime()
|
||||||
t_34 += t4 - t3
|
t_34 += t4 - t3
|
||||||
|
|
||||||
refer_audio_spec:torch.Tensor = self.prompt_cache["refer_spec"]\
|
refer_audio_spec:torch.Tensor = self.prompt_cache["refer_spec"]\
|
||||||
.to(dtype=self.precision, device=self.configs.device)
|
.to(dtype=self.precision, device=self.configs.device)
|
||||||
|
|
||||||
batch_audio_fragment = []
|
batch_audio_fragment = []
|
||||||
|
|
||||||
# ## vits并行推理 method 1
|
# 这里要记得加 torch.no_grad() 不然速度慢一大截
|
||||||
# pred_semantic_list = [item[-idx:] for item, idx in zip(pred_semantic_list, idx_list)]
|
# with torch.no_grad():
|
||||||
# pred_semantic_len = torch.LongTensor([item.shape[0] for item in pred_semantic_list]).to(self.configs.device)
|
|
||||||
# pred_semantic = self.batch_sequences(pred_semantic_list, axis=0, pad_value=0).unsqueeze(0)
|
|
||||||
# max_len = 0
|
|
||||||
# for i in range(0, len(batch_phones)):
|
|
||||||
# max_len = max(max_len, batch_phones[i].shape[-1])
|
|
||||||
# batch_phones = self.batch_sequences(batch_phones, axis=0, pad_value=0, max_length=max_len)
|
|
||||||
# batch_phones = batch_phones.to(self.configs.device)
|
|
||||||
# batch_audio_fragment = (self.vits_model.batched_decode(
|
|
||||||
# pred_semantic, pred_semantic_len, batch_phones, batch_phones_len,refer_audio_spec
|
|
||||||
# ))
|
|
||||||
|
|
||||||
# ## vits并行推理 method 2
|
# ## vits并行推理 method 1
|
||||||
pred_semantic_list = [item[-idx:] for item, idx in zip(pred_semantic_list, idx_list)]
|
# pred_semantic_list = [item[-idx:] for item, idx in zip(pred_semantic_list, idx_list)]
|
||||||
upsample_rate = math.prod(self.vits_model.upsample_rates)
|
# pred_semantic_len = torch.LongTensor([item.shape[0] for item in pred_semantic_list]).to(self.configs.device)
|
||||||
audio_frag_idx = [pred_semantic_list[i].shape[0]*2*upsample_rate for i in range(0, len(pred_semantic_list))]
|
# pred_semantic = self.batch_sequences(pred_semantic_list, axis=0, pad_value=0).unsqueeze(0)
|
||||||
audio_frag_end_idx = [ sum(audio_frag_idx[:i+1]) for i in range(0, len(audio_frag_idx))]
|
# max_len = 0
|
||||||
all_pred_semantic = torch.cat(pred_semantic_list).unsqueeze(0).unsqueeze(0).to(self.configs.device)
|
# for i in range(0, len(batch_phones)):
|
||||||
_batch_phones = torch.cat(batch_phones).unsqueeze(0).to(self.configs.device)
|
# max_len = max(max_len, batch_phones[i].shape[-1])
|
||||||
_batch_audio_fragment = (self.vits_model.decode(
|
# batch_phones = self.batch_sequences(batch_phones, axis=0, pad_value=0, max_length=max_len)
|
||||||
all_pred_semantic, _batch_phones, refer_audio_spec
|
# batch_phones = batch_phones.to(self.configs.device)
|
||||||
).detach()[0, 0, :])
|
# batch_audio_fragment = (self.vits_model.batched_decode(
|
||||||
audio_frag_end_idx.insert(0, 0)
|
# pred_semantic, pred_semantic_len, batch_phones, batch_phones_len,refer_audio_spec
|
||||||
batch_audio_fragment= [_batch_audio_fragment[audio_frag_end_idx[i-1]:audio_frag_end_idx[i]] for i in range(1, len(audio_frag_end_idx))]
|
# ))
|
||||||
|
|
||||||
|
# ## vits并行推理 method 2
|
||||||
|
pred_semantic_list = [item[-idx:] for item, idx in zip(pred_semantic_list, idx_list)]
|
||||||
|
upsample_rate = math.prod(self.vits_model.upsample_rates)
|
||||||
|
audio_frag_idx = [pred_semantic_list[i].shape[0]*2*upsample_rate for i in range(0, len(pred_semantic_list))]
|
||||||
|
audio_frag_end_idx = [ sum(audio_frag_idx[:i+1]) for i in range(0, len(audio_frag_idx))]
|
||||||
|
all_pred_semantic = torch.cat(pred_semantic_list).unsqueeze(0).unsqueeze(0).to(self.configs.device)
|
||||||
|
_batch_phones = torch.cat(batch_phones).unsqueeze(0).to(self.configs.device)
|
||||||
|
_batch_audio_fragment = (self.vits_model.decode(
|
||||||
|
all_pred_semantic, _batch_phones, refer_audio_spec
|
||||||
|
).detach()[0, 0, :])
|
||||||
|
audio_frag_end_idx.insert(0, 0)
|
||||||
|
batch_audio_fragment= [_batch_audio_fragment[audio_frag_end_idx[i-1]:audio_frag_end_idx[i]] for i in range(1, len(audio_frag_end_idx))]
|
||||||
|
|
||||||
# ## vits串行推理
|
# ## vits串行推理
|
||||||
# for i, idx in enumerate(idx_list):
|
# for i, idx in enumerate(idx_list):
|
||||||
# phones = batch_phones[i].unsqueeze(0).to(self.configs.device)
|
# phones = batch_phones[i].unsqueeze(0).to(self.configs.device)
|
||||||
# _pred_semantic = (pred_semantic_list[i][-idx:].unsqueeze(0).unsqueeze(0)) # .unsqueeze(0)#mq要多unsqueeze一次
|
# _pred_semantic = (pred_semantic_list[i][-idx:].unsqueeze(0).unsqueeze(0)) # .unsqueeze(0)#mq要多unsqueeze一次
|
||||||
# audio_fragment =(self.vits_model.decode(
|
# audio_fragment =(self.vits_model.decode(
|
||||||
# _pred_semantic, phones, refer_audio_spec
|
# _pred_semantic, phones, refer_audio_spec
|
||||||
# ).detach()[0, 0, :])
|
# ).detach()[0, 0, :])
|
||||||
# batch_audio_fragment.append(
|
# batch_audio_fragment.append(
|
||||||
# audio_fragment
|
# audio_fragment
|
||||||
# ) ###试试重建不带上prompt部分
|
# ) ###试试重建不带上prompt部分
|
||||||
|
|
||||||
t5 = ttime()
|
t5 = ttime()
|
||||||
t_45 += t5 - t4
|
t_45 += t5 - t4
|
||||||
if return_fragment:
|
if return_fragment:
|
||||||
print("%.3f\t%.3f\t%.3f\t%.3f" % (t1 - t0, t2 - t1, t4 - t3, t5 - t4))
|
print("%.3f\t%.3f\t%.3f\t%.3f" % (t1 - t0, t2 - t1, t4 - t3, t5 - t4))
|
||||||
yield self.audio_postprocess([batch_audio_fragment],
|
yield self.audio_postprocess([batch_audio_fragment],
|
||||||
|
self.configs.sampling_rate,
|
||||||
|
None,
|
||||||
|
speed_factor,
|
||||||
|
False,
|
||||||
|
fragment_interval
|
||||||
|
)
|
||||||
|
else:
|
||||||
|
audio.append(batch_audio_fragment)
|
||||||
|
|
||||||
|
if self.stop_flag:
|
||||||
|
yield self.configs.sampling_rate, np.zeros(int(self.configs.sampling_rate),
|
||||||
|
dtype=np.int16)
|
||||||
|
return
|
||||||
|
|
||||||
|
if not return_fragment:
|
||||||
|
print("%.3f\t%.3f\t%.3f\t%.3f" % (t1 - t0, t2 - t1, t_34, t_45))
|
||||||
|
yield self.audio_postprocess(audio,
|
||||||
self.configs.sampling_rate,
|
self.configs.sampling_rate,
|
||||||
None,
|
batch_index_list,
|
||||||
speed_factor,
|
speed_factor,
|
||||||
False,
|
split_bucket,
|
||||||
fragment_interval
|
fragment_interval
|
||||||
)
|
)
|
||||||
else:
|
|
||||||
audio.append(batch_audio_fragment)
|
|
||||||
|
|
||||||
if self.stop_flag:
|
except Exception as e:
|
||||||
yield self.configs.sampling_rate, np.zeros(int(self.configs.sampling_rate),
|
traceback.print_exc()
|
||||||
dtype=np.int16)
|
# 必须返回一个空音频, 否则会导致显存不释放。
|
||||||
return
|
yield self.configs.sampling_rate, np.zeros(int(self.configs.sampling_rate),
|
||||||
|
dtype=np.int16)
|
||||||
if not return_fragment:
|
# 重置模型, 否则会导致显存释放不完全。
|
||||||
print("%.3f\t%.3f\t%.3f\t%.3f" % (t1 - t0, t2 - t1, t_34, t_45))
|
del self.t2s_model
|
||||||
yield self.audio_postprocess(audio,
|
del self.vits_model
|
||||||
self.configs.sampling_rate,
|
self.t2s_model = None
|
||||||
batch_index_list,
|
self.vits_model = None
|
||||||
speed_factor,
|
self.init_t2s_weights(self.configs.t2s_weights_path)
|
||||||
split_bucket,
|
self.init_vits_weights(self.configs.vits_weights_path)
|
||||||
fragment_interval
|
raise e
|
||||||
)
|
finally:
|
||||||
|
self.empty_cache()
|
||||||
except Exception as e:
|
|
||||||
traceback.print_exc()
|
|
||||||
# 必须返回一个空音频, 否则会导致显存不释放。
|
|
||||||
yield self.configs.sampling_rate, np.zeros(int(self.configs.sampling_rate),
|
|
||||||
dtype=np.int16)
|
|
||||||
# 重置模型, 否则会导致显存释放不完全。
|
|
||||||
del self.t2s_model
|
|
||||||
del self.vits_model
|
|
||||||
self.t2s_model = None
|
|
||||||
self.vits_model = None
|
|
||||||
self.init_t2s_weights(self.configs.t2s_weights_path)
|
|
||||||
self.init_vits_weights(self.configs.vits_weights_path)
|
|
||||||
raise e
|
|
||||||
finally:
|
|
||||||
self.empty_cache()
|
|
||||||
|
|
||||||
def empty_cache(self):
|
def empty_cache(self):
|
||||||
try:
|
try:
|
||||||
@ -878,14 +881,12 @@ class TTS:
|
|||||||
audio_fragment:torch.Tensor = torch.cat([audio_fragment, zero_wav], dim=0)
|
audio_fragment:torch.Tensor = torch.cat([audio_fragment, zero_wav], dim=0)
|
||||||
audio[i][j] = audio_fragment.cpu().numpy()
|
audio[i][j] = audio_fragment.cpu().numpy()
|
||||||
|
|
||||||
|
|
||||||
if split_bucket:
|
if split_bucket:
|
||||||
audio = self.recovery_order(audio, batch_index_list)
|
audio = self.recovery_order(audio, batch_index_list)
|
||||||
else:
|
else:
|
||||||
# audio = [item for batch in audio for item in batch]
|
# audio = [item for batch in audio for item in batch]
|
||||||
audio = sum(audio, [])
|
audio = sum(audio, [])
|
||||||
|
|
||||||
|
|
||||||
audio = np.concatenate(audio, 0)
|
audio = np.concatenate(audio, 0)
|
||||||
audio = (audio * 32768).astype(np.int16)
|
audio = (audio * 32768).astype(np.int16)
|
||||||
|
|
||||||
@ -898,8 +899,6 @@ class TTS:
|
|||||||
return sr, audio
|
return sr, audio
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
def speed_change(input_audio:np.ndarray, speed:float, sr:int):
|
def speed_change(input_audio:np.ndarray, speed:float, sr:int):
|
||||||
# 将 NumPy 数组转换为原始 PCM 流
|
# 将 NumPy 数组转换为原始 PCM 流
|
||||||
raw_audio = input_audio.astype(np.int16).tobytes()
|
raw_audio = input_audio.astype(np.int16).tobytes()
|
||||||
|
Loading…
x
Reference in New Issue
Block a user