This commit is contained in:
Jarod Mica 2024-12-23 02:03:19 -08:00
parent 91c0466294
commit 55badf45c5
16 changed files with 19 additions and 19 deletions

View File

@ -15,7 +15,7 @@ from GPT_SoVITS.text import cleaned_text_to_sequence
from transformers import AutoModelForMaskedLM, AutoTokenizer from transformers import AutoModelForMaskedLM, AutoTokenizer
from TTS_infer_pack.text_segmentation_method import split_big_text, splits, get_method as get_seg_method from TTS_infer_pack.text_segmentation_method import split_big_text, splits, get_method as get_seg_method
from tools.i18n.i18n import I18nAuto, scan_language_list from GPT_SoVITS.tools..i18n.i18n import I18nAuto, scan_language_list
language=os.environ.get("language","Auto") language=os.environ.get("language","Auto")
language=sys.argv[-1] if sys.argv[-1] in scan_language_list() else language language=sys.argv[-1] if sys.argv[-1] in scan_language_list() else language

View File

@ -117,7 +117,7 @@ import uvicorn
from importlib.resources import files from importlib.resources import files
from io import BytesIO from io import BytesIO
from tools.i18n.i18n import I18nAuto from GPT_SoVITS.tools..i18n.i18n import I18nAuto
from GPT_SoVITS.TTS_infer_pack.TTS import TTS, TTS_Config from GPT_SoVITS.TTS_infer_pack.TTS import TTS, TTS_Config
from GPT_SoVITS.TTS_infer_pack.text_segmentation_method import get_method_names as get_cut_method_names from GPT_SoVITS.TTS_infer_pack.text_segmentation_method import get_method_names as get_cut_method_names
from fastapi.responses import StreamingResponse from fastapi.responses import StreamingResponse

View File

@ -2,7 +2,7 @@ import argparse
import os import os
import soundfile as sf import soundfile as sf
from tools.i18n.i18n import I18nAuto from GPT_SoVITS.tools..i18n.i18n import I18nAuto
from GPT_SoVITS.inference_webui import change_gpt_weights, change_sovits_weights, get_tts_wav from GPT_SoVITS.inference_webui import change_gpt_weights, change_sovits_weights, get_tts_wav
i18n = I18nAuto() i18n = I18nAuto()

View File

@ -5,7 +5,7 @@ from PyQt5.QtWidgets import QApplication, QMainWindow, QLabel, QLineEdit, QPushB
from PyQt5.QtWidgets import QGridLayout, QVBoxLayout, QWidget, QFileDialog, QStatusBar, QComboBox from PyQt5.QtWidgets import QGridLayout, QVBoxLayout, QWidget, QFileDialog, QStatusBar, QComboBox
import soundfile as sf import soundfile as sf
from tools.i18n.i18n import I18nAuto from GPT_SoVITS.tools..i18n.i18n import I18nAuto
i18n = I18nAuto() i18n = I18nAuto()
from inference_webui import gpt_path, sovits_path, change_gpt_weights, change_sovits_weights, get_tts_wav from inference_webui import gpt_path, sovits_path, change_gpt_weights, change_sovits_weights, get_tts_wav

View File

@ -89,8 +89,8 @@ from GPT_SoVITS.text import cleaned_text_to_sequence
from GPT_SoVITS.text.cleaner import clean_text from GPT_SoVITS.text.cleaner import clean_text
from time import time as ttime from time import time as ttime
from GPT_SoVITS.module.mel_processing import spectrogram_torch from GPT_SoVITS.module.mel_processing import spectrogram_torch
from tools.my_utils import load_audio from GPT_SoVITS.tools..my_utils import load_audio
from tools.i18n.i18n import I18nAuto, scan_language_list from GPT_SoVITS.tools..i18n.i18n import I18nAuto, scan_language_list
language=os.environ.get("language","Auto") language=os.environ.get("language","Auto")
language=sys.argv[-1] if sys.argv[-1] in scan_language_list() else language language=sys.argv[-1] if sys.argv[-1] in scan_language_list() else language

View File

@ -46,7 +46,7 @@ version=os.environ.get("version","v2")
import gradio as gr import gradio as gr
from TTS_infer_pack.TTS import TTS, TTS_Config from TTS_infer_pack.TTS import TTS, TTS_Config
from TTS_infer_pack.text_segmentation_method import get_method from TTS_infer_pack.text_segmentation_method import get_method
from tools.i18n.i18n import I18nAuto, scan_language_list from GPT_SoVITS.tools..i18n.i18n import I18nAuto, scan_language_list
language=os.environ.get("language","Auto") language=os.environ.get("language","Auto")
language=sys.argv[-1] if sys.argv[-1] in scan_language_list() else language language=sys.argv[-1] if sys.argv[-1] in scan_language_list() else language

View File

@ -10,7 +10,7 @@ cnhubert.cnhubert_base_path = cnhubert_base_path
ssl_model = cnhubert.get_model() ssl_model = cnhubert.get_model()
from GPT_SoVITS.text import cleaned_text_to_sequence from GPT_SoVITS.text import cleaned_text_to_sequence
import soundfile import soundfile
from tools.my_utils import load_audio from GPT_SoVITS.tools..my_utils import load_audio
import os import os
import json import json

View File

@ -21,7 +21,7 @@ from tqdm import tqdm
from GPT_SoVITS.text.cleaner import clean_text from GPT_SoVITS.text.cleaner import clean_text
from transformers import AutoModelForMaskedLM, AutoTokenizer from transformers import AutoModelForMaskedLM, AutoTokenizer
import numpy as np import numpy as np
from tools.my_utils import clean_path from GPT_SoVITS.tools..my_utils import clean_path
# inp_text=sys.argv[1] # inp_text=sys.argv[1]
# inp_wav_dir=sys.argv[2] # inp_wav_dir=sys.argv[2]

View File

@ -19,7 +19,7 @@ from scipy.io import wavfile
import librosa import librosa
now_dir = os.getcwd() now_dir = os.getcwd()
sys.path.append(now_dir) sys.path.append(now_dir)
from tools.my_utils import load_audio,clean_path from GPT_SoVITS.tools..my_utils import load_audio,clean_path
# from config import cnhubert_base_path # from config import cnhubert_base_path
# cnhubert.cnhubert_base_path=cnhubert_base_path # cnhubert.cnhubert_base_path=cnhubert_base_path

View File

@ -24,7 +24,7 @@ from glob import glob
from tqdm import tqdm from tqdm import tqdm
import logging, librosa, utils import logging, librosa, utils
from GPT_SoVITS.module.models import SynthesizerTrn from GPT_SoVITS.module.models import SynthesizerTrn
from tools.my_utils import clean_path from GPT_SoVITS.tools..my_utils import clean_path
logging.getLogger("numba").setLevel(logging.WARNING) logging.getLogger("numba").setLevel(logging.WARNING)
# from config import pretrained_s2G # from config import pretrained_s2G

View File

@ -3,7 +3,7 @@ from collections import OrderedDict
from time import time as ttime from time import time as ttime
import shutil,os import shutil,os
import torch import torch
from tools.i18n.i18n import I18nAuto from GPT_SoVITS.tools..i18n.i18n import I18nAuto
i18n = I18nAuto() i18n = I18nAuto()

View File

@ -9,7 +9,7 @@ import torch
from faster_whisper import WhisperModel from faster_whisper import WhisperModel
from tqdm import tqdm from tqdm import tqdm
from tools.asr.config import check_fw_local_models from GPT_SoVITS.tools..asr.config import check_fw_local_models
language_code_list = [ language_code_list = [
"af", "am", "ar", "as", "az", "af", "am", "ar", "as", "az",
@ -69,7 +69,7 @@ def execute_asr(input_folder, output_folder, model_size, language, precision):
if info.language == "zh": if info.language == "zh":
print("检测为中文文本, 转 FunASR 处理") print("检测为中文文本, 转 FunASR 处理")
if("only_asr" not in globals()): if("only_asr" not in globals()):
from tools.asr.funasr_asr import only_asr #如果用英文就不需要导入下载模型 from GPT_SoVITS.tools..asr.funasr_asr import only_asr #如果用英文就不需要导入下载模型
text = only_asr(file_path, language=info.language.lower()) text = only_asr(file_path, language=info.language.lower())
if text == '': if text == '':

View File

@ -2,7 +2,7 @@ import platform,os,traceback
import ffmpeg import ffmpeg
import numpy as np import numpy as np
import gradio as gr import gradio as gr
from tools.i18n.i18n import I18nAuto from GPT_SoVITS.tools..i18n.i18n import I18nAuto
import pandas as pd import pandas as pd
i18n = I18nAuto(language=os.environ.get('language','Auto')) i18n = I18nAuto(language=os.environ.get('language','Auto'))

View File

@ -3,7 +3,7 @@ import traceback
from scipy.io import wavfile from scipy.io import wavfile
# parent_directory = os.path.dirname(os.path.abspath(__file__)) # parent_directory = os.path.dirname(os.path.abspath(__file__))
# sys.path.append(parent_directory) # sys.path.append(parent_directory)
from tools.my_utils import load_audio from GPT_SoVITS.tools..my_utils import load_audio
from slicer2 import Slicer from slicer2 import Slicer
def slice(inp,opt_root,threshold,min_length,min_interval,hop_size,max_sil_kept,_max,alpha,i_part,all_part): def slice(inp,opt_root,threshold,min_length,min_interval,hop_size,max_sil_kept,_max,alpha,i_part,all_part):

View File

@ -1,8 +1,8 @@
import os import os
import traceback,gradio as gr import traceback,gradio as gr
import logging import logging
from tools.i18n.i18n import I18nAuto from GPT_SoVITS.tools..i18n.i18n import I18nAuto
from tools.my_utils import clean_path from GPT_SoVITS.tools..my_utils import clean_path
i18n = I18nAuto() i18n = I18nAuto()
logger = logging.getLogger(__name__) logger = logging.getLogger(__name__)

2
api.py
View File

@ -167,7 +167,7 @@ from GPT_SoVITS.AR.models.t2s_lightning_module import Text2SemanticLightningModu
from GPT_SoVITS.text import cleaned_text_to_sequence from GPT_SoVITS.text import cleaned_text_to_sequence
from GPT_SoVITS.text.cleaner import clean_text from GPT_SoVITS.text.cleaner import clean_text
from GPT_SoVITS.module.mel_processing import spectrogram_torch from GPT_SoVITS.module.mel_processing import spectrogram_torch
from tools.my_utils import load_audio from GPT_SoVITS.tools..my_utils import load_audio
import config as global_config import config as global_config
import logging import logging
import subprocess import subprocess