diff --git a/README.md b/README.md index 2283e13..397af31 100644 --- a/README.md +++ b/README.md @@ -65,7 +65,6 @@ bash install.sh pip install -r requirements.txt ``` - #### FFmpeg ##### Conda Users diff --git a/requirements.txt b/requirements.txt index 1bafeef..71576f1 100644 --- a/requirements.txt +++ b/requirements.txt @@ -8,7 +8,7 @@ gradio==3.38.0 ffmpeg-python onnxruntime tqdm -funasr==0.8.7 +funasr>=1.0.0 cn2an pypinyin pyopenjtalk diff --git a/tools/damo_asr/cmd-asr.py b/tools/damo_asr/cmd-asr.py index 2775562..70dd4d8 100644 --- a/tools/damo_asr/cmd-asr.py +++ b/tools/damo_asr/cmd-asr.py @@ -1,8 +1,9 @@ # -*- coding:utf-8 -*- -from modelscope.pipelines import pipeline -from modelscope.utils.constant import Tasks import sys,os,traceback + +from funasr import AutoModel + dir=sys.argv[1] # opt_name=dir.split("\\")[-1].split("/")[-1] opt_name=os.path.basename(dir) @@ -10,20 +11,22 @@ opt_name=os.path.basename(dir) path_asr='tools/damo_asr/models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch' path_vad='tools/damo_asr/models/speech_fsmn_vad_zh-cn-16k-common-pytorch' path_punc='tools/damo_asr/models/punc_ct-transformer_zh-cn-common-vocab272727-pytorch' -path_asr=path_asr if os.path.exists(path_asr)else "damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch" -path_vad=path_vad if os.path.exists(path_vad)else "damo/speech_fsmn_vad_zh-cn-16k-common-pytorch" -path_punc=path_punc if os.path.exists(path_punc)else "damo/punc_ct-transformer_zh-cn-common-vocab272727-pytorch" -inference_pipeline = pipeline( - task=Tasks.auto_speech_recognition, - model=path_asr, - vad_model=path_vad, - punc_model=path_punc, -) +path_asr=path_asr if os.path.exists(path_asr)else "iic/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch" +path_vad=path_vad if os.path.exists(path_vad)else "iic/speech_fsmn_vad_zh-cn-16k-common-pytorch" +path_punc=path_punc if os.path.exists(path_punc)else "iic/punc_ct-transformer_zh-cn-common-vocab272727-pytorch" + +model = AutoModel(model=path_asr, model_revision="v2.0.4", + vad_model=path_vad, + vad_model_revision="v2.0.4", + punc_model=path_punc, + punc_model_revision="v2.0.4", + ) + opt=[] for name in os.listdir(dir): try: - text = inference_pipeline(audio_in="%s/%s"%(dir,name))["text"] + text = model.generate(input="%s/%s"%(dir,name))[0]["text"] opt.append("%s/%s|%s|ZH|%s"%(dir,name,opt_name,text)) except: print(traceback.format_exc()) @@ -31,4 +34,3 @@ for name in os.listdir(dir): opt_dir="output/asr_opt" os.makedirs(opt_dir,exist_ok=True) with open("%s/%s.list"%(opt_dir,opt_name),"w",encoding="utf-8")as f:f.write("\n".join(opt)) -