From 03d1e0ace04189e4083efd2cb3e1f8f9877bda3a Mon Sep 17 00:00:00 2001 From: onlybetheone Date: Wed, 15 Feb 2023 21:28:25 +0800 Subject: [PATCH] add infer decoding model param --- .../infer.py | 2 +- .../infer.py | 2 +- .../infer.py | 2 +- .../infer.py | 2 +- .../infer.py | 2 +- .../infer.py | 2 +- .../infer.py | 2 +- .../infer.py | 2 +- .../infer.py | 2 +- .../infer.py | 2 +- .../infer.py | 2 +- .../infer.py | 2 +- 12 files changed, 12 insertions(+), 12 deletions(-) diff --git a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-cantonese-CHS-16k-common-vocab1468-tensorflow1-online/infer.py b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-cantonese-CHS-16k-common-vocab1468-tensorflow1-online/infer.py index 5cec9dd12..ac73adf72 100644 --- a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-cantonese-CHS-16k-common-vocab1468-tensorflow1-online/infer.py +++ b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-cantonese-CHS-16k-common-vocab1468-tensorflow1-online/infer.py @@ -9,5 +9,5 @@ if __name__ == "__main__": model="damo/speech_UniASR_asr_2pass-cantonese-CHS-16k-common-vocab1468-tensorflow1-online", output_dir=output_dir, ) - rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"online"}) + rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"normal"}) print(rec_result) diff --git a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-de-16k-common-vocab3690-tensorflow1-online/infer.py b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-de-16k-common-vocab3690-tensorflow1-online/infer.py index 5bd9755fb..f8d91b833 100644 --- a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-de-16k-common-vocab3690-tensorflow1-online/infer.py +++ b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-de-16k-common-vocab3690-tensorflow1-online/infer.py @@ -9,5 +9,5 @@ if __name__ == "__main__": model="damo/speech_UniASR_asr_2pass-de-16k-common-vocab3690-tensorflow1-online", output_dir=output_dir, ) - rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"online"}) + rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"normal"}) print(rec_result) diff --git a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-en-16k-common-vocab1080-tensorflow1-online/infer.py b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-en-16k-common-vocab1080-tensorflow1-online/infer.py index 195112ae8..57a3afdf9 100644 --- a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-en-16k-common-vocab1080-tensorflow1-online/infer.py +++ b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-en-16k-common-vocab1080-tensorflow1-online/infer.py @@ -9,5 +9,5 @@ if __name__ == "__main__": model="damo/speech_UniASR_asr_2pass-en-16k-common-vocab1080-tensorflow1-online", output_dir=output_dir, ) - rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"online"}) + rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"normal"}) print(rec_result) diff --git a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-es-16k-common-vocab3445-tensorflow1-online/infer.py b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-es-16k-common-vocab3445-tensorflow1-online/infer.py index 528d2bca6..2ec59402c 100644 --- a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-es-16k-common-vocab3445-tensorflow1-online/infer.py +++ b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-es-16k-common-vocab3445-tensorflow1-online/infer.py @@ -9,5 +9,5 @@ if __name__ == "__main__": model="damo/speech_UniASR_asr_2pass-es-16k-common-vocab3445-tensorflow1-online", output_dir=output_dir, ) - rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"online"}) + rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"normal"}) print(rec_result) diff --git a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-fa-16k-common-vocab1257-pytorch-online/infer.py b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-fa-16k-common-vocab1257-pytorch-online/infer.py index 40417ebd3..055e4ebdb 100644 --- a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-fa-16k-common-vocab1257-pytorch-online/infer.py +++ b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-fa-16k-common-vocab1257-pytorch-online/infer.py @@ -23,7 +23,7 @@ def modelscope_infer_core(output_dir, split_dir, njob, idx): batch_size=1 ) audio_in = os.path.join(split_dir, "wav.{}.scp".format(idx)) - inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"online"}) + inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"normal"}) def modelscope_infer(params): diff --git a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-fr-16k-common-vocab3472-tensorflow1-online/infer.py b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-fr-16k-common-vocab3472-tensorflow1-online/infer.py index caa0722b1..2f3e8330c 100644 --- a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-fr-16k-common-vocab3472-tensorflow1-online/infer.py +++ b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-fr-16k-common-vocab3472-tensorflow1-online/infer.py @@ -9,5 +9,5 @@ if __name__ == "__main__": model="damo/speech_UniASR_asr_2pass-fr-16k-common-vocab3472-tensorflow1-online", output_dir=output_dir, ) - rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"online"}) + rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"normal"}) print(rec_result) diff --git a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-id-16k-common-vocab1067-tensorflow1-online/infer.py b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-id-16k-common-vocab1067-tensorflow1-online/infer.py index 5da14a86f..ad2671a3e 100644 --- a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-id-16k-common-vocab1067-tensorflow1-online/infer.py +++ b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-id-16k-common-vocab1067-tensorflow1-online/infer.py @@ -9,5 +9,5 @@ if __name__ == "__main__": model="damo/speech_UniASR_asr_2pass-id-16k-common-vocab1067-tensorflow1-online", output_dir=output_dir, ) - rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"online"}) + rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"normal"}) print(rec_result) diff --git a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-ja-16k-common-vocab93-tensorflow1-online/infer.py b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-ja-16k-common-vocab93-tensorflow1-online/infer.py index 97446e390..f15bc2d2b 100644 --- a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-ja-16k-common-vocab93-tensorflow1-online/infer.py +++ b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-ja-16k-common-vocab93-tensorflow1-online/infer.py @@ -9,5 +9,5 @@ if __name__ == "__main__": model="damo/speech_UniASR_asr_2pass-ja-16k-common-vocab93-tensorflow1-online", output_dir=output_dir, ) - rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"online"}) + rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"normal"}) print(rec_result) diff --git a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-ko-16k-common-vocab6400-tensorflow1-online/infer.py b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-ko-16k-common-vocab6400-tensorflow1-online/infer.py index 0bfe50836..135e8f8b9 100644 --- a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-ko-16k-common-vocab6400-tensorflow1-online/infer.py +++ b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-ko-16k-common-vocab6400-tensorflow1-online/infer.py @@ -9,5 +9,5 @@ if __name__ == "__main__": model="damo/speech_UniASR_asr_2pass-ko-16k-common-vocab6400-tensorflow1-online", output_dir=output_dir, ) - rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"online"}) + rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"normal"}) print(rec_result) diff --git a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-pt-16k-common-vocab1617-tensorflow1-online/infer.py b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-pt-16k-common-vocab1617-tensorflow1-online/infer.py index 56a605638..aff2a9a51 100644 --- a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-pt-16k-common-vocab1617-tensorflow1-online/infer.py +++ b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-pt-16k-common-vocab1617-tensorflow1-online/infer.py @@ -9,5 +9,5 @@ if __name__ == "__main__": model="damo/speech_UniASR_asr_2pass-pt-16k-common-vocab1617-tensorflow1-online", output_dir=output_dir, ) - rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"online"}) + rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"normal"}) print(rec_result) diff --git a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-ru-16k-common-vocab1664-tensorflow1-online/infer.py b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-ru-16k-common-vocab1664-tensorflow1-online/infer.py index 7e113108d..88c06b4c6 100644 --- a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-ru-16k-common-vocab1664-tensorflow1-online/infer.py +++ b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-ru-16k-common-vocab1664-tensorflow1-online/infer.py @@ -9,5 +9,5 @@ if __name__ == "__main__": model="damo/speech_UniASR_asr_2pass-ru-16k-common-vocab1664-tensorflow1-online", output_dir=output_dir, ) - rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"online"}) + rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"normal"}) print(rec_result) diff --git a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-vi-16k-common-vocab1001-pytorch-online/infer.py b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-vi-16k-common-vocab1001-pytorch-online/infer.py index 15588e74c..4a844fc82 100644 --- a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-vi-16k-common-vocab1001-pytorch-online/infer.py +++ b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-vi-16k-common-vocab1001-pytorch-online/infer.py @@ -9,5 +9,5 @@ if __name__ == "__main__": model="damo/speech_UniASR_asr_2pass-vi-16k-common-vocab1001-pytorch-online", output_dir=output_dir, ) - rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"online"}) + rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"normal"}) print(rec_result)