diff --git a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-cantonese-CHS-16k-common-vocab1468-tensorflow1-offline/infer.py b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-cantonese-CHS-16k-common-vocab1468-tensorflow1-offline/infer.py index c06f934e4..c15114934 100644 --- a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-cantonese-CHS-16k-common-vocab1468-tensorflow1-offline/infer.py +++ b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-cantonese-CHS-16k-common-vocab1468-tensorflow1-offline/infer.py @@ -9,5 +9,5 @@ if __name__ == "__main__": model="damo/speech_UniASR_asr_2pass-cantonese-CHS-16k-common-vocab1468-tensorflow1-online", output_dir=output_dir, ) - rec_result = inference_pipline(audio_in=audio_in) + rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"offline"}) print(rec_result) diff --git a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-cantonese-CHS-16k-common-vocab1468-tensorflow1-online/infer.py b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-cantonese-CHS-16k-common-vocab1468-tensorflow1-online/infer.py index c06f934e4..ac73adf72 100644 --- a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-cantonese-CHS-16k-common-vocab1468-tensorflow1-online/infer.py +++ b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-cantonese-CHS-16k-common-vocab1468-tensorflow1-online/infer.py @@ -9,5 +9,5 @@ if __name__ == "__main__": model="damo/speech_UniASR_asr_2pass-cantonese-CHS-16k-common-vocab1468-tensorflow1-online", output_dir=output_dir, ) - rec_result = inference_pipline(audio_in=audio_in) + rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"normal"}) print(rec_result) diff --git a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-de-16k-common-vocab3690-tensorflow1-offline/infer.py b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-de-16k-common-vocab3690-tensorflow1-offline/infer.py index d23c7f4a3..5ace7e4cf 100644 --- a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-de-16k-common-vocab3690-tensorflow1-offline/infer.py +++ b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-de-16k-common-vocab3690-tensorflow1-offline/infer.py @@ -9,5 +9,5 @@ if __name__ == "__main__": model="damo/speech_UniASR_asr_2pass-de-16k-common-vocab3690-tensorflow1-offline", output_dir=output_dir, ) - rec_result = inference_pipline(audio_in=audio_in) + rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"offline"}) print(rec_result) diff --git a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-de-16k-common-vocab3690-tensorflow1-online/infer.py b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-de-16k-common-vocab3690-tensorflow1-online/infer.py index d7840c29f..f8d91b833 100644 --- a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-de-16k-common-vocab3690-tensorflow1-online/infer.py +++ b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-de-16k-common-vocab3690-tensorflow1-online/infer.py @@ -9,5 +9,5 @@ if __name__ == "__main__": model="damo/speech_UniASR_asr_2pass-de-16k-common-vocab3690-tensorflow1-online", output_dir=output_dir, ) - rec_result = inference_pipline(audio_in=audio_in) + rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"normal"}) print(rec_result) diff --git a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-en-16k-common-vocab1080-tensorflow1-offline/infer.py b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-en-16k-common-vocab1080-tensorflow1-offline/infer.py index 0cfe93995..49b884b2f 100644 --- a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-en-16k-common-vocab1080-tensorflow1-offline/infer.py +++ b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-en-16k-common-vocab1080-tensorflow1-offline/infer.py @@ -9,5 +9,5 @@ if __name__ == "__main__": model="damo/speech_UniASR_asr_2pass-en-16k-common-vocab1080-tensorflow1-offline", output_dir=output_dir, ) - rec_result = inference_pipline(audio_in=audio_in) + rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"offline"}) print(rec_result) diff --git a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-en-16k-common-vocab1080-tensorflow1-online/infer.py b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-en-16k-common-vocab1080-tensorflow1-online/infer.py index 201f7942e..57a3afdf9 100644 --- a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-en-16k-common-vocab1080-tensorflow1-online/infer.py +++ b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-en-16k-common-vocab1080-tensorflow1-online/infer.py @@ -9,5 +9,5 @@ if __name__ == "__main__": model="damo/speech_UniASR_asr_2pass-en-16k-common-vocab1080-tensorflow1-online", output_dir=output_dir, ) - rec_result = inference_pipline(audio_in=audio_in) + rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"normal"}) print(rec_result) diff --git a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-es-16k-common-vocab3445-tensorflow1-offline/infer.py b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-es-16k-common-vocab3445-tensorflow1-offline/infer.py index 6c416b2ae..510f00828 100644 --- a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-es-16k-common-vocab3445-tensorflow1-offline/infer.py +++ b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-es-16k-common-vocab3445-tensorflow1-offline/infer.py @@ -9,5 +9,5 @@ if __name__ == "__main__": model="damo/speech_UniASR_asr_2pass-es-16k-common-vocab3445-tensorflow1-offline", output_dir=output_dir, ) - rec_result = inference_pipline(audio_in=audio_in) + rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"offline"}) print(rec_result) diff --git a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-es-16k-common-vocab3445-tensorflow1-online/infer.py b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-es-16k-common-vocab3445-tensorflow1-online/infer.py index cfd9e9d9c..2ec59402c 100644 --- a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-es-16k-common-vocab3445-tensorflow1-online/infer.py +++ b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-es-16k-common-vocab3445-tensorflow1-online/infer.py @@ -9,5 +9,5 @@ if __name__ == "__main__": model="damo/speech_UniASR_asr_2pass-es-16k-common-vocab3445-tensorflow1-online", output_dir=output_dir, ) - rec_result = inference_pipline(audio_in=audio_in) + rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"normal"}) print(rec_result) diff --git a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-fa-16k-common-vocab1257-pytorch-offline/infer.py b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-fa-16k-common-vocab1257-pytorch-offline/infer.py index 3a8954640..040265d22 100644 --- a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-fa-16k-common-vocab1257-pytorch-offline/infer.py +++ b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-fa-16k-common-vocab1257-pytorch-offline/infer.py @@ -23,7 +23,7 @@ def modelscope_infer_core(output_dir, split_dir, njob, idx): batch_size=1 ) audio_in = os.path.join(split_dir, "wav.{}.scp".format(idx)) - inference_pipline(audio_in=audio_in) + inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"offline"}) def modelscope_infer(params): diff --git a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-fa-16k-common-vocab1257-pytorch-online/infer.py b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-fa-16k-common-vocab1257-pytorch-online/infer.py index ecb138181..055e4ebdb 100644 --- a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-fa-16k-common-vocab1257-pytorch-online/infer.py +++ b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-fa-16k-common-vocab1257-pytorch-online/infer.py @@ -23,7 +23,7 @@ def modelscope_infer_core(output_dir, split_dir, njob, idx): batch_size=1 ) audio_in = os.path.join(split_dir, "wav.{}.scp".format(idx)) - inference_pipline(audio_in=audio_in) + inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"normal"}) def modelscope_infer(params): diff --git a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-fr-16k-common-vocab3472-tensorflow1-offline/infer.py b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-fr-16k-common-vocab3472-tensorflow1-offline/infer.py index e541f2709..6aedeeaa8 100644 --- a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-fr-16k-common-vocab3472-tensorflow1-offline/infer.py +++ b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-fr-16k-common-vocab3472-tensorflow1-offline/infer.py @@ -9,5 +9,5 @@ if __name__ == "__main__": model="damo/speech_UniASR_asr_2pass-fr-16k-common-vocab3472-tensorflow1-offline", output_dir=output_dir, ) - rec_result = inference_pipline(audio_in=audio_in) + rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"offline"}) print(rec_result) diff --git a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-fr-16k-common-vocab3472-tensorflow1-online/infer.py b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-fr-16k-common-vocab3472-tensorflow1-online/infer.py index f87166510..2f3e8330c 100644 --- a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-fr-16k-common-vocab3472-tensorflow1-online/infer.py +++ b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-fr-16k-common-vocab3472-tensorflow1-online/infer.py @@ -9,5 +9,5 @@ if __name__ == "__main__": model="damo/speech_UniASR_asr_2pass-fr-16k-common-vocab3472-tensorflow1-online", output_dir=output_dir, ) - rec_result = inference_pipline(audio_in=audio_in) + rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"normal"}) print(rec_result) diff --git a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-id-16k-common-vocab1067-tensorflow1-offline/infer.py b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-id-16k-common-vocab1067-tensorflow1-offline/infer.py index 82d19d0c5..219c9ec42 100644 --- a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-id-16k-common-vocab1067-tensorflow1-offline/infer.py +++ b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-id-16k-common-vocab1067-tensorflow1-offline/infer.py @@ -9,5 +9,5 @@ if __name__ == "__main__": model="damo/speech_UniASR_asr_2pass-id-16k-common-vocab1067-tensorflow1-online", output_dir=output_dir, ) - rec_result = inference_pipline(audio_in=audio_in) + rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"offline"}) print(rec_result) diff --git a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-id-16k-common-vocab1067-tensorflow1-online/infer.py b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-id-16k-common-vocab1067-tensorflow1-online/infer.py index 82d19d0c5..ad2671a3e 100644 --- a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-id-16k-common-vocab1067-tensorflow1-online/infer.py +++ b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-id-16k-common-vocab1067-tensorflow1-online/infer.py @@ -9,5 +9,5 @@ if __name__ == "__main__": model="damo/speech_UniASR_asr_2pass-id-16k-common-vocab1067-tensorflow1-online", output_dir=output_dir, ) - rec_result = inference_pipline(audio_in=audio_in) + rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"normal"}) print(rec_result) diff --git a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-ja-16k-common-vocab93-tensorflow1-offline/infer.py b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-ja-16k-common-vocab93-tensorflow1-offline/infer.py index f1e156bb3..a053957d3 100644 --- a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-ja-16k-common-vocab93-tensorflow1-offline/infer.py +++ b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-ja-16k-common-vocab93-tensorflow1-offline/infer.py @@ -9,5 +9,5 @@ if __name__ == "__main__": model="damo/speech_UniASR_asr_2pass-ja-16k-common-vocab93-tensorflow1-online", output_dir=output_dir, ) - rec_result = inference_pipline(audio_in=audio_in) + rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"offline"}) print(rec_result) diff --git a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-ja-16k-common-vocab93-tensorflow1-online/infer.py b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-ja-16k-common-vocab93-tensorflow1-online/infer.py index f1e156bb3..f15bc2d2b 100644 --- a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-ja-16k-common-vocab93-tensorflow1-online/infer.py +++ b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-ja-16k-common-vocab93-tensorflow1-online/infer.py @@ -9,5 +9,5 @@ if __name__ == "__main__": model="damo/speech_UniASR_asr_2pass-ja-16k-common-vocab93-tensorflow1-online", output_dir=output_dir, ) - rec_result = inference_pipline(audio_in=audio_in) + rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"normal"}) print(rec_result) diff --git a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-ko-16k-common-vocab6400-tensorflow1-offline/infer.py b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-ko-16k-common-vocab6400-tensorflow1-offline/infer.py index 7aba7ee5b..618b3f601 100644 --- a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-ko-16k-common-vocab6400-tensorflow1-offline/infer.py +++ b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-ko-16k-common-vocab6400-tensorflow1-offline/infer.py @@ -9,5 +9,5 @@ if __name__ == "__main__": model="damo/speech_UniASR_asr_2pass-ko-16k-common-vocab6400-tensorflow1-offline", output_dir=output_dir, ) - rec_result = inference_pipline(audio_in=audio_in) + rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"offline"}) print(rec_result) diff --git a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-ko-16k-common-vocab6400-tensorflow1-online/infer.py b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-ko-16k-common-vocab6400-tensorflow1-online/infer.py index 5fbfdbbb3..135e8f8b9 100644 --- a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-ko-16k-common-vocab6400-tensorflow1-online/infer.py +++ b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-ko-16k-common-vocab6400-tensorflow1-online/infer.py @@ -9,5 +9,5 @@ if __name__ == "__main__": model="damo/speech_UniASR_asr_2pass-ko-16k-common-vocab6400-tensorflow1-online", output_dir=output_dir, ) - rec_result = inference_pipline(audio_in=audio_in) + rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"normal"}) print(rec_result) diff --git a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-pt-16k-common-vocab1617-tensorflow1-offline/infer.py b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-pt-16k-common-vocab1617-tensorflow1-offline/infer.py index fdab5a36b..30a11ffd3 100644 --- a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-pt-16k-common-vocab1617-tensorflow1-offline/infer.py +++ b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-pt-16k-common-vocab1617-tensorflow1-offline/infer.py @@ -9,5 +9,5 @@ if __name__ == "__main__": model="damo/speech_UniASR_asr_2pass-pt-16k-common-vocab1617-tensorflow1-online", output_dir=output_dir, ) - rec_result = inference_pipline(audio_in=audio_in) + rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"offline"}) print(rec_result) diff --git a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-pt-16k-common-vocab1617-tensorflow1-online/infer.py b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-pt-16k-common-vocab1617-tensorflow1-online/infer.py index fdab5a36b..aff2a9a51 100644 --- a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-pt-16k-common-vocab1617-tensorflow1-online/infer.py +++ b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-pt-16k-common-vocab1617-tensorflow1-online/infer.py @@ -9,5 +9,5 @@ if __name__ == "__main__": model="damo/speech_UniASR_asr_2pass-pt-16k-common-vocab1617-tensorflow1-online", output_dir=output_dir, ) - rec_result = inference_pipline(audio_in=audio_in) + rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"normal"}) print(rec_result) diff --git a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-ru-16k-common-vocab1664-tensorflow1-offline/infer.py b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-ru-16k-common-vocab1664-tensorflow1-offline/infer.py index 8b96ffb18..95f447d13 100644 --- a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-ru-16k-common-vocab1664-tensorflow1-offline/infer.py +++ b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-ru-16k-common-vocab1664-tensorflow1-offline/infer.py @@ -9,5 +9,5 @@ if __name__ == "__main__": model="damo/speech_UniASR_asr_2pass-ru-16k-common-vocab1664-tensorflow1-offline", output_dir=output_dir, ) - rec_result = inference_pipline(audio_in=audio_in) + rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"offline"}) print(rec_result) diff --git a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-ru-16k-common-vocab1664-tensorflow1-online/infer.py b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-ru-16k-common-vocab1664-tensorflow1-online/infer.py index 1051b1fb6..88c06b4c6 100644 --- a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-ru-16k-common-vocab1664-tensorflow1-online/infer.py +++ b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-ru-16k-common-vocab1664-tensorflow1-online/infer.py @@ -9,5 +9,5 @@ if __name__ == "__main__": model="damo/speech_UniASR_asr_2pass-ru-16k-common-vocab1664-tensorflow1-online", output_dir=output_dir, ) - rec_result = inference_pipline(audio_in=audio_in) + rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"normal"}) print(rec_result) diff --git a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-vi-16k-common-vocab1001-pytorch-offline/infer.py b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-vi-16k-common-vocab1001-pytorch-offline/infer.py index b7fcd5933..9472104e5 100644 --- a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-vi-16k-common-vocab1001-pytorch-offline/infer.py +++ b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-vi-16k-common-vocab1001-pytorch-offline/infer.py @@ -9,5 +9,5 @@ if __name__ == "__main__": model="damo/speech_UniASR_asr_2pass-vi-16k-common-vocab1001-pytorch-offline", output_dir=output_dir, ) - rec_result = inference_pipline(audio_in=audio_in) + rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"offline"}) print(rec_result) diff --git a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-vi-16k-common-vocab1001-pytorch-online/infer.py b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-vi-16k-common-vocab1001-pytorch-online/infer.py index 869082b07..4a844fc82 100644 --- a/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-vi-16k-common-vocab1001-pytorch-online/infer.py +++ b/egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-vi-16k-common-vocab1001-pytorch-online/infer.py @@ -9,5 +9,5 @@ if __name__ == "__main__": model="damo/speech_UniASR_asr_2pass-vi-16k-common-vocab1001-pytorch-online", output_dir=output_dir, ) - rec_result = inference_pipline(audio_in=audio_in) + rec_result = inference_pipline(audio_in=audio_in, param_dict={"decoding_model":"normal"}) print(rec_result)