mirror of
https://github.com/modelscope/FunASR
synced 2025-09-15 14:48:36 +08:00
commit
b2acffedd4
@ -1,6 +1,6 @@
|
||||
# ModelScope Model
|
||||
|
||||
## How to finetune and infer using a pretrained Paraformer-large Model
|
||||
## How to finetune and infer using a pretrained UniASR Model
|
||||
|
||||
### Finetune
|
||||
|
||||
|
||||
@ -23,7 +23,7 @@ def modelscope_infer_core(output_dir, split_dir, njob, idx):
|
||||
batch_size=1
|
||||
)
|
||||
audio_in = os.path.join(split_dir, "wav.{}.scp".format(idx))
|
||||
inference_pipline(audio_in=audio_in)
|
||||
inference_pipline(audio_in=audio_in, param_dict={"decoding_model": "normal"})
|
||||
|
||||
|
||||
def modelscope_infer(params):
|
||||
|
||||
@ -34,7 +34,7 @@ def modelscope_infer_after_finetune(params):
|
||||
batch_size=1
|
||||
)
|
||||
audio_in = os.path.join(params["data_dir"], "wav.scp")
|
||||
inference_pipeline(audio_in=audio_in)
|
||||
inference_pipeline(audio_in=audio_in, param_dict={"decoding_model": "normal"})
|
||||
|
||||
# computer CER if GT text is set
|
||||
text_in = os.path.join(params["data_dir"], "text")
|
||||
|
||||
@ -23,7 +23,7 @@ def modelscope_infer_core(output_dir, split_dir, njob, idx):
|
||||
batch_size=1
|
||||
)
|
||||
audio_in = os.path.join(split_dir, "wav.{}.scp".format(idx))
|
||||
inference_pipline(audio_in=audio_in)
|
||||
inference_pipline(audio_in=audio_in, param_dict={"decoding_model": "offline"})
|
||||
|
||||
|
||||
def modelscope_infer(params):
|
||||
|
||||
@ -34,7 +34,7 @@ def modelscope_infer_after_finetune(params):
|
||||
batch_size=1
|
||||
)
|
||||
audio_in = os.path.join(params["data_dir"], "wav.scp")
|
||||
inference_pipeline(audio_in=audio_in)
|
||||
inference_pipeline(audio_in=audio_in, param_dict={"decoding_model": "offline"})
|
||||
|
||||
# computer CER if GT text is set
|
||||
text_in = os.path.join(params["data_dir"], "text")
|
||||
|
||||
@ -23,7 +23,7 @@ def modelscope_infer_core(output_dir, split_dir, njob, idx):
|
||||
batch_size=1
|
||||
)
|
||||
audio_in = os.path.join(split_dir, "wav.{}.scp".format(idx))
|
||||
inference_pipline(audio_in=audio_in)
|
||||
inference_pipline(audio_in=audio_in, param_dict={"decoding_model": "normal"})
|
||||
|
||||
|
||||
def modelscope_infer(params):
|
||||
|
||||
@ -34,7 +34,7 @@ def modelscope_infer_after_finetune(params):
|
||||
batch_size=1
|
||||
)
|
||||
audio_in = os.path.join(params["data_dir"], "wav.scp")
|
||||
inference_pipeline(audio_in=audio_in)
|
||||
inference_pipeline(audio_in=audio_in, param_dict={"decoding_model": "normal"})
|
||||
|
||||
# computer CER if GT text is set
|
||||
text_in = os.path.join(params["data_dir"], "text")
|
||||
|
||||
Loading…
Reference in New Issue
Block a user