mirror of
https://github.com/modelscope/FunASR
synced 2025-09-15 14:48:36 +08:00
rtf benchmark
This commit is contained in:
parent
ed3de87cd9
commit
50571828c6
@ -17,7 +17,7 @@ The installation is the same as [funasr](../../README.md)
|
|||||||
--model-name [model_name] \
|
--model-name [model_name] \
|
||||||
--export-dir [export_dir] \
|
--export-dir [export_dir] \
|
||||||
--type [onnx, torch] \
|
--type [onnx, torch] \
|
||||||
--quantize \
|
--quantize [true, false] \
|
||||||
--fallback-num [fallback_num]
|
--fallback-num [fallback_num]
|
||||||
```
|
```
|
||||||
`model-name`: the model is to export. It could be the models from modelscope, or local finetuned model(named: model.pb).
|
`model-name`: the model is to export. It could be the models from modelscope, or local finetuned model(named: model.pb).
|
||||||
|
|||||||
@ -10,7 +10,7 @@ import torch
|
|||||||
from funasr.export.models import get_model
|
from funasr.export.models import get_model
|
||||||
import numpy as np
|
import numpy as np
|
||||||
import random
|
import random
|
||||||
|
from funasr.utils.types import str2bool
|
||||||
# torch_version = float(".".join(torch.__version__.split(".")[:2]))
|
# torch_version = float(".".join(torch.__version__.split(".")[:2]))
|
||||||
# assert torch_version > 1.9
|
# assert torch_version > 1.9
|
||||||
|
|
||||||
@ -234,7 +234,7 @@ if __name__ == '__main__':
|
|||||||
parser.add_argument('--model-name', type=str, required=True)
|
parser.add_argument('--model-name', type=str, required=True)
|
||||||
parser.add_argument('--export-dir', type=str, required=True)
|
parser.add_argument('--export-dir', type=str, required=True)
|
||||||
parser.add_argument('--type', type=str, default='onnx', help='["onnx", "torch"]')
|
parser.add_argument('--type', type=str, default='onnx', help='["onnx", "torch"]')
|
||||||
parser.add_argument('--quantize', action='store_true', help='export quantized model')
|
parser.add_argument('--quantize', type=str2bool, default=False, help='export quantized model')
|
||||||
parser.add_argument('--fallback-num', type=int, default=0, help='amp fallback number')
|
parser.add_argument('--fallback-num', type=int, default=0, help='amp fallback number')
|
||||||
parser.add_argument('--audio_in', type=str, default=None, help='["wav", "wav.scp"]')
|
parser.add_argument('--audio_in', type=str, default=None, help='["wav", "wav.scp"]')
|
||||||
parser.add_argument('--calib_num', type=int, default=200, help='calib max num')
|
parser.add_argument('--calib_num', type=int, default=200, help='calib max num')
|
||||||
|
|||||||
@ -2,14 +2,14 @@
|
|||||||
import time
|
import time
|
||||||
import sys
|
import sys
|
||||||
import librosa
|
import librosa
|
||||||
|
from funasr.utils.types import str2bool
|
||||||
|
|
||||||
import argparse
|
import argparse
|
||||||
parser = argparse.ArgumentParser()
|
parser = argparse.ArgumentParser()
|
||||||
parser.add_argument('--model_dir', type=str, required=True)
|
parser.add_argument('--model_dir', type=str, required=True)
|
||||||
parser.add_argument('--backend', type=str, default='onnx', help='["onnx", "torch"]')
|
parser.add_argument('--backend', type=str, default='onnx', help='["onnx", "torch"]')
|
||||||
parser.add_argument('--wav_file', type=str, default=None, help='amp fallback number')
|
parser.add_argument('--wav_file', type=str, default=None, help='amp fallback number')
|
||||||
parser.add_argument('--quantize', type=bool, default=False, help='quantized model')
|
parser.add_argument('--quantize', type=str2bool, default=False, help='quantized model')
|
||||||
parser.add_argument('--intra_op_num_threads', type=int, default=1, help='intra_op_num_threads for onnx')
|
parser.add_argument('--intra_op_num_threads', type=int, default=1, help='intra_op_num_threads for onnx')
|
||||||
args = parser.parse_args()
|
args = parser.parse_args()
|
||||||
|
|
||||||
|
|||||||
@ -10,7 +10,7 @@ rtf_tool=test_rtf.py
|
|||||||
#:<<!
|
#:<<!
|
||||||
model_name="damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch"
|
model_name="damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch"
|
||||||
backend="onnx" # "torch"
|
backend="onnx" # "torch"
|
||||||
quantize='True' # 'False'
|
quantize='true' # 'False'
|
||||||
tag=${model_name}/${backend}_${quantize}
|
tag=${model_name}/${backend}_${quantize}
|
||||||
!
|
!
|
||||||
|
|
||||||
@ -21,11 +21,7 @@ echo ${logs_outputs_dir}
|
|||||||
|
|
||||||
if [ $stage == 0 ];then
|
if [ $stage == 0 ];then
|
||||||
|
|
||||||
if [ $quantize == 'True' ];then
|
|
||||||
python -m funasr.export.export_model --model-name ${model_name} --export-dir ${export_root} --type ${backend} --quantize --audio_in ${scp}
|
python -m funasr.export.export_model --model-name ${model_name} --export-dir ${export_root} --type ${backend} --quantize --audio_in ${scp}
|
||||||
else
|
|
||||||
python -m funasr.export.export_model --model-name ${model_name} --export-dir ${export_root} --type ${backend}
|
|
||||||
fi
|
|
||||||
|
|
||||||
fi
|
fi
|
||||||
|
|
||||||
|
|||||||
0
push_scan.py
Normal file
0
push_scan.py
Normal file
Loading…
Reference in New Issue
Block a user