mirror of
https://github.com/modelscope/FunASR
synced 2025-09-15 14:48:36 +08:00
* contextual&seaco ONNX export * update ContextualEmbedderExport2 * update ContextualEmbedderExport2 * update code * onnx (#1482) * qwenaudio qwenaudiochat * qwenaudio qwenaudiochat * whisper * whisper * llm * llm * llm * llm * llm * llm * llm * llm * export onnx * export onnx * export onnx * dingding * dingding * llm * doc * onnx * onnx * onnx * onnx * onnx * onnx * v1.0.15 * qwenaudio * qwenaudio * issue doc * update * update * bugfix * onnx * update export calling * update codes * remove useless code * update code --------- Co-authored-by: zhifu gao <zhifu.gzf@alibaba-inc.com>
16 lines
719 B
Python
16 lines
719 B
Python
from funasr_onnx import Paraformer
|
|
from pathlib import Path
|
|
|
|
model_dir = "damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch"
|
|
# model_dir = "damo/speech_paraformer-large-vad-punc_asr_nat-zh-cn-16k-common-vocab8404-pytorch"
|
|
model = Paraformer(model_dir, batch_size=1, quantize=False)
|
|
# model = Paraformer(model_dir, batch_size=1, device_id=0) # gpu
|
|
|
|
# when using paraformer-large-vad-punc model, you can set plot_timestamp_to="./xx.png" to get figure of alignment besides timestamps
|
|
# model = Paraformer(model_dir, batch_size=1, plot_timestamp_to="test.png")
|
|
|
|
wav_path = ['{}/.cache/modelscope/hub/{}/example/asr_example.wav'.format(Path.home(), model_dir)]
|
|
|
|
result = model(wav_path)
|
|
print(result)
|