mirror of
https://github.com/modelscope/FunASR
synced 2025-09-15 14:48:36 +08:00
Merge branch 'dev_infer' of https://github.com/alibaba/FunASR into dev_infer
This commit is contained in:
commit
94a4dbba3d
@ -1,7 +1,3 @@
|
||||
|
||||
##################text二进制数据#####################
|
||||
inputs = "跨境河流是养育沿岸|人民的生命之源长期以来为帮助下游地区防灾减灾中方技术人员|在上游地区极为恶劣的自然条件下克服巨大困难甚至冒着生命危险|向印方提供汛期水文资料处理紧急事件中方重视印方在跨境河流问题上的关切|愿意进一步完善双方联合工作机制|凡是|中方能做的我们|都会去做而且会做得更好我请印度朋友们放心中国在上游的|任何开发利用都会经过科学|规划和论证兼顾上下游的利益"
|
||||
|
||||
from modelscope.pipelines import pipeline
|
||||
from modelscope.utils.constant import Tasks
|
||||
from modelscope.utils.logger import get_logger
|
||||
@ -16,6 +12,9 @@ inference_pipeline = pipeline(
|
||||
output_dir="./tmp/"
|
||||
)
|
||||
|
||||
##################text二进制数据#####################
|
||||
inputs = "跨境河流是养育沿岸|人民的生命之源长期以来为帮助下游地区防灾减灾中方技术人员|在上游地区极为恶劣的自然条件下克服巨大困难甚至冒着生命危险|向印方提供汛期水文资料处理紧急事件中方重视印方在跨境河流问题上的关切|愿意进一步完善双方联合工作机制|凡是|中方能做的我们|都会去做而且会做得更好我请印度朋友们放心中国在上游的|任何开发利用都会经过科学|规划和论证兼顾上下游的利益"
|
||||
|
||||
vads = inputs.split("|")
|
||||
rec_result_all="outputs:"
|
||||
param_dict = {"cache": []}
|
||||
|
||||
@ -1,14 +1,4 @@
|
||||
|
||||
##################text.scp###################
|
||||
# inputs = "./egs_modelscope/punctuation/punc_ct-transformer_zh-cn-common-vocab272727-pytorch/data/punc_example.txt"
|
||||
|
||||
##################text#####################
|
||||
#inputs = "我们都是木头人不会讲话不会动"
|
||||
|
||||
##################text file url#######################
|
||||
inputs = "https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_text/punc_example.txt"
|
||||
|
||||
|
||||
from modelscope.pipelines import pipeline
|
||||
from modelscope.utils.constant import Tasks
|
||||
|
||||
@ -19,5 +9,14 @@ inference_pipeline = pipeline(
|
||||
output_dir="./tmp/"
|
||||
)
|
||||
|
||||
##################text.scp###################
|
||||
# inputs = "./egs_modelscope/punctuation/punc_ct-transformer_zh-cn-common-vocab272727-pytorch/data/punc_example.txt"
|
||||
|
||||
##################text#####################
|
||||
#inputs = "我们都是木头人不会讲话不会动"
|
||||
|
||||
##################text file url#######################
|
||||
inputs = "https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_text/punc_example.txt"
|
||||
|
||||
rec_result = inference_pipeline(text_in=inputs)
|
||||
print(rec_result)
|
||||
|
||||
@ -8,6 +8,7 @@ def modelscope_infer(args):
|
||||
inference_pipeline = pipeline(
|
||||
task=Tasks.speech_timestamp,
|
||||
model=args.model,
|
||||
model_revision='v1.1.0',
|
||||
output_dir=args.output_dir,
|
||||
batch_size=args.batch_size,
|
||||
)
|
||||
@ -21,7 +22,7 @@ if __name__ == "__main__":
|
||||
parser.add_argument('--model', type=str, default="damo/speech_timestamp_prediction-v1-16k-offline")
|
||||
parser.add_argument('--audio_in', type=str, default="https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_timestamps.wav")
|
||||
parser.add_argument('--text_in', type=str, default="一 个 东 太 平 洋 国 家 为 什 么 跑 到 西 太 平 洋 来 了 呢")
|
||||
parser.add_argument('--output_dir', type=str, default="./results/")
|
||||
parser.add_argument('--output_dir', type=str, default=None)
|
||||
parser.add_argument('--batch_size', type=int, default=1)
|
||||
parser.add_argument('--gpuid', type=str, default="0")
|
||||
args = parser.parse_args()
|
||||
|
||||
@ -4,6 +4,7 @@ from modelscope.utils.constant import Tasks
|
||||
inference_pipeline = pipeline(
|
||||
task=Tasks.speech_timestamp,
|
||||
model='damo/speech_timestamp_prediction-v1-16k-offline',
|
||||
model_revision='v1.1.0',
|
||||
output_dir=None)
|
||||
|
||||
rec_result = inference_pipeline(
|
||||
|
||||
@ -1526,6 +1526,12 @@ def get_parser():
|
||||
action="append",
|
||||
)
|
||||
group.add_argument("--key_file", type=str_or_none)
|
||||
parser.add_argument(
|
||||
"--hotword",
|
||||
type=str_or_none,
|
||||
default=None,
|
||||
help="hotword file path or hotwords seperated by space"
|
||||
)
|
||||
group.add_argument("--allow_variable_data_keys", type=str2bool, default=False)
|
||||
group.add_argument(
|
||||
"--mc",
|
||||
|
||||
Loading…
Reference in New Issue
Block a user