mirror of
https://github.com/modelscope/FunASR
synced 2025-09-15 14:48:36 +08:00
* update with main (#1817) * add cmakelist * add paraformer-torch * add debug for funasr-onnx-offline * fix redefinition of jieba StdExtension.hpp * add loading torch models * update funasr-onnx-offline * add SwitchArg for wss-server * add SwitchArg for funasr-onnx-offline * update cmakelist * update funasr-onnx-offline-rtf * add define condition * add gpu define for offlne-stream * update com define * update offline-stream * update cmakelist * update func CompileHotwordEmbedding * add timestamp for paraformer-torch * add C10_USE_GLOG for paraformer-torch * update paraformer-torch * fix func FunASRWfstDecoderInit * update model.h * fix func FunASRWfstDecoderInit * fix tpass_stream * update paraformer-torch * add bladedisc for funasr-onnx-offline * update comdefine * update funasr-wss-server * add log for torch * fix GetValue BLADEDISC * fix log * update cmakelist * update warmup to 10 * update funasrruntime * add batch_size for wss-server * add batch for bins * add batch for offline-stream * add batch for paraformer * add batch for offline-stream * fix func SetBatchSize * add SetBatchSize for model * add SetBatchSize for model * fix func Forward * fix padding * update funasrruntime * add dec reset for batch * set batch default value * add argv for CutSplit * sort frame_queue * sorted msgs * fix FunOfflineInfer * add dynamic batch for fetch * fix FetchDynamic * update run_server.sh * update run_server.sh * cpp http post server support (#1739) * add cpp http server * add some comment * remove some comments * del debug infos * restore run_server.sh * adapt to new model struct * 修复了onnxruntime在macos下编译失败的错误 (#1748) * Add files via upload 增加macos的编译支持 * Add files via upload 增加macos支持 * Add files via upload target_link_directories(funasr PUBLIC ${ONNXRUNTIME_DIR}/lib) target_link_directories(funasr PUBLIC ${FFMPEG_DIR}/lib) 添加 if(APPLE) 限制 --------- Co-authored-by: Yabin Li <wucong.lyb@alibaba-inc.com> * Delete docs/images/wechat.png * Add files via upload * fixed the issues about seaco-onnx timestamp * fix bug (#1764) 当语音识别结果包含 `http` 时,标点符号预测会把它会被当成 url * fix empty asr result (#1765) 解码结果为空的语音片段,text 用空字符串 * update export * update export * docs * docs * update export name * docs * update * docs * docs * keep empty speech result (#1772) * docs * docs * update wechat QRcode * Add python funasr api support for websocket srv (#1777) * add python funasr_api supoort * change little to README.md * add core tools stream * modified a little * fix bug for timeout * support for buffer decode * add ffmpeg decode for buffer * libtorch demo * update libtorch infer * update utils * update demo * update demo * update libtorch inference * update model class * update seaco paraformer * bug fix * bug fix * auto frontend * auto frontend * auto frontend * auto frontend * auto frontend * auto frontend * auto frontend * auto frontend * Dev gzf exp (#1785) * resume from step * batch * batch * batch * batch * batch * batch * batch * batch * batch * batch * batch * batch * batch * batch * batch * train_loss_avg train_acc_avg * train_loss_avg train_acc_avg * train_loss_avg train_acc_avg * log step * wav is not exist * wav is not exist * decoding * decoding * decoding * wechat * decoding key * decoding key * decoding key * decoding key * decoding key * decoding key * dynamic batch * start_data_split_i=0 * total_time/accum_grad * total_time/accum_grad * total_time/accum_grad * update avg slice * update avg slice * sensevoice sanm * sensevoice sanm * sensevoice sanm --------- Co-authored-by: 北念 <lzr265946@alibaba-inc.com> * auto frontend * update paraformer timestamp * [Optimization] support bladedisc fp16 optimization (#1790) * add cif_v1 and cif_export * Update SDK_advanced_guide_offline_zh.md * add cif_wo_hidden_v1 * [fix] fix empty asr result (#1794) * english timestamp for valilla paraformer * wechat * [fix] better solution for handling empty result (#1796) * update scripts * modify the qformer adaptor (#1804) Co-authored-by: nichongjia-2007 <nichongjia@gmail.com> * add ctc inference code (#1806) Co-authored-by: haoneng.lhn <haoneng.lhn@alibaba-inc.com> * Update auto_model.py 修复空字串进入speaker model时报raw_text变量不存在的bug * Update auto_model.py 修复识别出空串后spk_model内变量未定义问题 * update model name * fix paramter 'quantize' unused issue (#1813) Co-authored-by: ZihanLiao <liaozihan1@xdf.cn> * wechat * Update cif_predictor.py (#1811) * Update cif_predictor.py * modify cif_v1_export under extreme cases, max_label_len calculated by batch_len misaligns with token_num * Update cif_predictor.py torch.cumsum precision degradation, using float64 instead * update code --------- Co-authored-by: 雾聪 <wucong.lyb@alibaba-inc.com> Co-authored-by: zhaomingwork <61895407+zhaomingwork@users.noreply.github.com> Co-authored-by: szsteven008 <97944818+szsteven008@users.noreply.github.com> Co-authored-by: Ephemeroptera <605686962@qq.com> Co-authored-by: 彭震东 <zhendong.peng@qq.com> Co-authored-by: Shi Xian <40013335+R1ckShi@users.noreply.github.com> Co-authored-by: 维石 <shixian.shi@alibaba-inc.com> Co-authored-by: 北念 <lzr265946@alibaba-inc.com> Co-authored-by: xiaowan0322 <wanchen.swc@alibaba-inc.com> Co-authored-by: zhuangzhong <zhuangzhong@corp.netease.com> Co-authored-by: Xingchen Song(宋星辰) <xingchensong1996@163.com> Co-authored-by: nichongjia-2007 <nichongjia@gmail.com> Co-authored-by: haoneng.lhn <haoneng.lhn@alibaba-inc.com> Co-authored-by: liugz18 <57401541+liugz18@users.noreply.github.com> Co-authored-by: Marlowe <54339989+ZihanLiao@users.noreply.github.com> Co-authored-by: ZihanLiao <liaozihan1@xdf.cn> Co-authored-by: zhong zhuang <zhuangz@lamda.nju.edu.cn> * sensevoice * sensevoice * sensevoice * sensevoice * sensevoice * sensevoice * sensevoice * sensevoice * sensevoice * sensevoice --------- Co-authored-by: 雾聪 <wucong.lyb@alibaba-inc.com> Co-authored-by: zhaomingwork <61895407+zhaomingwork@users.noreply.github.com> Co-authored-by: szsteven008 <97944818+szsteven008@users.noreply.github.com> Co-authored-by: Ephemeroptera <605686962@qq.com> Co-authored-by: 彭震东 <zhendong.peng@qq.com> Co-authored-by: Shi Xian <40013335+R1ckShi@users.noreply.github.com> Co-authored-by: 维石 <shixian.shi@alibaba-inc.com> Co-authored-by: 北念 <lzr265946@alibaba-inc.com> Co-authored-by: xiaowan0322 <wanchen.swc@alibaba-inc.com> Co-authored-by: zhuangzhong <zhuangzhong@corp.netease.com> Co-authored-by: Xingchen Song(宋星辰) <xingchensong1996@163.com> Co-authored-by: nichongjia-2007 <nichongjia@gmail.com> Co-authored-by: haoneng.lhn <haoneng.lhn@alibaba-inc.com> Co-authored-by: liugz18 <57401541+liugz18@users.noreply.github.com> Co-authored-by: Marlowe <54339989+ZihanLiao@users.noreply.github.com> Co-authored-by: ZihanLiao <liaozihan1@xdf.cn> Co-authored-by: zhong zhuang <zhuangz@lamda.nju.edu.cn>
225 lines
9.4 KiB
Python
225 lines
9.4 KiB
Python
import os
|
|
import json
|
|
from omegaconf import OmegaConf, DictConfig
|
|
|
|
from funasr.download.name_maps_from_hub import name_maps_ms, name_maps_hf, name_maps_openai
|
|
|
|
|
|
def download_model(**kwargs):
|
|
hub = kwargs.get("hub", "ms")
|
|
if hub == "ms":
|
|
kwargs = download_from_ms(**kwargs)
|
|
elif hub == "hf":
|
|
kwargs = download_from_hf(**kwargs)
|
|
elif hub == "openai":
|
|
model_or_path = kwargs.get("model")
|
|
if os.path.exists(model_or_path):
|
|
# local path
|
|
kwargs["model_path"] = model_or_path
|
|
kwargs["model"] = "WhisperWarp"
|
|
else:
|
|
# model name
|
|
if model_or_path in name_maps_openai:
|
|
model_or_path = name_maps_openai[model_or_path]
|
|
kwargs["model_path"] = model_or_path
|
|
|
|
return kwargs
|
|
|
|
|
|
def download_from_ms(**kwargs):
|
|
model_or_path = kwargs.get("model")
|
|
if model_or_path in name_maps_ms:
|
|
model_or_path = name_maps_ms[model_or_path]
|
|
model_revision = kwargs.get("model_revision", "master")
|
|
if not os.path.exists(model_or_path) and "model_path" not in kwargs:
|
|
try:
|
|
model_or_path = get_or_download_model_dir(
|
|
model_or_path,
|
|
model_revision,
|
|
is_training=kwargs.get("is_training"),
|
|
check_latest=kwargs.get("check_latest", True),
|
|
)
|
|
except Exception as e:
|
|
print(f"Download: {model_or_path} failed!: {e}")
|
|
|
|
kwargs["model_path"] = model_or_path if "model_path" not in kwargs else kwargs["model_path"]
|
|
|
|
if os.path.exists(os.path.join(model_or_path, "configuration.json")):
|
|
with open(os.path.join(model_or_path, "configuration.json"), "r", encoding="utf-8") as f:
|
|
conf_json = json.load(f)
|
|
|
|
cfg = {}
|
|
if "file_path_metas" in conf_json:
|
|
add_file_root_path(model_or_path, conf_json["file_path_metas"], cfg)
|
|
cfg.update(kwargs)
|
|
if "config" in cfg:
|
|
config = OmegaConf.load(cfg["config"])
|
|
kwargs = OmegaConf.merge(config, cfg)
|
|
kwargs["model"] = config["model"]
|
|
elif os.path.exists(os.path.join(model_or_path, "config.yaml")):
|
|
config = OmegaConf.load(os.path.join(model_or_path, "config.yaml"))
|
|
kwargs = OmegaConf.merge(config, kwargs)
|
|
init_param = os.path.join(model_or_path, "model.pt")
|
|
if "init_param" not in kwargs or not os.path.exists(kwargs["init_param"]):
|
|
kwargs["init_param"] = init_param
|
|
assert os.path.exists(kwargs["init_param"]), "init_param does not exist"
|
|
if os.path.exists(os.path.join(model_or_path, "tokens.txt")):
|
|
kwargs["tokenizer_conf"]["token_list"] = os.path.join(model_or_path, "tokens.txt")
|
|
if os.path.exists(os.path.join(model_or_path, "tokens.json")):
|
|
kwargs["tokenizer_conf"]["token_list"] = os.path.join(model_or_path, "tokens.json")
|
|
if os.path.exists(os.path.join(model_or_path, "seg_dict")):
|
|
kwargs["tokenizer_conf"]["seg_dict"] = os.path.join(model_or_path, "seg_dict")
|
|
if os.path.exists(os.path.join(model_or_path, "bpe.model")):
|
|
kwargs["tokenizer_conf"]["bpemodel"] = os.path.join(model_or_path, "bpe.model")
|
|
kwargs["model"] = config["model"]
|
|
if os.path.exists(os.path.join(model_or_path, "am.mvn")):
|
|
kwargs["frontend_conf"]["cmvn_file"] = os.path.join(model_or_path, "am.mvn")
|
|
if os.path.exists(os.path.join(model_or_path, "jieba_usr_dict")):
|
|
kwargs["jieba_usr_dict"] = os.path.join(model_or_path, "jieba_usr_dict")
|
|
if isinstance(kwargs, DictConfig):
|
|
kwargs = OmegaConf.to_container(kwargs, resolve=True)
|
|
if os.path.exists(os.path.join(model_or_path, "requirements.txt")):
|
|
requirements = os.path.join(model_or_path, "requirements.txt")
|
|
print(f"Detect model requirements, begin to install it: {requirements}")
|
|
from funasr.utils.install_model_requirements import install_requirements
|
|
|
|
install_requirements(requirements)
|
|
if kwargs.get("trust_remote_code", False):
|
|
|
|
import model
|
|
|
|
# from funasr.register import tables
|
|
# tables.print("model")
|
|
return kwargs
|
|
|
|
|
|
def download_from_hf(**kwargs):
|
|
model_or_path = kwargs.get("model")
|
|
if model_or_path in name_maps_hf:
|
|
model_or_path = name_maps_hf[model_or_path]
|
|
model_revision = kwargs.get("model_revision", "master")
|
|
if not os.path.exists(model_or_path) and "model_path" not in kwargs:
|
|
try:
|
|
model_or_path = get_or_download_model_dir_hf(
|
|
model_or_path,
|
|
model_revision,
|
|
is_training=kwargs.get("is_training"),
|
|
check_latest=kwargs.get("check_latest", True),
|
|
)
|
|
except Exception as e:
|
|
print(f"Download: {model_or_path} failed!: {e}")
|
|
|
|
kwargs["model_path"] = model_or_path if "model_path" not in kwargs else kwargs["model_path"]
|
|
|
|
if os.path.exists(os.path.join(model_or_path, "configuration.json")):
|
|
with open(os.path.join(model_or_path, "configuration.json"), "r", encoding="utf-8") as f:
|
|
conf_json = json.load(f)
|
|
|
|
cfg = {}
|
|
if "file_path_metas" in conf_json:
|
|
add_file_root_path(model_or_path, conf_json["file_path_metas"], cfg)
|
|
cfg.update(kwargs)
|
|
if "config" in cfg:
|
|
config = OmegaConf.load(cfg["config"])
|
|
kwargs = OmegaConf.merge(config, cfg)
|
|
kwargs["model"] = config["model"]
|
|
elif os.path.exists(os.path.join(model_or_path, "config.yaml")) and os.path.exists(
|
|
os.path.join(model_or_path, "model.pt")
|
|
):
|
|
config = OmegaConf.load(os.path.join(model_or_path, "config.yaml"))
|
|
kwargs = OmegaConf.merge(config, kwargs)
|
|
init_param = os.path.join(model_or_path, "model.pt")
|
|
kwargs["init_param"] = init_param
|
|
if os.path.exists(os.path.join(model_or_path, "tokens.txt")):
|
|
kwargs["tokenizer_conf"]["token_list"] = os.path.join(model_or_path, "tokens.txt")
|
|
if os.path.exists(os.path.join(model_or_path, "tokens.json")):
|
|
kwargs["tokenizer_conf"]["token_list"] = os.path.join(model_or_path, "tokens.json")
|
|
if os.path.exists(os.path.join(model_or_path, "seg_dict")):
|
|
kwargs["tokenizer_conf"]["seg_dict"] = os.path.join(model_or_path, "seg_dict")
|
|
if os.path.exists(os.path.join(model_or_path, "bpe.model")):
|
|
kwargs["tokenizer_conf"]["bpemodel"] = os.path.join(model_or_path, "bpe.model")
|
|
kwargs["model"] = config["model"]
|
|
if os.path.exists(os.path.join(model_or_path, "am.mvn")):
|
|
kwargs["frontend_conf"]["cmvn_file"] = os.path.join(model_or_path, "am.mvn")
|
|
if os.path.exists(os.path.join(model_or_path, "jieba_usr_dict")):
|
|
kwargs["jieba_usr_dict"] = os.path.join(model_or_path, "jieba_usr_dict")
|
|
if isinstance(kwargs, DictConfig):
|
|
kwargs = OmegaConf.to_container(kwargs, resolve=True)
|
|
if os.path.exists(os.path.join(model_or_path, "requirements.txt")):
|
|
requirements = os.path.join(model_or_path, "requirements.txt")
|
|
print(f"Detect model requirements, begin to install it: {requirements}")
|
|
from funasr.utils.install_model_requirements import install_requirements
|
|
|
|
install_requirements(requirements)
|
|
return kwargs
|
|
|
|
|
|
def add_file_root_path(model_or_path: str, file_path_metas: dict, cfg={}):
|
|
|
|
if isinstance(file_path_metas, dict):
|
|
for k, v in file_path_metas.items():
|
|
if isinstance(v, str):
|
|
p = os.path.join(model_or_path, v)
|
|
if os.path.exists(p):
|
|
cfg[k] = p
|
|
elif isinstance(v, dict):
|
|
if k not in cfg:
|
|
cfg[k] = {}
|
|
add_file_root_path(model_or_path, v, cfg[k])
|
|
|
|
return cfg
|
|
|
|
|
|
def get_or_download_model_dir(
|
|
model,
|
|
model_revision=None,
|
|
is_training=False,
|
|
check_latest=True,
|
|
):
|
|
"""Get local model directory or download model if necessary.
|
|
|
|
Args:
|
|
model (str): model id or path to local model directory.
|
|
model_revision (str, optional): model version number.
|
|
:param is_training:
|
|
"""
|
|
from modelscope.hub.check_model import check_local_model_is_latest
|
|
from modelscope.hub.snapshot_download import snapshot_download
|
|
|
|
from modelscope.utils.constant import Invoke, ThirdParty
|
|
|
|
key = Invoke.LOCAL_TRAINER if is_training else Invoke.PIPELINE
|
|
|
|
if os.path.exists(model) and check_latest:
|
|
model_cache_dir = model if os.path.isdir(model) else os.path.dirname(model)
|
|
try:
|
|
check_local_model_is_latest(
|
|
model_cache_dir, user_agent={Invoke.KEY: key, ThirdParty.KEY: "funasr"}
|
|
)
|
|
except:
|
|
print("could not check the latest version")
|
|
else:
|
|
model_cache_dir = snapshot_download(
|
|
model, revision=model_revision, user_agent={Invoke.KEY: key, ThirdParty.KEY: "funasr"}
|
|
)
|
|
return model_cache_dir
|
|
|
|
|
|
def get_or_download_model_dir_hf(
|
|
model,
|
|
model_revision=None,
|
|
is_training=False,
|
|
check_latest=True,
|
|
):
|
|
"""Get local model directory or download model if necessary.
|
|
|
|
Args:
|
|
model (str): model id or path to local model directory.
|
|
model_revision (str, optional): model version number.
|
|
:param is_training:
|
|
"""
|
|
from huggingface_hub import snapshot_download
|
|
|
|
model_cache_dir = snapshot_download(model)
|
|
return model_cache_dir
|