From 179a3f99c45d21cec3ea17e3b9265bcf1e49c617 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E8=AF=AD=E5=B8=86?= Date: Wed, 28 Feb 2024 19:30:15 +0800 Subject: [PATCH] test --- .../lcbnet/demo_nj.sh | 77 ++++++++++--------- 1 file changed, 39 insertions(+), 38 deletions(-) diff --git a/examples/industrial_data_pretraining/lcbnet/demo_nj.sh b/examples/industrial_data_pretraining/lcbnet/demo_nj.sh index 5e634c315..d9f42a033 100755 --- a/examples/industrial_data_pretraining/lcbnet/demo_nj.sh +++ b/examples/industrial_data_pretraining/lcbnet/demo_nj.sh @@ -16,46 +16,46 @@ inference_dir="outputs/test" _logdir="${inference_dir}/logdir" echo "inference_dir: ${inference_dir}" -mkdir -p "${_logdir}" -key_file1=${file_dir}/wav.scp -key_file2=${file_dir}/ocr.txt -split_scps1= -split_scps2= -for JOB in $(seq "${nj}"); do - split_scps1+=" ${_logdir}/wav.${JOB}.scp" - split_scps2+=" ${_logdir}/ocr.${JOB}.txt" -done -utils/split_scp.pl "${key_file1}" ${split_scps1} -utils/split_scp.pl "${key_file2}" ${split_scps2} +# mkdir -p "${_logdir}" +# key_file1=${file_dir}/wav.scp +# key_file2=${file_dir}/ocr.txt +# split_scps1= +# split_scps2= +# for JOB in $(seq "${nj}"); do +# split_scps1+=" ${_logdir}/wav.${JOB}.scp" +# split_scps2+=" ${_logdir}/ocr.${JOB}.txt" +# done +# utils/split_scp.pl "${key_file1}" ${split_scps1} +# utils/split_scp.pl "${key_file2}" ${split_scps2} -gpuid_list_array=(${CUDA_VISIBLE_DEVICES//,/ }) -for JOB in $(seq ${nj}); do - { - id=$((JOB-1)) - gpuid=${gpuid_list_array[$id]} +# gpuid_list_array=(${CUDA_VISIBLE_DEVICES//,/ }) +# for JOB in $(seq ${nj}); do +# { +# id=$((JOB-1)) +# gpuid=${gpuid_list_array[$id]} - export CUDA_VISIBLE_DEVICES=${gpuid} +# export CUDA_VISIBLE_DEVICES=${gpuid} - python -m funasr.bin.inference \ - --config-path=${file_dir} \ - --config-name="config.yaml" \ - ++init_param=${file_dir}/model.pb \ - ++tokenizer_conf.token_list=${file_dir}/tokens.txt \ - ++input=[${_logdir}/wav.${JOB}.scp,${_logdir}/ocr.${JOB}.txt] \ - +data_type='["kaldi_ark", "text"]' \ - ++tokenizer_conf.bpemodel=${file_dir}/bpe.model \ - ++output_dir="${inference_dir}/${JOB}" \ - ++device="${inference_device}" \ - ++ncpu=1 \ - ++disable_log=true &> ${_logdir}/log.${JOB}.txt +# python -m funasr.bin.inference \ +# --config-path=${file_dir} \ +# --config-name="config.yaml" \ +# ++init_param=${file_dir}/model.pb \ +# ++tokenizer_conf.token_list=${file_dir}/tokens.txt \ +# ++input=[${_logdir}/wav.${JOB}.scp,${_logdir}/ocr.${JOB}.txt] \ +# +data_type='["kaldi_ark", "text"]' \ +# ++tokenizer_conf.bpemodel=${file_dir}/bpe.model \ +# ++output_dir="${inference_dir}/${JOB}" \ +# ++device="${inference_device}" \ +# ++ncpu=1 \ +# ++disable_log=true &> ${_logdir}/log.${JOB}.txt - }& -done -wait +# }& +# done +# wait -mkdir -p ${inference_dir}/1best_recog -for f in token score text; do +#mkdir -p ${inference_dir}/1best_recog +for f in token; do if [ -f "${inference_dir}/${JOB}/1best_recog/${f}" ]; then for JOB in $(seq "${nj}"); do cat "${inference_dir}/${JOB}/1best_recog/${f}" @@ -65,7 +65,8 @@ done echo "Computing WER ..." echo "Computing WER ..." -python utils/postprocess_text_zh.py ${inference_dir}/1best_recog/text ${inference_dir}/1best_recog/text.proc -python utils/postprocess_text_zh.py ${data_dir}/text ${inference_dir}/1best_recog/text.ref -python utils/compute_wer.py ${inference_dir}/1best_recog/text.ref ${inference_dir}/1best_recog/text.proc ${inference_dir}/1best_recog/text.cer -tail -n 3 ${inference_dir}/1best_recog/text.cer \ No newline at end of file +#python utils/postprocess_text_zh.py ${inference_dir}/1best_recog/text ${inference_dir}/1best_recog/text.proc + +#cp ${data_dir}/text ${inference_dir}/1best_recog/text.ref +#python utils/compute_wer.py ${inference_dir}/1best_recog/text.ref ${inference_dir}/1best_recog/text.proc ${inference_dir}/1best_recog/text.cer +#tail -n 3 ${inference_dir}/1best_recog/text.cer \ No newline at end of file