diff --git a/.github/ISSUE_TEMPLATE/ask_questions.md b/.github/ISSUE_TEMPLATE/ask_questions.md new file mode 100644 index 0000000..4716faa --- /dev/null +++ b/.github/ISSUE_TEMPLATE/ask_questions.md @@ -0,0 +1,38 @@ +--- +name: ❓ Questions/Help +about: If you have questions, please first search existing issues and docs +labels: 'question, needs triage' +--- + +Notice: In order to resolve issues more efficiently, please raise issue following the template. +(注意:为了更加高效率解决您遇到的问题,请按照模板提问,补充细节) + +## ❓ Questions and Help + + +### Before asking: +1. search the issues. +2. search the docs. + + + +#### What is your question? + +#### Code + + + +#### What have you tried? + +#### What's your environment? + + - OS (e.g., Linux): + - FunASR Version (e.g., 1.0.0): + - ModelScope Version (e.g., 1.11.0): + - PyTorch Version (e.g., 2.0.0): + - How you installed funasr (`pip`, source): + - Python version: + - GPU (e.g., V100M32) + - CUDA/cuDNN version (e.g., cuda11.7): + - Docker version (e.g., funasr-runtime-sdk-cpu-0.4.1) + - Any other relevant information: \ No newline at end of file diff --git a/.github/ISSUE_TEMPLATE/bug_report.md b/.github/ISSUE_TEMPLATE/bug_report.md new file mode 100644 index 0000000..80b0317 --- /dev/null +++ b/.github/ISSUE_TEMPLATE/bug_report.md @@ -0,0 +1,47 @@ +--- +name: 🐛 Bug Report +about: Submit a bug report to help us improve +labels: 'bug, needs triage' +--- + +Notice: In order to resolve issues more efficiently, please raise issue following the template. +(注意:为了更加高效率解决您遇到的问题,请按照模板提问,补充细节) + +## 🐛 Bug + + + +### To Reproduce + +Steps to reproduce the behavior (**always include the command you ran**): + +1. Run cmd '....' +2. See error + + + + +#### Code sample + + +### Expected behavior + + + +### Environment + + - OS (e.g., Linux): + - FunASR Version (e.g., 1.0.0): + - ModelScope Version (e.g., 1.11.0): + - PyTorch Version (e.g., 2.0.0): + - How you installed funasr (`pip`, source): + - Python version: + - GPU (e.g., V100M32) + - CUDA/cuDNN version (e.g., cuda11.7): + - Docker version (e.g., funasr-runtime-sdk-cpu-0.4.1) + - Any other relevant information: + +### Additional context + + \ No newline at end of file diff --git a/.github/ISSUE_TEMPLATE/config.yaml b/.github/ISSUE_TEMPLATE/config.yaml new file mode 100644 index 0000000..ec4bb38 --- /dev/null +++ b/.github/ISSUE_TEMPLATE/config.yaml @@ -0,0 +1 @@ +blank_issues_enabled: false \ No newline at end of file diff --git a/.github/ISSUE_TEMPLATE/error_docs.md b/.github/ISSUE_TEMPLATE/error_docs.md new file mode 100644 index 0000000..ee731c1 --- /dev/null +++ b/.github/ISSUE_TEMPLATE/error_docs.md @@ -0,0 +1,15 @@ +--- +name: 📚 Documentation/Typos +about: Report an issue related to documentation or a typo +labels: 'documentation, needs triage' +--- + +## 📚 Documentation + +For typos and doc fixes, please go ahead and: + +1. Create an issue. +2. Fix the typo. +3. Submit a PR. + +Thanks! \ No newline at end of file diff --git a/README.md b/README.md index 98a4236..4c8a49c 100644 --- a/README.md +++ b/README.md @@ -121,7 +121,7 @@ model = AutoModel( res = model.generate( input=f"{model.model_path}/example/en.mp3", cache={}, - language="auto", # "zn", "en", "yue", "ja", "ko", "nospeech" + language="auto", # "zh", "en", "yue", "ja", "ko", "nospeech" use_itn=True, batch_size_s=60, merge_vad=True, # @@ -150,7 +150,7 @@ model = AutoModel(model=model_dir, trust_remote_code=True, device="cuda:0") res = model.generate( input=f"{model.model_path}/example/en.mp3", cache={}, - language="zh", # "zn", "en", "yue", "ja", "ko", "nospeech" + language="zh", # "zh", "en", "yue", "ja", "ko", "nospeech" use_itn=False, batch_size=64, ) @@ -172,7 +172,7 @@ m, kwargs = SenseVoiceSmall.from_pretrained(model=model_dir, device="cuda:0") res = m.inference( data_in=f"{kwargs['model_path']}/example/en.mp3", - language="auto", # "zn", "en", "yue", "ja", "ko", "nospeech" + language="auto", # "zh", "en", "yue", "ja", "ko", "nospeech" use_itn=False, **kwargs, ) diff --git a/README_ja.md b/README_ja.md index 61dd605..44df014 100644 --- a/README_ja.md +++ b/README_ja.md @@ -121,7 +121,7 @@ model = AutoModel( res = model.generate( input=f"{model.model_path}/example/en.mp3", cache={}, - language="auto", # "zn", "en", "yue", "ja", "ko", "nospeech" + language="auto", # "zh", "en", "yue", "ja", "ko", "nospeech" use_itn=True, batch_size_s=60, merge_vad=True, # @@ -150,7 +150,7 @@ model = AutoModel(model=model_dir, trust_remote_code=True, device="cuda:0") res = model.generate( input=f"{model.model_path}/example/en.mp3", cache={}, - language="auto", # "zn", "en", "yue", "ja", "ko", "nospeech" + language="auto", # "zh", "en", "yue", "ja", "ko", "nospeech" use_itn=True, batch_size=64, ) @@ -172,7 +172,7 @@ m, kwargs = SenseVoiceSmall.from_pretrained(model=model_dir, device="cuda:0") res = m.inference( data_in=f"{kwargs['model_path']}/example/en.mp3", - language="auto", # "zn", "en", "yue", "ja", "ko", "nospeech" + language="auto", # "zh", "en", "yue", "ja", "ko", "nospeech" use_itn=False, **kwargs, ) diff --git a/README_zh.md b/README_zh.md index c3475a3..86a3b27 100644 --- a/README_zh.md +++ b/README_zh.md @@ -125,7 +125,7 @@ model = AutoModel( res = model.generate( input=f"{model.model_path}/example/en.mp3", cache={}, - language="auto", # "zn", "en", "yue", "ja", "ko", "nospeech" + language="auto", # "zh", "en", "yue", "ja", "ko", "nospeech" use_itn=True, batch_size_s=60, merge_vad=True, # @@ -154,7 +154,7 @@ model = AutoModel(model=model_dir, trust_remote_code=True, device="cuda:0") res = model.generate( input=f"{model.model_path}/example/en.mp3", cache={}, - language="auto", # "zn", "en", "yue", "ja", "ko", "nospeech" + language="auto", # "zh", "en", "yue", "ja", "ko", "nospeech" use_itn=True, batch_size=64, ) @@ -176,7 +176,7 @@ m, kwargs = SenseVoiceSmall.from_pretrained(model=model_dir, device="cuda:0") res = m.inference( data_in=f"{kwargs['model_path']}/example/en.mp3", - language="auto", # "zn", "en", "yue", "ja", "ko", "nospeech" + language="auto", # "zh", "en", "yue", "ja", "ko", "nospeech" use_itn=False, **kwargs, ) diff --git a/demo1.py b/demo1.py index 946adc0..c47b985 100644 --- a/demo1.py +++ b/demo1.py @@ -22,7 +22,7 @@ model = AutoModel( res = model.generate( input=f"{model.model_path}/example/en.mp3", cache={}, - language="auto", # "zn", "en", "yue", "ja", "ko", "nospeech" + language="auto", # "zh", "en", "yue", "ja", "ko", "nospeech" use_itn=True, batch_size_s=60, merge_vad=True, # @@ -35,7 +35,7 @@ print(text) res = model.generate( input=f"{model.model_path}/example/zh.mp3", cache={}, - language="auto", # "zn", "en", "yue", "ja", "ko", "nospeech" + language="auto", # "zh", "en", "yue", "ja", "ko", "nospeech" use_itn=True, batch_size_s=60, merge_vad=True, # @@ -48,7 +48,7 @@ print(text) res = model.generate( input=f"{model.model_path}/example/yue.mp3", cache={}, - language="auto", # "zn", "en", "yue", "ja", "ko", "nospeech" + language="auto", # "zh", "en", "yue", "ja", "ko", "nospeech" use_itn=True, batch_size_s=60, merge_vad=True, # @@ -61,7 +61,7 @@ print(text) res = model.generate( input=f"{model.model_path}/example/ja.mp3", cache={}, - language="auto", # "zn", "en", "yue", "ja", "ko", "nospeech" + language="auto", # "zh", "en", "yue", "ja", "ko", "nospeech" use_itn=True, batch_size_s=60, merge_vad=True, # @@ -75,7 +75,7 @@ print(text) res = model.generate( input=f"{model.model_path}/example/ko.mp3", cache={}, - language="auto", # "zn", "en", "yue", "ja", "ko", "nospeech" + language="auto", # "zh", "en", "yue", "ja", "ko", "nospeech" use_itn=True, batch_size_s=60, merge_vad=True, # diff --git a/demo2.py b/demo2.py index a6dbc1d..818336d 100644 --- a/demo2.py +++ b/demo2.py @@ -13,7 +13,7 @@ m, kwargs = SenseVoiceSmall.from_pretrained(model=model_dir, device="cuda:0") res = m.inference( data_in=f"{kwargs['model_path']}/example/en.mp3", - language="auto", # "zn", "en", "yue", "ja", "ko", "nospeech" + language="auto", # "zh", "en", "yue", "ja", "ko", "nospeech" use_itn=False, **kwargs, )