diff --git a/docs/installation/installation.md b/docs/installation/installation.md index 89000eca0..d020b51ff 100755 --- a/docs/installation/installation.md +++ b/docs/installation/installation.md @@ -34,9 +34,7 @@ Ref to [docs](https://docs.conda.io/en/latest/miniconda.html#windows-installers) ```sh pip install torch torchaudio ``` - -For more details about torch, please see [https://pytorch.org/get-started/locally](https://pytorch.org/get-started/locally) - +If there exists CUDAs in your environments, you should install the pytorch with the version matching the CUDA. The matching list could be found in [docs](https://pytorch.org/get-started/previous-versions/). ### Install funasr #### Install from pip diff --git a/docs/reference/papers.md b/docs/reference/papers.md index 33bf72fb3..fc82fceb5 100644 --- a/docs/reference/papers.md +++ b/docs/reference/papers.md @@ -3,7 +3,8 @@ FunASR have implemented the following paper code ### Speech Recognition -- [Paraformer: Fast and Accurate Parallel Transformer for Non-autoregressive End-to-End Speech Recognition](https://arxiv.org/abs/2206.08317), INTERSPEECH 2022. +- [FunASR: A Fundamental End-to-End Speech Recognition Toolkit](https://arxiv.org/abs/2305.11013), INTERSPEECH 2023 +- [Paraformer: Fast and Accurate Parallel Transformer for Non-autoregressive End-to-End Speech Recognition](https://arxiv.org/abs/2206.08317), INTERSPEECH 2022 - [Universal ASR: Unifying Streaming and Non-Streaming ASR Using a Single Encoder-Decoder Model](https://arxiv.org/abs/2010.14099), arXiv preprint arXiv:2010.14099, 2020. - [San-m: Memory equipped self-attention for end-to-end speech recognition](https://arxiv.org/pdf/2006.01713), INTERSPEECH 2020 - [Streaming Chunk-Aware Multihead Attention for Online End-to-End Speech Recognition](https://arxiv.org/abs/2006.01712), INTERSPEECH 2020