mirror of
https://github.com/modelscope/FunASR
synced 2025-09-15 14:48:36 +08:00
206 lines
17 KiB
HTML
206 lines
17 KiB
HTML
|
||
<!DOCTYPE html>
|
||
|
||
<html lang="en">
|
||
<head>
|
||
<meta charset="utf-8" />
|
||
<meta name="viewport" content="width=device-width, initial-scale=1.0" /><meta name="generator" content="Docutils 0.18.1: http://docutils.sourceforge.net/" />
|
||
|
||
|
||
<!-- Licensed under the Apache 2.0 License -->
|
||
<link rel="stylesheet" type="text/css" href="_static/fonts/open-sans/stylesheet.css" />
|
||
<!-- Licensed under the SIL Open Font License -->
|
||
<link rel="stylesheet" type="text/css" href="_static/fonts/source-serif-pro/source-serif-pro.css" />
|
||
<link rel="stylesheet" type="text/css" href="_static/css/bootstrap.min.css" />
|
||
<link rel="stylesheet" type="text/css" href="_static/css/bootstrap-theme.min.css" />
|
||
<meta name="viewport" content="width=device-width, initial-scale=1.0">
|
||
|
||
<title>Organizers — m2met2 documentation</title>
|
||
<link rel="stylesheet" type="text/css" href="_static/pygments.css" />
|
||
<link rel="stylesheet" type="text/css" href="_static/guzzle.css" />
|
||
<script data-url_root="./" id="documentation_options" src="_static/documentation_options.js"></script>
|
||
<script src="_static/jquery.js"></script>
|
||
<script src="_static/underscore.js"></script>
|
||
<script src="_static/_sphinx_javascript_frameworks_compat.js"></script>
|
||
<script src="_static/doctools.js"></script>
|
||
<script src="_static/sphinx_highlight.js"></script>
|
||
<link rel="index" title="Index" href="genindex.html" />
|
||
<link rel="search" title="Search" href="search.html" />
|
||
<link rel="next" title="Contact" href="Contact.html" />
|
||
<link rel="prev" title="Rules" href="Rules.html" />
|
||
|
||
|
||
|
||
</head><body>
|
||
<div class="related" role="navigation" aria-label="related navigation">
|
||
<h3>Navigation</h3>
|
||
<ul>
|
||
<li class="right" style="margin-right: 10px">
|
||
<a href="genindex.html" title="General Index"
|
||
accesskey="I">index</a></li>
|
||
<li class="right" >
|
||
<a href="Contact.html" title="Contact"
|
||
accesskey="N">next</a> |</li>
|
||
<li class="right" >
|
||
<a href="Rules.html" title="Rules"
|
||
accesskey="P">previous</a> |</li>
|
||
<li class="nav-item nav-item-0"><a href="index.html">m2met2 documentation</a> »</li>
|
||
<li class="nav-item nav-item-this"><a href="">Organizers</a></li>
|
||
</ul>
|
||
</div>
|
||
<div class="container-wrapper">
|
||
|
||
<div id="mobile-toggle">
|
||
<a href="#"><span class="glyphicon glyphicon-align-justify" aria-hidden="true"></span></a>
|
||
</div>
|
||
<div id="left-column">
|
||
<div class="sphinxsidebar"><a href="
|
||
index.html" class="text-logo">m2met2 documentation</a>
|
||
<div class="sidebar-block">
|
||
<div class="sidebar-wrapper">
|
||
<div id="main-search">
|
||
<form class="form-inline" action="search.html" method="GET" role="form">
|
||
<div class="input-group">
|
||
<input name="q" type="text" class="form-control" placeholder="Search...">
|
||
</div>
|
||
<input type="hidden" name="check_keywords" value="yes" />
|
||
<input type="hidden" name="area" value="default" />
|
||
</form>
|
||
</div>
|
||
</div>
|
||
</div>
|
||
<div class="sidebar-block">
|
||
<div class="sidebar-toc">
|
||
|
||
|
||
<p class="caption" role="heading"><span class="caption-text">Contents:</span></p>
|
||
<ul class="current">
|
||
<li class="toctree-l1"><a class="reference internal" href="Introduction.html">Introduction</a><ul>
|
||
<li class="toctree-l2"><a class="reference internal" href="Introduction.html#call-for-participation">Call for participation</a></li>
|
||
<li class="toctree-l2"><a class="reference internal" href="Introduction.html#timeline-aoe-time">Timeline(AOE Time)</a></li>
|
||
<li class="toctree-l2"><a class="reference internal" href="Introduction.html#guidelines">Guidelines</a></li>
|
||
</ul>
|
||
</li>
|
||
<li class="toctree-l1"><a class="reference internal" href="Dataset.html">Datasets</a><ul>
|
||
<li class="toctree-l2"><a class="reference internal" href="Dataset.html#overview-of-training-data">Overview of training data</a></li>
|
||
<li class="toctree-l2"><a class="reference internal" href="Dataset.html#detail-of-alimeeting-corpus">Detail of AliMeeting corpus</a></li>
|
||
<li class="toctree-l2"><a class="reference internal" href="Dataset.html#get-the-data">Get the data</a></li>
|
||
</ul>
|
||
</li>
|
||
<li class="toctree-l1"><a class="reference internal" href="Track_setting_and_evaluation.html">Track & Evaluation</a><ul>
|
||
<li class="toctree-l2"><a class="reference internal" href="Track_setting_and_evaluation.html#speaker-attributed-asr-main-track">Speaker-Attributed ASR (Main Track)</a></li>
|
||
<li class="toctree-l2"><a class="reference internal" href="Track_setting_and_evaluation.html#evaluation-metric">Evaluation metric</a></li>
|
||
<li class="toctree-l2"><a class="reference internal" href="Track_setting_and_evaluation.html#sub-track-arrangement">Sub-track arrangement</a></li>
|
||
</ul>
|
||
</li>
|
||
<li class="toctree-l1"><a class="reference internal" href="Baseline.html">Baseline</a><ul>
|
||
<li class="toctree-l2"><a class="reference internal" href="Baseline.html#overview">Overview</a></li>
|
||
<li class="toctree-l2"><a class="reference internal" href="Baseline.html#quick-start">Quick start</a></li>
|
||
<li class="toctree-l2"><a class="reference internal" href="Baseline.html#baseline-results">Baseline results</a></li>
|
||
</ul>
|
||
</li>
|
||
<li class="toctree-l1"><a class="reference internal" href="Rules.html">Rules</a></li>
|
||
<li class="toctree-l1 current"><a class="current reference internal" href="#">Organizers</a></li>
|
||
<li class="toctree-l1"><a class="reference internal" href="Contact.html">Contact</a></li>
|
||
</ul>
|
||
|
||
|
||
</div>
|
||
</div>
|
||
|
||
</div>
|
||
</div>
|
||
<div id="right-column">
|
||
|
||
<div role="navigation" aria-label="breadcrumbs navigation">
|
||
<ol class="breadcrumb">
|
||
<li><a href="index.html">Docs</a></li>
|
||
|
||
<li>Organizers</li>
|
||
</ol>
|
||
</div>
|
||
|
||
<div class="document clearer body">
|
||
|
||
<section id="organizers">
|
||
<h1>Organizers<a class="headerlink" href="#organizers" title="Permalink to this heading">¶</a></h1>
|
||
<p><em><strong>Lei Xie, Professor, Northwestern Polytechnical University, China</strong></em></p>
|
||
<p>Email: <a class="reference external" href="mailto:lxie%40nwpu.edu.cn">lxie<span>@</span>nwpu<span>.</span>edu<span>.</span>cn</a></p>
|
||
<a class="reference internal image-reference" href="_images/lxie.jpeg"><img alt="lxie" src="_images/lxie.jpeg" style="width: 20%;" /></a>
|
||
<p>Lei Xie received the Ph.D. degree in computer science from Northwestern Polytechnical University, Xi’an, China, in 2004. From 2001 to 2002, he was with the Department of Electronics and Information Processing, Vrije Universiteit Brussel (VUB), Brussels, Belgium, as a Visiting Scientist. From 2004 to 2006, he was a Senior Research Associate with the Center for Media Technology, School of Creative Media, City University of Hong Kong, Hong Kong, China. From 2006 to 2007, he was a Postdoctoral Fellow with the Human-Computer Communications Laboratory (HCCL), The Chinese University of Hong Kong, Hong Kong, China. He is currently a Professor with School of Computer Science, Northwestern Polytechnical University, Xian, China and leads the Audio, Speech and Language Processing Group (ASLP@NPU). He has published over 200 papers in referred journals and conferences, such as IEEE/ACM Transactions on Audio, Speech and Language Processing, IEEE Transactions on Multimedia, Interspeech, ICASSP, ASRU, ACL and ACM Multimedia. He has achieved several best paper awards in flagship conferences. His current research interests include general topics in speech and language processing, multimedia, and human-computer interaction. Dr. Xie is currently an associate editor (AE) of IEEE/ACM Trans. on Audio, Speech and language Processing. He has actively served as Chairs in many conferences and technical committees. He serves as an IEEE Speech and Language Processing
|
||
Technical Committee Member.</p>
|
||
<p><em><strong>Kong Aik Lee, Senior Scientist at Institute for Infocomm Research, A*Star, Singapore</strong></em></p>
|
||
<p>Email: <a class="reference external" href="mailto:kongaik.lee%40ieee.org">kongaik<span>.</span>lee<span>@</span>ieee<span>.</span>org</a></p>
|
||
<a class="reference internal image-reference" href="_images/kong.png"><img alt="kong" src="_images/kong.png" style="width: 20%;" /></a>
|
||
<p>Kong Aik Lee started off him career as a researcher, then a team leader and a strategic planning manager, at the Institute Infocomm Research, A*STAR, Singapore, working on speaker and language recognition research. From 2018 to 2020, he spent two and a half years in NEC Corporation, Japan, focusing very much on voice biometrics and multi-modal biometrics products. He is proud to work with a great team on voice biometrics featured on NEC Bio-Idiom platform. He returned to Singapore in July 2020, and now leading the speech and audio analytics research at the Institute for Infocomm Research, as a Senior Scientist and PI. He also serve as an Editor for Elsevier Computer Speech and Language (since 2016), and was an Associate Editor for IEEE/ACM Transactions on Audio, Speech and Language Processing (2017 - 2021), and am an elected member of IEEE Speech and Language Technical Committee (2019 - 2021).</p>
|
||
<p><em><strong>Zhijie Yan, Principal Engineer at Alibaba, China</strong></em>
|
||
Email: <a class="reference external" href="mailto:zhijie.yzj%40alibaba-inc.com">zhijie<span>.</span>yzj<span>@</span>alibaba-inc<span>.</span>com</a></p>
|
||
<a class="reference internal image-reference" href="_images/zhijie.jpg"><img alt="zhijie" src="_images/zhijie.jpg" style="width: 20%;" /></a>
|
||
<p>Zhijie Yan holds a PhD from the University of Science and Technology of China, and is a senior member of the Institute of Electrical and Electronics Engineers (IEEE). He is also an expert reviewer of top academic conferences and journals in the speech field. His research fields include speech recognition, speech synthesis, voiceprints, and speech interaction. His research results are applied in speech services provided by Alibaba Group and Ant Financial. He was awarded the title of “One of the Top 100 Grassroots Scientists” by the China Association for Science and Technology.</p>
|
||
<p><em><strong>Shiliang Zhang, Senior Engineer at Alibaba, China</strong></em>
|
||
Email: <a class="reference external" href="mailto:sly.zsl%40alibaba-inc.com">sly<span>.</span>zsl<span>@</span>alibaba-inc<span>.</span>com</a></p>
|
||
<a class="reference internal image-reference" href="_images/zsl.JPG"><img alt="zsl" src="_images/zsl.JPG" style="width: 20%;" /></a>
|
||
<p>Shiliang Zhang graduated with a Ph.D. from the University of Science and Technology of China in 2017. His research areas mainly include speech recognition, natural language understanding, and machine learning. Currently, he has published over 40 papers in mainstream academic journals and conferences in the fields of speech and machine learning, and has applied for dozens of patents. After obtaining his doctorate degree, he joined the Alibaba Intelligent Speech team. He is currently leading the direction of speech recognition and fundamental technology at DAMO Academy’s speech laboratory.</p>
|
||
<p><em><strong>Yanmin Qian, Professor, Shanghai Jiao Tong University, China</strong></em></p>
|
||
<p>Email: <a class="reference external" href="mailto:yanminqian%40sjtu.edu.cn">yanminqian<span>@</span>sjtu<span>.</span>edu<span>.</span>cn</a></p>
|
||
<a class="reference internal image-reference" href="_images/qian.jpeg"><img alt="qian" src="_images/qian.jpeg" style="width: 20%;" /></a>
|
||
<p>Yanmin Qian received the B.S. degree from the Department of Electronic and Information Engineering,Huazhong University of Science and Technology, Wuhan, China, in 2007, and the Ph.D. degree from the Department of Electronic Engineering, Tsinghua University, Beijing, China, in 2012. Since 2013, he has been with the Department of Computer Science and Engineering, Shanghai Jiao Tong University (SJTU), Shanghai, China, where he is currently an Associate Professor. From 2015 to 2016, he also worked as an Associate Research in the Speech Group, Cambridge University Engineering Department, Cambridge, U.K. He is a senior member of IEEE and a member of ISCA, and one of the founding members of Kaldi Speech Recognition Toolkit. He has published more than 110 papers on speech and language processing with 4000+ citations, including the top conference: ICASSP, INTERSPEECH and ASRU. His current research interests include the acoustic and language modeling in speech recognition, speaker and language recognition, key word spotting, and multimedia signal processing.</p>
|
||
<p><em><strong>Zhuo Chen, Applied Scientist in Microsoft, USA</strong></em></p>
|
||
<p>Email: <a class="reference external" href="mailto:zhuc%40microsoft.com">zhuc<span>@</span>microsoft<span>.</span>com</a></p>
|
||
<a class="reference internal image-reference" href="_images/chenzhuo.jpg"><img alt="chenzhuo" src="_images/chenzhuo.jpg" style="width: 20%;" /></a>
|
||
<p>Zhuo Chen received the Ph.D. degree from Columbia University, New York, NY, USA, in 2017. He is currently a Principal Applied Data Scientist with Microsoft. He has authored or coauthored more than 80 papers in peer-reviewed journals and conferences with around 6000 citations. He is a reviewer or technical committee member for more than ten journals and conferences. His research interests include automatic conversation recognition, speech separation, diarisation, and speaker information extraction. He actively participated in the academic events and challenges, and won several awards. Meanwhile, he contributed to open-sourced datasets, such as WSJ0-2mix, LibriCSS, and AISHELL-4, that have been main benchmark datasets for multi-speaker processing research. In 2020, he was the Team Leader in 2020 Jelinek workshop, leading more than 30 researchers and students to push the state of the art in conversation transcription.</p>
|
||
<p><em><strong>Jian Wu, Applied Scientist in Microsoft, USA</strong></em></p>
|
||
<p>Email: <a class="reference external" href="mailto:wujian%40microsoft.com">wujian<span>@</span>microsoft<span>.</span>com</a></p>
|
||
<a class="reference internal image-reference" href="_images/wujian.jpg"><img alt="wujian" src="_images/wujian.jpg" style="width: 20%;" /></a>
|
||
<p>Jian Wu received a master degree from Northwestern Polytechnical University, Xi’an, China, in 2020 and currently he is a Applied Scientist in Microsoft, USA. His research interests cover multi-channel signal processing, robust and multi-talker speech recognition, speech enhancement, dereverberation and separation. He has around 30 conference publications with a total citation over 1200. He participated in several challenges such as CHiME5, DNS 2020 and FFSVC 2020 and contributed to the open-sourced datasets including LibriCSS and AISHELL-4. He is also a reviewer for several journals and conferences such as ICASSP, SLT, TASLP and SPL.</p>
|
||
<p><em><strong>Hui Bu, CEO, AISHELL foundation, China</strong></em></p>
|
||
<p>Email: <a class="reference external" href="mailto:buhui%40aishelldata.com">buhui<span>@</span>aishelldata<span>.</span>com</a></p>
|
||
<a class="reference internal image-reference" href="_images/buhui.jpeg"><img alt="buhui" src="_images/buhui.jpeg" style="width: 20%;" /></a>
|
||
<p>Hui Bu received his master degree in the Artificial Intelligence Laboratory of Korea University in 2014. He is the founder and the CEO of AISHELL and AISHELL foundation. He participated in the release of AISHELL 1 & 2 & 3 & 4, DMASH and HI-MIA open source database project and is the co-founder of China Kaldi offline Technology Forum.</p>
|
||
</section>
|
||
|
||
|
||
</div>
|
||
|
||
<div class="footer-relations">
|
||
|
||
<div class="pull-left">
|
||
<a class="btn btn-default" href="Rules.html" title="previous chapter (use the left arrow)">Rules</a>
|
||
</div>
|
||
|
||
<div class="pull-right">
|
||
<a class="btn btn-default" href="Contact.html" title="next chapter (use the right arrow)">Contact</a>
|
||
</div>
|
||
</div>
|
||
<div class="clearer"></div>
|
||
|
||
</div>
|
||
<div class="clearfix"></div>
|
||
</div>
|
||
<div class="related" role="navigation" aria-label="related navigation">
|
||
<h3>Navigation</h3>
|
||
<ul>
|
||
<li class="right" style="margin-right: 10px">
|
||
<a href="genindex.html" title="General Index"
|
||
>index</a></li>
|
||
<li class="right" >
|
||
<a href="Contact.html" title="Contact"
|
||
>next</a> |</li>
|
||
<li class="right" >
|
||
<a href="Rules.html" title="Rules"
|
||
>previous</a> |</li>
|
||
<li class="nav-item nav-item-0"><a href="index.html">m2met2 documentation</a> »</li>
|
||
<li class="nav-item nav-item-this"><a href="">Organizers</a></li>
|
||
</ul>
|
||
</div>
|
||
<script type="text/javascript">
|
||
$("#mobile-toggle a").click(function () {
|
||
$("#left-column").toggle();
|
||
});
|
||
</script>
|
||
<script type="text/javascript" src="_static/js/bootstrap.js"></script>
|
||
<div class="footer">
|
||
© Copyright 2023, Speech Lab, Alibaba Group; ASLP Group, Northwestern Polytechnical University. Created using <a href="http://sphinx.pocoo.org/">Sphinx</a>.
|
||
</div>
|
||
</body>
|
||
</html> |