Alert button
Picture for Aswin Shanmugam Subramanian

Aswin Shanmugam Subramanian

Alert button

TS-SEP: Joint Diarization and Separation Conditioned on Estimated Speaker Embeddings

Mar 08, 2023
Christoph Boeddeker, Aswin Shanmugam Subramanian, Gordon Wichern, Reinhold Haeb-Umbach, Jonathan Le Roux

Figure 1 for TS-SEP: Joint Diarization and Separation Conditioned on Estimated Speaker Embeddings
Figure 2 for TS-SEP: Joint Diarization and Separation Conditioned on Estimated Speaker Embeddings
Figure 3 for TS-SEP: Joint Diarization and Separation Conditioned on Estimated Speaker Embeddings
Figure 4 for TS-SEP: Joint Diarization and Separation Conditioned on Estimated Speaker Embeddings
Viaarxiv icon

Tackling the Cocktail Fork Problem for Separation and Transcription of Real-World Soundtracks

Dec 14, 2022
Darius Petermann, Gordon Wichern, Aswin Shanmugam Subramanian, Zhong-Qiu Wang, Jonathan Le Roux

Figure 1 for Tackling the Cocktail Fork Problem for Separation and Transcription of Real-World Soundtracks
Figure 2 for Tackling the Cocktail Fork Problem for Separation and Transcription of Real-World Soundtracks
Figure 3 for Tackling the Cocktail Fork Problem for Separation and Transcription of Real-World Soundtracks
Figure 4 for Tackling the Cocktail Fork Problem for Separation and Transcription of Real-World Soundtracks
Viaarxiv icon

Reverberation as Supervision for Speech Separation

Nov 15, 2022
Rohith Aralikatti, Christoph Boeddeker, Gordon Wichern, Aswin Shanmugam Subramanian, Jonathan Le Roux

Figure 1 for Reverberation as Supervision for Speech Separation
Figure 2 for Reverberation as Supervision for Speech Separation
Figure 3 for Reverberation as Supervision for Speech Separation
Figure 4 for Reverberation as Supervision for Speech Separation
Viaarxiv icon

An Exploration of Self-Supervised Pretrained Representations for End-to-End Speech Recognition

Oct 09, 2021
Xuankai Chang, Takashi Maekaku, Pengcheng Guo, Jing Shi, Yen-Ju Lu, Aswin Shanmugam Subramanian, Tianzi Wang, Shu-wen Yang, Yu Tsao, Hung-yi Lee, Shinji Watanabe

Figure 1 for An Exploration of Self-Supervised Pretrained Representations for End-to-End Speech Recognition
Figure 2 for An Exploration of Self-Supervised Pretrained Representations for End-to-End Speech Recognition
Figure 3 for An Exploration of Self-Supervised Pretrained Representations for End-to-End Speech Recognition
Figure 4 for An Exploration of Self-Supervised Pretrained Representations for End-to-End Speech Recognition
Viaarxiv icon

Deep Learning based Multi-Source Localization with Source Splitting and its Effectiveness in Multi-Talker Speech Recognition

Feb 16, 2021
Aswin Shanmugam Subramanian, Chao Weng, Shinji Watanabe, Meng Yu, Dong Yu

Figure 1 for Deep Learning based Multi-Source Localization with Source Splitting and its Effectiveness in Multi-Talker Speech Recognition
Figure 2 for Deep Learning based Multi-Source Localization with Source Splitting and its Effectiveness in Multi-Talker Speech Recognition
Figure 3 for Deep Learning based Multi-Source Localization with Source Splitting and its Effectiveness in Multi-Talker Speech Recognition
Figure 4 for Deep Learning based Multi-Source Localization with Source Splitting and its Effectiveness in Multi-Talker Speech Recognition
Viaarxiv icon

The 2020 ESPnet update: new features, broadened applications, performance improvements, and future plans

Dec 23, 2020
Shinji Watanabe, Florian Boyer, Xuankai Chang, Pengcheng Guo, Tomoki Hayashi, Yosuke Higuchi, Takaaki Hori, Wen-Chin Huang, Hirofumi Inaguma, Naoyuki Kamo, Shigeki Karita, Chenda Li, Jing Shi, Aswin Shanmugam Subramanian, Wangyou Zhang

Figure 1 for The 2020 ESPnet update: new features, broadened applications, performance improvements, and future plans
Figure 2 for The 2020 ESPnet update: new features, broadened applications, performance improvements, and future plans
Viaarxiv icon

Directional ASR: A New Paradigm for E2E Multi-Speaker Speech Recognition with Source Localization

Oct 30, 2020
Aswin Shanmugam Subramanian, Chao Weng, Shinji Watanabe, Meng Yu, Yong Xu, Shi-Xiong Zhang, Dong Yu

Figure 1 for Directional ASR: A New Paradigm for E2E Multi-Speaker Speech Recognition with Source Localization
Figure 2 for Directional ASR: A New Paradigm for E2E Multi-Speaker Speech Recognition with Source Localization
Figure 3 for Directional ASR: A New Paradigm for E2E Multi-Speaker Speech Recognition with Source Localization
Figure 4 for Directional ASR: A New Paradigm for E2E Multi-Speaker Speech Recognition with Source Localization
Viaarxiv icon

CHiME-6 Challenge:Tackling Multispeaker Speech Recognition for Unsegmented Recordings

May 02, 2020
Shinji Watanabe, Michael Mandel, Jon Barker, Emmanuel Vincent, Ashish Arora, Xuankai Chang, Sanjeev Khudanpur, Vimal Manohar, Daniel Povey, Desh Raj, David Snyder, Aswin Shanmugam Subramanian, Jan Trmal, Bar Ben Yair, Christoph Boeddeker, Zhaoheng Ni, Yusuke Fujita, Shota Horiguchi, Naoyuki Kanda, Takuya Yoshioka, Neville Ryant

Figure 1 for CHiME-6 Challenge:Tackling Multispeaker Speech Recognition for Unsegmented Recordings
Figure 2 for CHiME-6 Challenge:Tackling Multispeaker Speech Recognition for Unsegmented Recordings
Figure 3 for CHiME-6 Challenge:Tackling Multispeaker Speech Recognition for Unsegmented Recordings
Figure 4 for CHiME-6 Challenge:Tackling Multispeaker Speech Recognition for Unsegmented Recordings
Viaarxiv icon

An Investigation of End-to-End Multichannel Speech Recognition for Reverberant and Mismatch Conditions

Apr 28, 2019
Aswin Shanmugam Subramanian, Xiaofei Wang, Shinji Watanabe, Toru Taniguchi, Dung Tran, Yuya Fujita

Figure 1 for An Investigation of End-to-End Multichannel Speech Recognition for Reverberant and Mismatch Conditions
Figure 2 for An Investigation of End-to-End Multichannel Speech Recognition for Reverberant and Mismatch Conditions
Figure 3 for An Investigation of End-to-End Multichannel Speech Recognition for Reverberant and Mismatch Conditions
Viaarxiv icon

Building state-of-the-art distant speech recognition using the CHiME-4 challenge with a setup of speech enhancement baseline

Mar 27, 2018
Szu-Jui Chen, Aswin Shanmugam Subramanian, Hainan Xu, Shinji Watanabe

Figure 1 for Building state-of-the-art distant speech recognition using the CHiME-4 challenge with a setup of speech enhancement baseline
Viaarxiv icon