Alert button
Picture for Desh Raj

Desh Raj

Alert button

Listening to Multi-talker Conversations: Modular and End-to-end Perspectives

Add code
Bookmark button
Alert button
Feb 14, 2024
Desh Raj

Viaarxiv icon

On Speaker Attribution with SURT

Add code
Bookmark button
Alert button
Jan 28, 2024
Desh Raj, Matthew Wiesner, Matthew Maciejewski, Leibny Paola Garcia-Perera, Daniel Povey, Sanjeev Khudanpur

Viaarxiv icon

Updated Corpora and Benchmarks for Long-Form Speech Recognition

Add code
Bookmark button
Alert button
Sep 26, 2023
Jennifer Drexler Fox, Desh Raj, Natalie Delworth, Quinn McNamara, Corey Miller, Migüel Jetté

Viaarxiv icon

Learning from Flawed Data: Weakly Supervised Automatic Speech Recognition

Add code
Bookmark button
Alert button
Sep 26, 2023
Dongji Gao, Hainan Xu, Desh Raj, Leibny Paola Garcia Perera, Daniel Povey, Sanjeev Khudanpur

Figure 1 for Learning from Flawed Data: Weakly Supervised Automatic Speech Recognition
Figure 2 for Learning from Flawed Data: Weakly Supervised Automatic Speech Recognition
Figure 3 for Learning from Flawed Data: Weakly Supervised Automatic Speech Recognition
Figure 4 for Learning from Flawed Data: Weakly Supervised Automatic Speech Recognition
Viaarxiv icon

Training dynamic models using early exits for automatic speech recognition on resource-constrained devices

Add code
Bookmark button
Alert button
Sep 18, 2023
George August Wright, Umberto Cappellazzo, Salah Zaiem, Desh Raj, Lucas Ondel Yang, Daniele Falavigna, Alessio Brutti

Figure 1 for Training dynamic models using early exits for automatic speech recognition on resource-constrained devices
Figure 2 for Training dynamic models using early exits for automatic speech recognition on resource-constrained devices
Figure 3 for Training dynamic models using early exits for automatic speech recognition on resource-constrained devices
Figure 4 for Training dynamic models using early exits for automatic speech recognition on resource-constrained devices
Viaarxiv icon

The CHiME-7 DASR Challenge: Distant Meeting Transcription with Multiple Devices in Diverse Scenarios

Add code
Bookmark button
Alert button
Jul 14, 2023
Samuele Cornell, Matthew Wiesner, Shinji Watanabe, Desh Raj, Xuankai Chang, Paola Garcia, Matthew Maciejewski, Yoshiki Masuyama, Zhong-Qiu Wang, Stefano Squartini, Sanjeev Khudanpur

Figure 1 for The CHiME-7 DASR Challenge: Distant Meeting Transcription with Multiple Devices in Diverse Scenarios
Figure 2 for The CHiME-7 DASR Challenge: Distant Meeting Transcription with Multiple Devices in Diverse Scenarios
Figure 3 for The CHiME-7 DASR Challenge: Distant Meeting Transcription with Multiple Devices in Diverse Scenarios
Figure 4 for The CHiME-7 DASR Challenge: Distant Meeting Transcription with Multiple Devices in Diverse Scenarios
Viaarxiv icon

SURT 2.0: Advances in Transducer-based Multi-talker Speech Recognition

Add code
Bookmark button
Alert button
Jun 18, 2023
Desh Raj, Daniel Povey, Sanjeev Khudanpur

Figure 1 for SURT 2.0: Advances in Transducer-based Multi-talker Speech Recognition
Figure 2 for SURT 2.0: Advances in Transducer-based Multi-talker Speech Recognition
Figure 3 for SURT 2.0: Advances in Transducer-based Multi-talker Speech Recognition
Figure 4 for SURT 2.0: Advances in Transducer-based Multi-talker Speech Recognition
Viaarxiv icon

GPU-accelerated Guided Source Separation for Meeting Transcription

Add code
Bookmark button
Alert button
Dec 10, 2022
Desh Raj, Daniel Povey, Sanjeev Khudanpur

Figure 1 for GPU-accelerated Guided Source Separation for Meeting Transcription
Figure 2 for GPU-accelerated Guided Source Separation for Meeting Transcription
Figure 3 for GPU-accelerated Guided Source Separation for Meeting Transcription
Figure 4 for GPU-accelerated Guided Source Separation for Meeting Transcription
Viaarxiv icon

Adapting self-supervised models to multi-talker speech recognition using speaker embeddings

Add code
Bookmark button
Alert button
Nov 01, 2022
Zili Huang, Desh Raj, Paola García, Sanjeev Khudanpur

Figure 1 for Adapting self-supervised models to multi-talker speech recognition using speaker embeddings
Figure 2 for Adapting self-supervised models to multi-talker speech recognition using speaker embeddings
Figure 3 for Adapting self-supervised models to multi-talker speech recognition using speaker embeddings
Figure 4 for Adapting self-supervised models to multi-talker speech recognition using speaker embeddings
Viaarxiv icon