Alert button
Picture for Mohan Li

Mohan Li

Alert button

DiaLoc: An Iterative Approach to Embodied Dialog Localization

Add code
Bookmark button
Alert button
Mar 11, 2024
Chao Zhang, Mohan Li, Ignas Budvytis, Stephan Liwicki

Figure 1 for DiaLoc: An Iterative Approach to Embodied Dialog Localization
Figure 2 for DiaLoc: An Iterative Approach to Embodied Dialog Localization
Figure 3 for DiaLoc: An Iterative Approach to Embodied Dialog Localization
Figure 4 for DiaLoc: An Iterative Approach to Embodied Dialog Localization
Viaarxiv icon

Self-regularised Minimum Latency Training for Streaming Transformer-based Speech Recognition

Add code
Bookmark button
Alert button
Apr 24, 2023
Mohan Li, Rama Doddipatla, Catalin Zorila

Figure 1 for Self-regularised Minimum Latency Training for Streaming Transformer-based Speech Recognition
Figure 2 for Self-regularised Minimum Latency Training for Streaming Transformer-based Speech Recognition
Figure 3 for Self-regularised Minimum Latency Training for Streaming Transformer-based Speech Recognition
Figure 4 for Self-regularised Minimum Latency Training for Streaming Transformer-based Speech Recognition
Viaarxiv icon

Non-autoregressive End-to-end Approaches for Joint Automatic Speech Recognition and Spoken Language Understanding

Add code
Bookmark button
Alert button
Apr 21, 2023
Mohan Li, Rama Doddipatla

Figure 1 for Non-autoregressive End-to-end Approaches for Joint Automatic Speech Recognition and Spoken Language Understanding
Figure 2 for Non-autoregressive End-to-end Approaches for Joint Automatic Speech Recognition and Spoken Language Understanding
Figure 3 for Non-autoregressive End-to-end Approaches for Joint Automatic Speech Recognition and Spoken Language Understanding
Viaarxiv icon

Multiple-hypothesis RNN-T Loss for Unsupervised Fine-tuning and Self-training of Neural Transducer

Add code
Bookmark button
Alert button
Jul 29, 2022
Cong-Thanh Do, Mohan Li, Rama Doddipatla

Figure 1 for Multiple-hypothesis RNN-T Loss for Unsupervised Fine-tuning and Self-training of Neural Transducer
Figure 2 for Multiple-hypothesis RNN-T Loss for Unsupervised Fine-tuning and Self-training of Neural Transducer
Figure 3 for Multiple-hypothesis RNN-T Loss for Unsupervised Fine-tuning and Self-training of Neural Transducer
Figure 4 for Multiple-hypothesis RNN-T Loss for Unsupervised Fine-tuning and Self-training of Neural Transducer
Viaarxiv icon

Transformer-based Streaming ASR with Cumulative Attention

Add code
Bookmark button
Alert button
Mar 11, 2022
Mohan Li, Shucong Zhang, Catalin Zorila, Rama Doddipatla

Figure 1 for Transformer-based Streaming ASR with Cumulative Attention
Figure 2 for Transformer-based Streaming ASR with Cumulative Attention
Figure 3 for Transformer-based Streaming ASR with Cumulative Attention
Figure 4 for Transformer-based Streaming ASR with Cumulative Attention
Viaarxiv icon

Head-synchronous Decoding for Transformer-based Streaming ASR

Add code
Bookmark button
Alert button
Apr 26, 2021
Mohan Li, Catalin Zorila, Rama Doddipatla

Figure 1 for Head-synchronous Decoding for Transformer-based Streaming ASR
Figure 2 for Head-synchronous Decoding for Transformer-based Streaming ASR
Figure 3 for Head-synchronous Decoding for Transformer-based Streaming ASR
Figure 4 for Head-synchronous Decoding for Transformer-based Streaming ASR
Viaarxiv icon

End-to-end Speech Recognition with Adaptive Computation Steps

Add code
Bookmark button
Alert button
Sep 26, 2018
Mohan Li, Min Liu, Masanori Hattori

Figure 1 for End-to-end Speech Recognition with Adaptive Computation Steps
Figure 2 for End-to-end Speech Recognition with Adaptive Computation Steps
Figure 3 for End-to-end Speech Recognition with Adaptive Computation Steps
Viaarxiv icon