Alert button
Picture for Martin Radfar

Martin Radfar

Alert button

Lookahead When It Matters: Adaptive Non-causal Transformers for Streaming Neural Transducers

Add code
Bookmark button
Alert button
May 09, 2023
Grant P. Strimel, Yi Xie, Brian King, Martin Radfar, Ariya Rastrow, Athanasios Mouchtaris

Figure 1 for Lookahead When It Matters: Adaptive Non-causal Transformers for Streaming Neural Transducers
Figure 2 for Lookahead When It Matters: Adaptive Non-causal Transformers for Streaming Neural Transducers
Figure 3 for Lookahead When It Matters: Adaptive Non-causal Transformers for Streaming Neural Transducers
Figure 4 for Lookahead When It Matters: Adaptive Non-causal Transformers for Streaming Neural Transducers
Viaarxiv icon

End-to-end spoken language understanding using joint CTC loss and self-supervised, pretrained acoustic encoders

Add code
Bookmark button
Alert button
May 04, 2023
Jixuan Wang, Martin Radfar, Kai Wei, Clement Chung

Figure 1 for End-to-end spoken language understanding using joint CTC loss and self-supervised, pretrained acoustic encoders
Figure 2 for End-to-end spoken language understanding using joint CTC loss and self-supervised, pretrained acoustic encoders
Figure 3 for End-to-end spoken language understanding using joint CTC loss and self-supervised, pretrained acoustic encoders
Figure 4 for End-to-end spoken language understanding using joint CTC loss and self-supervised, pretrained acoustic encoders
Viaarxiv icon

Leveraging Redundancy in Multiple Audio Signals for Far-Field Speech Recognition

Add code
Bookmark button
Alert button
Mar 01, 2023
Feng-Ju Chang, Anastasios Alexandridis, Rupak Vignesh Swaminathan, Martin Radfar, Harish Mallidi, Maurizio Omologo, Athanasios Mouchtaris, Brian King, Roland Maas

Figure 1 for Leveraging Redundancy in Multiple Audio Signals for Far-Field Speech Recognition
Figure 2 for Leveraging Redundancy in Multiple Audio Signals for Far-Field Speech Recognition
Figure 3 for Leveraging Redundancy in Multiple Audio Signals for Far-Field Speech Recognition
Figure 4 for Leveraging Redundancy in Multiple Audio Signals for Far-Field Speech Recognition
Viaarxiv icon

Sub-8-bit quantization for on-device speech recognition: a regularization-free approach

Add code
Bookmark button
Alert button
Oct 17, 2022
Kai Zhen, Martin Radfar, Hieu Duy Nguyen, Grant P. Strimel, Nathan Susanj, Athanasios Mouchtaris

Figure 1 for Sub-8-bit quantization for on-device speech recognition: a regularization-free approach
Figure 2 for Sub-8-bit quantization for on-device speech recognition: a regularization-free approach
Figure 3 for Sub-8-bit quantization for on-device speech recognition: a regularization-free approach
Figure 4 for Sub-8-bit quantization for on-device speech recognition: a regularization-free approach
Viaarxiv icon

ConvRNN-T: Convolutional Augmented Recurrent Neural Network Transducers for Streaming Speech Recognition

Add code
Bookmark button
Alert button
Sep 29, 2022
Martin Radfar, Rohit Barnwal, Rupak Vignesh Swaminathan, Feng-Ju Chang, Grant P. Strimel, Nathan Susanj, Athanasios Mouchtaris

Figure 1 for ConvRNN-T: Convolutional Augmented Recurrent Neural Network Transducers for Streaming Speech Recognition
Figure 2 for ConvRNN-T: Convolutional Augmented Recurrent Neural Network Transducers for Streaming Speech Recognition
Figure 3 for ConvRNN-T: Convolutional Augmented Recurrent Neural Network Transducers for Streaming Speech Recognition
Figure 4 for ConvRNN-T: Convolutional Augmented Recurrent Neural Network Transducers for Streaming Speech Recognition
Viaarxiv icon

Compute Cost Amortized Transformer for Streaming ASR

Add code
Bookmark button
Alert button
Jul 05, 2022
Yi Xie, Jonathan Macoskey, Martin Radfar, Feng-Ju Chang, Brian King, Ariya Rastrow, Athanasios Mouchtaris, Grant P. Strimel

Figure 1 for Compute Cost Amortized Transformer for Streaming ASR
Figure 2 for Compute Cost Amortized Transformer for Streaming ASR
Figure 3 for Compute Cost Amortized Transformer for Streaming ASR
Figure 4 for Compute Cost Amortized Transformer for Streaming ASR
Viaarxiv icon

A neural prosody encoder for end-ro-end dialogue act classification

Add code
Bookmark button
Alert button
May 11, 2022
Kai Wei, Dillon Knox, Martin Radfar, Thanh Tran, Markus Muller, Grant P. Strimel, Nathan Susanj, Athanasios Mouchtaris, Maurizio Omologo

Figure 1 for A neural prosody encoder for end-ro-end dialogue act classification
Figure 2 for A neural prosody encoder for end-ro-end dialogue act classification
Figure 3 for A neural prosody encoder for end-ro-end dialogue act classification
Figure 4 for A neural prosody encoder for end-ro-end dialogue act classification
Viaarxiv icon

Multi-task RNN-T with Semantic Decoder for Streamable Spoken Language Understanding

Add code
Bookmark button
Alert button
Apr 01, 2022
Xuandi Fu, Feng-Ju Chang, Martin Radfar, Kai Wei, Jing Liu, Grant P. Strimel, Kanthashree Mysore Sathyendra

Figure 1 for Multi-task RNN-T with Semantic Decoder for Streamable Spoken Language Understanding
Figure 2 for Multi-task RNN-T with Semantic Decoder for Streamable Spoken Language Understanding
Figure 3 for Multi-task RNN-T with Semantic Decoder for Streamable Spoken Language Understanding
Figure 4 for Multi-task RNN-T with Semantic Decoder for Streamable Spoken Language Understanding
Viaarxiv icon

Context-Aware Transformer Transducer for Speech Recognition

Add code
Bookmark button
Alert button
Nov 05, 2021
Feng-Ju Chang, Jing Liu, Martin Radfar, Athanasios Mouchtaris, Maurizio Omologo, Ariya Rastrow, Siegfried Kunzmann

Figure 1 for Context-Aware Transformer Transducer for Speech Recognition
Figure 2 for Context-Aware Transformer Transducer for Speech Recognition
Figure 3 for Context-Aware Transformer Transducer for Speech Recognition
Figure 4 for Context-Aware Transformer Transducer for Speech Recognition
Viaarxiv icon

Speech Emotion Recognition Using Quaternion Convolutional Neural Networks

Add code
Bookmark button
Alert button
Oct 31, 2021
Aneesh Muppidi, Martin Radfar

Figure 1 for Speech Emotion Recognition Using Quaternion Convolutional Neural Networks
Figure 2 for Speech Emotion Recognition Using Quaternion Convolutional Neural Networks
Figure 3 for Speech Emotion Recognition Using Quaternion Convolutional Neural Networks
Figure 4 for Speech Emotion Recognition Using Quaternion Convolutional Neural Networks
Viaarxiv icon