Alert button
Picture for Nanxin Chen

Nanxin Chen

Alert button

A Quantum Kernel Learning Approach to Acoustic Modeling for Spoken Command Recognition

Add code
Bookmark button
Alert button
Nov 02, 2022
Chao-Han Huck Yang, Bo Li, Yu Zhang, Nanxin Chen, Tara N. Sainath, Sabato Marco Siniscalchi, Chin-Hui Lee

Figure 1 for A Quantum Kernel Learning Approach to Acoustic Modeling for Spoken Command Recognition
Figure 2 for A Quantum Kernel Learning Approach to Acoustic Modeling for Spoken Command Recognition
Figure 3 for A Quantum Kernel Learning Approach to Acoustic Modeling for Spoken Command Recognition
Figure 4 for A Quantum Kernel Learning Approach to Acoustic Modeling for Spoken Command Recognition
Viaarxiv icon

Residual Adapters for Few-Shot Text-to-Speech Speaker Adaptation

Add code
Bookmark button
Alert button
Oct 28, 2022
Nobuyuki Morioka, Heiga Zen, Nanxin Chen, Yu Zhang, Yifan Ding

Figure 1 for Residual Adapters for Few-Shot Text-to-Speech Speaker Adaptation
Figure 2 for Residual Adapters for Few-Shot Text-to-Speech Speaker Adaptation
Figure 3 for Residual Adapters for Few-Shot Text-to-Speech Speaker Adaptation
Figure 4 for Residual Adapters for Few-Shot Text-to-Speech Speaker Adaptation
Viaarxiv icon

Maestro-U: Leveraging joint speech-text representation learning for zero supervised speech ASR

Add code
Bookmark button
Alert button
Oct 18, 2022
Zhehuai Chen, Ankur Bapna, Andrew Rosenberg, Yu Zhang, Bhuvana Ramabhadran, Pedro Moreno, Nanxin Chen

Figure 1 for Maestro-U: Leveraging joint speech-text representation learning for zero supervised speech ASR
Figure 2 for Maestro-U: Leveraging joint speech-text representation learning for zero supervised speech ASR
Figure 3 for Maestro-U: Leveraging joint speech-text representation learning for zero supervised speech ASR
Figure 4 for Maestro-U: Leveraging joint speech-text representation learning for zero supervised speech ASR
Viaarxiv icon

SpecGrad: Diffusion Probabilistic Model based Neural Vocoder with Adaptive Noise Spectral Shaping

Add code
Bookmark button
Alert button
Mar 31, 2022
Yuma Koizumi, Heiga Zen, Kohei Yatabe, Nanxin Chen, Michiel Bacchiani

Figure 1 for SpecGrad: Diffusion Probabilistic Model based Neural Vocoder with Adaptive Noise Spectral Shaping
Figure 2 for SpecGrad: Diffusion Probabilistic Model based Neural Vocoder with Adaptive Noise Spectral Shaping
Figure 3 for SpecGrad: Diffusion Probabilistic Model based Neural Vocoder with Adaptive Noise Spectral Shaping
Figure 4 for SpecGrad: Diffusion Probabilistic Model based Neural Vocoder with Adaptive Noise Spectral Shaping
Viaarxiv icon

A Comparative Study on Non-Autoregressive Modelings for Speech-to-Text Generation

Add code
Bookmark button
Alert button
Oct 11, 2021
Yosuke Higuchi, Nanxin Chen, Yuya Fujita, Hirofumi Inaguma, Tatsuya Komatsu, Jaesong Lee, Jumon Nozaki, Tianzi Wang, Shinji Watanabe

Figure 1 for A Comparative Study on Non-Autoregressive Modelings for Speech-to-Text Generation
Figure 2 for A Comparative Study on Non-Autoregressive Modelings for Speech-to-Text Generation
Figure 3 for A Comparative Study on Non-Autoregressive Modelings for Speech-to-Text Generation
Figure 4 for A Comparative Study on Non-Autoregressive Modelings for Speech-to-Text Generation
Viaarxiv icon

WaveGrad 2: Iterative Refinement for Text-to-Speech Synthesis

Add code
Bookmark button
Alert button
Jun 19, 2021
Nanxin Chen, Yu Zhang, Heiga Zen, Ron J. Weiss, Mohammad Norouzi, Najim Dehak, William Chan

Figure 1 for WaveGrad 2: Iterative Refinement for Text-to-Speech Synthesis
Figure 2 for WaveGrad 2: Iterative Refinement for Text-to-Speech Synthesis
Figure 3 for WaveGrad 2: Iterative Refinement for Text-to-Speech Synthesis
Figure 4 for WaveGrad 2: Iterative Refinement for Text-to-Speech Synthesis
Viaarxiv icon

Focus on the present: a regularization method for the ASR source-target attention layer

Add code
Bookmark button
Alert button
Nov 02, 2020
Nanxin Chen, Piotr Żelasko, Jesús Villalba, Najim Dehak

Figure 1 for Focus on the present: a regularization method for the ASR source-target attention layer
Figure 2 for Focus on the present: a regularization method for the ASR source-target attention layer
Figure 3 for Focus on the present: a regularization method for the ASR source-target attention layer
Figure 4 for Focus on the present: a regularization method for the ASR source-target attention layer
Viaarxiv icon

WaveGrad: Estimating Gradients for Waveform Generation

Add code
Bookmark button
Alert button
Sep 02, 2020
Nanxin Chen, Yu Zhang, Heiga Zen, Ron J. Weiss, Mohammad Norouzi, William Chan

Figure 1 for WaveGrad: Estimating Gradients for Waveform Generation
Figure 2 for WaveGrad: Estimating Gradients for Waveform Generation
Figure 3 for WaveGrad: Estimating Gradients for Waveform Generation
Figure 4 for WaveGrad: Estimating Gradients for Waveform Generation
Viaarxiv icon

Robust Training of Vector Quantized Bottleneck Models

Add code
Bookmark button
Alert button
May 18, 2020
Adrian Łańcucki, Jan Chorowski, Guillaume Sanchez, Ricard Marxer, Nanxin Chen, Hans J. G. A. Dolfing, Sameer Khurana, Tanel Alumäe, Antoine Laurent

Figure 1 for Robust Training of Vector Quantized Bottleneck Models
Figure 2 for Robust Training of Vector Quantized Bottleneck Models
Figure 3 for Robust Training of Vector Quantized Bottleneck Models
Figure 4 for Robust Training of Vector Quantized Bottleneck Models
Viaarxiv icon

x-vectors meet emotions: A study on dependencies between emotion and speaker recognition

Add code
Bookmark button
Alert button
Feb 12, 2020
Raghavendra Pappagari, Tianzi Wang, Jesus Villalba, Nanxin Chen, Najim Dehak

Figure 1 for x-vectors meet emotions: A study on dependencies between emotion and speaker recognition
Figure 2 for x-vectors meet emotions: A study on dependencies between emotion and speaker recognition
Figure 3 for x-vectors meet emotions: A study on dependencies between emotion and speaker recognition
Figure 4 for x-vectors meet emotions: A study on dependencies between emotion and speaker recognition
Viaarxiv icon