Alert button
Picture for Shinji Watanabe

Shinji Watanabe

Alert button

Structured Pruning of Self-Supervised Pre-trained Models for Speech Recognition and Understanding

Add code
Bookmark button
Alert button
Feb 27, 2023
Yifan Peng, Kwangyoun Kim, Felix Wu, Prashant Sridhar, Shinji Watanabe

Figure 1 for Structured Pruning of Self-Supervised Pre-trained Models for Speech Recognition and Understanding
Figure 2 for Structured Pruning of Self-Supervised Pre-trained Models for Speech Recognition and Understanding
Figure 3 for Structured Pruning of Self-Supervised Pre-trained Models for Speech Recognition and Understanding
Figure 4 for Structured Pruning of Self-Supervised Pre-trained Models for Speech Recognition and Understanding
Viaarxiv icon

Improving Massively Multilingual ASR With Auxiliary CTC Objectives

Add code
Bookmark button
Alert button
Feb 27, 2023
William Chen, Brian Yan, Jiatong Shi, Yifan Peng, Soumi Maiti, Shinji Watanabe

Figure 1 for Improving Massively Multilingual ASR With Auxiliary CTC Objectives
Figure 2 for Improving Massively Multilingual ASR With Auxiliary CTC Objectives
Figure 3 for Improving Massively Multilingual ASR With Auxiliary CTC Objectives
Figure 4 for Improving Massively Multilingual ASR With Auxiliary CTC Objectives
Viaarxiv icon

PAAPLoss: A Phonetic-Aligned Acoustic Parameter Loss for Speech Enhancement

Add code
Bookmark button
Alert button
Feb 16, 2023
Muqiao Yang, Joseph Konan, David Bick, Yunyang Zeng, Shuo Han, Anurag Kumar, Shinji Watanabe, Bhiksha Raj

Figure 1 for PAAPLoss: A Phonetic-Aligned Acoustic Parameter Loss for Speech Enhancement
Figure 2 for PAAPLoss: A Phonetic-Aligned Acoustic Parameter Loss for Speech Enhancement
Figure 3 for PAAPLoss: A Phonetic-Aligned Acoustic Parameter Loss for Speech Enhancement
Viaarxiv icon

TAPLoss: A Temporal Acoustic Parameter Loss for Speech Enhancement

Add code
Bookmark button
Alert button
Feb 16, 2023
Yunyang Zeng, Joseph Konan, Shuo Han, David Bick, Muqiao Yang, Anurag Kumar, Shinji Watanabe, Bhiksha Raj

Figure 1 for TAPLoss: A Temporal Acoustic Parameter Loss for Speech Enhancement
Figure 2 for TAPLoss: A Temporal Acoustic Parameter Loss for Speech Enhancement
Figure 3 for TAPLoss: A Temporal Acoustic Parameter Loss for Speech Enhancement
Figure 4 for TAPLoss: A Temporal Acoustic Parameter Loss for Speech Enhancement
Viaarxiv icon

Multi-Channel Target Speaker Extraction with Refinement: The WavLab Submission to the Second Clarity Enhancement Challenge

Add code
Bookmark button
Alert button
Feb 15, 2023
Samuele Cornell, Zhong-Qiu Wang, Yoshiki Masuyama, Shinji Watanabe, Manuel Pariente, Nobutaka Ono

Figure 1 for Multi-Channel Target Speaker Extraction with Refinement: The WavLab Submission to the Second Clarity Enhancement Challenge
Figure 2 for Multi-Channel Target Speaker Extraction with Refinement: The WavLab Submission to the Second Clarity Enhancement Challenge
Figure 3 for Multi-Channel Target Speaker Extraction with Refinement: The WavLab Submission to the Second Clarity Enhancement Challenge
Viaarxiv icon

Speaker-Independent Acoustic-to-Articulatory Speech Inversion

Add code
Bookmark button
Alert button
Feb 14, 2023
Peter Wu, Li-Wei Chen, Cheol Jun Cho, Shinji Watanabe, Louis Goldstein, Alan W Black, Gopala K. Anumanchipalli

Figure 1 for Speaker-Independent Acoustic-to-Articulatory Speech Inversion
Figure 2 for Speaker-Independent Acoustic-to-Articulatory Speech Inversion
Figure 3 for Speaker-Independent Acoustic-to-Articulatory Speech Inversion
Figure 4 for Speaker-Independent Acoustic-to-Articulatory Speech Inversion
Viaarxiv icon

A Vector Quantized Approach for Text to Speech Synthesis on Real-World Spontaneous Speech

Add code
Bookmark button
Alert button
Feb 08, 2023
Li-Wei Chen, Shinji Watanabe, Alexander Rudnicky

Figure 1 for A Vector Quantized Approach for Text to Speech Synthesis on Real-World Spontaneous Speech
Figure 2 for A Vector Quantized Approach for Text to Speech Synthesis on Real-World Spontaneous Speech
Figure 3 for A Vector Quantized Approach for Text to Speech Synthesis on Real-World Spontaneous Speech
Figure 4 for A Vector Quantized Approach for Text to Speech Synthesis on Real-World Spontaneous Speech
Viaarxiv icon

Learning to Speak from Text: Zero-Shot Multilingual Text-to-Speech with Unsupervised Text Pretraining

Add code
Bookmark button
Alert button
Feb 05, 2023
Takaaki Saeki, Soumi Maiti, Xinjian Li, Shinji Watanabe, Shinnosuke Takamichi, Hiroshi Saruwatari

Figure 1 for Learning to Speak from Text: Zero-Shot Multilingual Text-to-Speech with Unsupervised Text Pretraining
Figure 2 for Learning to Speak from Text: Zero-Shot Multilingual Text-to-Speech with Unsupervised Text Pretraining
Figure 3 for Learning to Speak from Text: Zero-Shot Multilingual Text-to-Speech with Unsupervised Text Pretraining
Figure 4 for Learning to Speak from Text: Zero-Shot Multilingual Text-to-Speech with Unsupervised Text Pretraining
Viaarxiv icon