Picture for Andros Tjandra

Andros Tjandra

Sequence-to-Sequence Learning via Attention Transfer for Incremental Speech Recognition

Add code
Nov 04, 2020
Figure 1 for Sequence-to-Sequence Learning via Attention Transfer for Incremental Speech Recognition
Figure 2 for Sequence-to-Sequence Learning via Attention Transfer for Incremental Speech Recognition
Figure 3 for Sequence-to-Sequence Learning via Attention Transfer for Incremental Speech Recognition
Figure 4 for Sequence-to-Sequence Learning via Attention Transfer for Incremental Speech Recognition
Viaarxiv icon

Incremental Machine Speech Chain Towards Enabling Listening while Speaking in Real-time

Add code
Nov 04, 2020
Figure 1 for Incremental Machine Speech Chain Towards Enabling Listening while Speaking in Real-time
Figure 2 for Incremental Machine Speech Chain Towards Enabling Listening while Speaking in Real-time
Figure 3 for Incremental Machine Speech Chain Towards Enabling Listening while Speaking in Real-time
Viaarxiv icon

Augmenting Images for ASR and TTS through Single-loop and Dual-loop Multimodal Chain Framework

Add code
Nov 04, 2020
Figure 1 for Augmenting Images for ASR and TTS through Single-loop and Dual-loop Multimodal Chain Framework
Figure 2 for Augmenting Images for ASR and TTS through Single-loop and Dual-loop Multimodal Chain Framework
Figure 3 for Augmenting Images for ASR and TTS through Single-loop and Dual-loop Multimodal Chain Framework
Figure 4 for Augmenting Images for ASR and TTS through Single-loop and Dual-loop Multimodal Chain Framework
Viaarxiv icon

Unsupervised Learning of Disentangled Speech Content and Style Representation

Add code
Oct 24, 2020
Figure 1 for Unsupervised Learning of Disentangled Speech Content and Style Representation
Figure 2 for Unsupervised Learning of Disentangled Speech Content and Style Representation
Figure 3 for Unsupervised Learning of Disentangled Speech Content and Style Representation
Figure 4 for Unsupervised Learning of Disentangled Speech Content and Style Representation
Viaarxiv icon

Transformer VQ-VAE for Unsupervised Unit Discovery and Speech Synthesis: ZeroSpeech 2020 Challenge

Add code
May 24, 2020
Figure 1 for Transformer VQ-VAE for Unsupervised Unit Discovery and Speech Synthesis: ZeroSpeech 2020 Challenge
Figure 2 for Transformer VQ-VAE for Unsupervised Unit Discovery and Speech Synthesis: ZeroSpeech 2020 Challenge
Figure 3 for Transformer VQ-VAE for Unsupervised Unit Discovery and Speech Synthesis: ZeroSpeech 2020 Challenge
Figure 4 for Transformer VQ-VAE for Unsupervised Unit Discovery and Speech Synthesis: ZeroSpeech 2020 Challenge
Viaarxiv icon

Deja-vu: Double Feature Presentation in Deep Transformer Networks

Add code
Oct 23, 2019
Figure 1 for Deja-vu: Double Feature Presentation in Deep Transformer Networks
Figure 2 for Deja-vu: Double Feature Presentation in Deep Transformer Networks
Figure 3 for Deja-vu: Double Feature Presentation in Deep Transformer Networks
Figure 4 for Deja-vu: Double Feature Presentation in Deep Transformer Networks
Viaarxiv icon

Transformer-based Acoustic Modeling for Hybrid Speech Recognition

Add code
Oct 22, 2019
Figure 1 for Transformer-based Acoustic Modeling for Hybrid Speech Recognition
Figure 2 for Transformer-based Acoustic Modeling for Hybrid Speech Recognition
Figure 3 for Transformer-based Acoustic Modeling for Hybrid Speech Recognition
Figure 4 for Transformer-based Acoustic Modeling for Hybrid Speech Recognition
Viaarxiv icon

Speech-to-speech Translation between Untranscribed Unknown Languages

Add code
Oct 05, 2019
Figure 1 for Speech-to-speech Translation between Untranscribed Unknown Languages
Figure 2 for Speech-to-speech Translation between Untranscribed Unknown Languages
Figure 3 for Speech-to-speech Translation between Untranscribed Unknown Languages
Figure 4 for Speech-to-speech Translation between Untranscribed Unknown Languages
Viaarxiv icon

From Speech Chain to Multimodal Chain: Leveraging Cross-modal Data Augmentation for Semi-supervised Learning

Add code
Jun 03, 2019
Figure 1 for From Speech Chain to Multimodal Chain: Leveraging Cross-modal Data Augmentation for Semi-supervised Learning
Figure 2 for From Speech Chain to Multimodal Chain: Leveraging Cross-modal Data Augmentation for Semi-supervised Learning
Figure 3 for From Speech Chain to Multimodal Chain: Leveraging Cross-modal Data Augmentation for Semi-supervised Learning
Figure 4 for From Speech Chain to Multimodal Chain: Leveraging Cross-modal Data Augmentation for Semi-supervised Learning
Viaarxiv icon

VQVAE Unsupervised Unit Discovery and Multi-scale Code2Spec Inverter for Zerospeech Challenge 2019

Add code
May 29, 2019
Figure 1 for VQVAE Unsupervised Unit Discovery and Multi-scale Code2Spec Inverter for Zerospeech Challenge 2019
Figure 2 for VQVAE Unsupervised Unit Discovery and Multi-scale Code2Spec Inverter for Zerospeech Challenge 2019
Figure 3 for VQVAE Unsupervised Unit Discovery and Multi-scale Code2Spec Inverter for Zerospeech Challenge 2019
Figure 4 for VQVAE Unsupervised Unit Discovery and Multi-scale Code2Spec Inverter for Zerospeech Challenge 2019
Viaarxiv icon