Picture for Yung-Sung Chuang

Yung-Sung Chuang

C2KD: Cross-Lingual Cross-Modal Knowledge Distillation for Multilingual Text-Video Retrieval

Add code
Oct 07, 2022
Figure 1 for C2KD: Cross-Lingual Cross-Modal Knowledge Distillation for Multilingual Text-Video Retrieval
Figure 2 for C2KD: Cross-Lingual Cross-Modal Knowledge Distillation for Multilingual Text-Video Retrieval
Figure 3 for C2KD: Cross-Lingual Cross-Modal Knowledge Distillation for Multilingual Text-Video Retrieval
Figure 4 for C2KD: Cross-Lingual Cross-Modal Knowledge Distillation for Multilingual Text-Video Retrieval
Viaarxiv icon

DiffCSE: Difference-based Contrastive Learning for Sentence Embeddings

Add code
Apr 21, 2022
Figure 1 for DiffCSE: Difference-based Contrastive Learning for Sentence Embeddings
Figure 2 for DiffCSE: Difference-based Contrastive Learning for Sentence Embeddings
Figure 3 for DiffCSE: Difference-based Contrastive Learning for Sentence Embeddings
Figure 4 for DiffCSE: Difference-based Contrastive Learning for Sentence Embeddings
Viaarxiv icon

DUAL: Discrete Spoken Unit Adaptive Learning for Textless Spoken Question Answering

Add code
Mar 26, 2022
Figure 1 for DUAL: Discrete Spoken Unit Adaptive Learning for Textless Spoken Question Answering
Figure 2 for DUAL: Discrete Spoken Unit Adaptive Learning for Textless Spoken Question Answering
Figure 3 for DUAL: Discrete Spoken Unit Adaptive Learning for Textless Spoken Question Answering
Figure 4 for DUAL: Discrete Spoken Unit Adaptive Learning for Textless Spoken Question Answering
Viaarxiv icon

On the Interplay Between Sparsity, Naturalness, Intelligibility, and Prosody in Speech Synthesis

Add code
Oct 04, 2021
Figure 1 for On the Interplay Between Sparsity, Naturalness, Intelligibility, and Prosody in Speech Synthesis
Figure 2 for On the Interplay Between Sparsity, Naturalness, Intelligibility, and Prosody in Speech Synthesis
Figure 3 for On the Interplay Between Sparsity, Naturalness, Intelligibility, and Prosody in Speech Synthesis
Figure 4 for On the Interplay Between Sparsity, Naturalness, Intelligibility, and Prosody in Speech Synthesis
Viaarxiv icon

Mitigating Biases in Toxic Language Detection through Invariant Rationalization

Add code
Jun 14, 2021
Figure 1 for Mitigating Biases in Toxic Language Detection through Invariant Rationalization
Figure 2 for Mitigating Biases in Toxic Language Detection through Invariant Rationalization
Figure 3 for Mitigating Biases in Toxic Language Detection through Invariant Rationalization
Viaarxiv icon

PARP: Prune, Adjust and Re-Prune for Self-Supervised Speech Recognition

Add code
Jun 10, 2021
Figure 1 for PARP: Prune, Adjust and Re-Prune for Self-Supervised Speech Recognition
Figure 2 for PARP: Prune, Adjust and Re-Prune for Self-Supervised Speech Recognition
Figure 3 for PARP: Prune, Adjust and Re-Prune for Self-Supervised Speech Recognition
Figure 4 for PARP: Prune, Adjust and Re-Prune for Self-Supervised Speech Recognition
Viaarxiv icon

Meta-learning for downstream aware and agnostic pretraining

Add code
Jun 06, 2021
Viaarxiv icon

Investigating the Reordering Capability in CTC-based Non-Autoregressive End-to-End Speech Translation

Add code
May 11, 2021
Figure 1 for Investigating the Reordering Capability in CTC-based Non-Autoregressive End-to-End Speech Translation
Figure 2 for Investigating the Reordering Capability in CTC-based Non-Autoregressive End-to-End Speech Translation
Figure 3 for Investigating the Reordering Capability in CTC-based Non-Autoregressive End-to-End Speech Translation
Figure 4 for Investigating the Reordering Capability in CTC-based Non-Autoregressive End-to-End Speech Translation
Viaarxiv icon

SUPERB: Speech processing Universal PERformance Benchmark

Add code
May 03, 2021
Figure 1 for SUPERB: Speech processing Universal PERformance Benchmark
Figure 2 for SUPERB: Speech processing Universal PERformance Benchmark
Viaarxiv icon

Semi-Supervised Spoken Language Understanding via Self-Supervised Speech and Language Model Pretraining

Add code
Oct 26, 2020
Figure 1 for Semi-Supervised Spoken Language Understanding via Self-Supervised Speech and Language Model Pretraining
Figure 2 for Semi-Supervised Spoken Language Understanding via Self-Supervised Speech and Language Model Pretraining
Figure 3 for Semi-Supervised Spoken Language Understanding via Self-Supervised Speech and Language Model Pretraining
Figure 4 for Semi-Supervised Spoken Language Understanding via Self-Supervised Speech and Language Model Pretraining
Viaarxiv icon