Picture for Ramon Sanabria

Ramon Sanabria

Looking Enhances Listening: Recovering Missing Speech Using Images

Add code
Feb 13, 2020
Figure 1 for Looking Enhances Listening: Recovering Missing Speech Using Images
Figure 2 for Looking Enhances Listening: Recovering Missing Speech Using Images
Figure 3 for Looking Enhances Listening: Recovering Missing Speech Using Images
Figure 4 for Looking Enhances Listening: Recovering Missing Speech Using Images
Viaarxiv icon

Multitask Learning For Different Subword Segmentations In Neural Machine Translation

Add code
Oct 27, 2019
Figure 1 for Multitask Learning For Different Subword Segmentations In Neural Machine Translation
Figure 2 for Multitask Learning For Different Subword Segmentations In Neural Machine Translation
Figure 3 for Multitask Learning For Different Subword Segmentations In Neural Machine Translation
Figure 4 for Multitask Learning For Different Subword Segmentations In Neural Machine Translation
Viaarxiv icon

Analyzing Utility of Visual Context in Multimodal Speech Recognition Under Noisy Conditions

Add code
Jun 30, 2019
Figure 1 for Analyzing Utility of Visual Context in Multimodal Speech Recognition Under Noisy Conditions
Figure 2 for Analyzing Utility of Visual Context in Multimodal Speech Recognition Under Noisy Conditions
Figure 3 for Analyzing Utility of Visual Context in Multimodal Speech Recognition Under Noisy Conditions
Figure 4 for Analyzing Utility of Visual Context in Multimodal Speech Recognition Under Noisy Conditions
Viaarxiv icon

Multimodal Grounding for Sequence-to-Sequence Speech Recognition

Add code
Nov 09, 2018
Figure 1 for Multimodal Grounding for Sequence-to-Sequence Speech Recognition
Figure 2 for Multimodal Grounding for Sequence-to-Sequence Speech Recognition
Figure 3 for Multimodal Grounding for Sequence-to-Sequence Speech Recognition
Figure 4 for Multimodal Grounding for Sequence-to-Sequence Speech Recognition
Viaarxiv icon

How2: A Large-scale Dataset for Multimodal Language Understanding

Add code
Nov 01, 2018
Figure 1 for How2: A Large-scale Dataset for Multimodal Language Understanding
Figure 2 for How2: A Large-scale Dataset for Multimodal Language Understanding
Figure 3 for How2: A Large-scale Dataset for Multimodal Language Understanding
Figure 4 for How2: A Large-scale Dataset for Multimodal Language Understanding
Viaarxiv icon

Hierarchical Multi Task Learning With CTC

Add code
Jul 25, 2018
Figure 1 for Hierarchical Multi Task Learning With CTC
Figure 2 for Hierarchical Multi Task Learning With CTC
Figure 3 for Hierarchical Multi Task Learning With CTC
Figure 4 for Hierarchical Multi Task Learning With CTC
Viaarxiv icon

Subword and Crossword Units for CTC Acoustic Models

Add code
Jun 18, 2018
Figure 1 for Subword and Crossword Units for CTC Acoustic Models
Figure 2 for Subword and Crossword Units for CTC Acoustic Models
Figure 3 for Subword and Crossword Units for CTC Acoustic Models
Figure 4 for Subword and Crossword Units for CTC Acoustic Models
Viaarxiv icon

End-to-End Multimodal Speech Recognition

Add code
Apr 25, 2018
Figure 1 for End-to-End Multimodal Speech Recognition
Viaarxiv icon

Sequence-based Multi-lingual Low Resource Speech Recognition

Add code
Mar 06, 2018
Figure 1 for Sequence-based Multi-lingual Low Resource Speech Recognition
Viaarxiv icon

Comparison of Decoding Strategies for CTC Acoustic Models

Add code
Aug 15, 2017
Figure 1 for Comparison of Decoding Strategies for CTC Acoustic Models
Figure 2 for Comparison of Decoding Strategies for CTC Acoustic Models
Figure 3 for Comparison of Decoding Strategies for CTC Acoustic Models
Viaarxiv icon