Picture for Florian Metze

Florian Metze

Subword and Crossword Units for CTC Acoustic Models

Add code
Jun 18, 2018
Figure 1 for Subword and Crossword Units for CTC Acoustic Models
Figure 2 for Subword and Crossword Units for CTC Acoustic Models
Figure 3 for Subword and Crossword Units for CTC Acoustic Models
Figure 4 for Subword and Crossword Units for CTC Acoustic Models
Viaarxiv icon

End-to-End Multimodal Speech Recognition

Add code
Apr 25, 2018
Figure 1 for End-to-End Multimodal Speech Recognition
Viaarxiv icon

Sequence-based Multi-lingual Low Resource Speech Recognition

Add code
Mar 06, 2018
Figure 1 for Sequence-based Multi-lingual Low Resource Speech Recognition
Viaarxiv icon

Annotating High-Level Structures of Short Stories and Personal Anecdotes

Add code
Feb 27, 2018
Figure 1 for Annotating High-Level Structures of Short Stories and Personal Anecdotes
Figure 2 for Annotating High-Level Structures of Short Stories and Personal Anecdotes
Figure 3 for Annotating High-Level Structures of Short Stories and Personal Anecdotes
Figure 4 for Annotating High-Level Structures of Short Stories and Personal Anecdotes
Viaarxiv icon

Linguistic unit discovery from multi-modal inputs in unwritten languages: Summary of the "Speaking Rosetta" JSALT 2017 Workshop

Add code
Feb 14, 2018
Figure 1 for Linguistic unit discovery from multi-modal inputs in unwritten languages: Summary of the "Speaking Rosetta" JSALT 2017 Workshop
Figure 2 for Linguistic unit discovery from multi-modal inputs in unwritten languages: Summary of the "Speaking Rosetta" JSALT 2017 Workshop
Figure 3 for Linguistic unit discovery from multi-modal inputs in unwritten languages: Summary of the "Speaking Rosetta" JSALT 2017 Workshop
Viaarxiv icon

Visual Features for Context-Aware Speech Recognition

Add code
Dec 01, 2017
Figure 1 for Visual Features for Context-Aware Speech Recognition
Figure 2 for Visual Features for Context-Aware Speech Recognition
Figure 3 for Visual Features for Context-Aware Speech Recognition
Viaarxiv icon

Comparison of Decoding Strategies for CTC Acoustic Models

Add code
Aug 15, 2017
Figure 1 for Comparison of Decoding Strategies for CTC Acoustic Models
Figure 2 for Comparison of Decoding Strategies for CTC Acoustic Models
Figure 3 for Comparison of Decoding Strategies for CTC Acoustic Models
Viaarxiv icon

A Comparison of deep learning methods for environmental sound

Add code
Mar 20, 2017
Figure 1 for A Comparison of deep learning methods for environmental sound
Figure 2 for A Comparison of deep learning methods for environmental sound
Figure 3 for A Comparison of deep learning methods for environmental sound
Figure 4 for A Comparison of deep learning methods for environmental sound
Viaarxiv icon

Robust end-to-end deep audiovisual speech recognition

Add code
Nov 21, 2016
Figure 1 for Robust end-to-end deep audiovisual speech recognition
Figure 2 for Robust end-to-end deep audiovisual speech recognition
Figure 3 for Robust end-to-end deep audiovisual speech recognition
Figure 4 for Robust end-to-end deep audiovisual speech recognition
Viaarxiv icon

EESEN: End-to-End Speech Recognition using Deep RNN Models and WFST-based Decoding

Add code
Oct 18, 2015
Figure 1 for EESEN: End-to-End Speech Recognition using Deep RNN Models and WFST-based Decoding
Figure 2 for EESEN: End-to-End Speech Recognition using Deep RNN Models and WFST-based Decoding
Figure 3 for EESEN: End-to-End Speech Recognition using Deep RNN Models and WFST-based Decoding
Figure 4 for EESEN: End-to-End Speech Recognition using Deep RNN Models and WFST-based Decoding
Viaarxiv icon