Picture for Korin Richmond

Korin Richmond

CSTR

An Initial Investigation of Language Adaptation for TTS Systems under Low-resource Scenarios

Add code
Jun 13, 2024
Figure 1 for An Initial Investigation of Language Adaptation for TTS Systems under Low-resource Scenarios
Figure 2 for An Initial Investigation of Language Adaptation for TTS Systems under Low-resource Scenarios
Figure 3 for An Initial Investigation of Language Adaptation for TTS Systems under Low-resource Scenarios
Figure 4 for An Initial Investigation of Language Adaptation for TTS Systems under Low-resource Scenarios
Viaarxiv icon

ZMM-TTS: Zero-shot Multilingual and Multispeaker Speech Synthesis Conditioned on Self-supervised Discrete Speech Representations

Add code
Dec 22, 2023
Viaarxiv icon

Predicting pairwise preferences between TTS audio stimuli using parallel ratings data and anti-symmetric twin neural networks

Add code
Sep 22, 2022
Figure 1 for Predicting pairwise preferences between TTS audio stimuli using parallel ratings data and anti-symmetric twin neural networks
Figure 2 for Predicting pairwise preferences between TTS audio stimuli using parallel ratings data and anti-symmetric twin neural networks
Figure 3 for Predicting pairwise preferences between TTS audio stimuli using parallel ratings data and anti-symmetric twin neural networks
Figure 4 for Predicting pairwise preferences between TTS audio stimuli using parallel ratings data and anti-symmetric twin neural networks
Viaarxiv icon

Automatic audiovisual synchronisation for ultrasound tongue imaging

Add code
May 31, 2021
Figure 1 for Automatic audiovisual synchronisation for ultrasound tongue imaging
Figure 2 for Automatic audiovisual synchronisation for ultrasound tongue imaging
Figure 3 for Automatic audiovisual synchronisation for ultrasound tongue imaging
Figure 4 for Automatic audiovisual synchronisation for ultrasound tongue imaging
Viaarxiv icon

Silent versus modal multi-speaker speech recognition from ultrasound and video

Add code
Feb 27, 2021
Figure 1 for Silent versus modal multi-speaker speech recognition from ultrasound and video
Figure 2 for Silent versus modal multi-speaker speech recognition from ultrasound and video
Figure 3 for Silent versus modal multi-speaker speech recognition from ultrasound and video
Figure 4 for Silent versus modal multi-speaker speech recognition from ultrasound and video
Viaarxiv icon

Exploiting ultrasound tongue imaging for the automatic detection of speech articulation errors

Add code
Feb 27, 2021
Figure 1 for Exploiting ultrasound tongue imaging for the automatic detection of speech articulation errors
Figure 2 for Exploiting ultrasound tongue imaging for the automatic detection of speech articulation errors
Figure 3 for Exploiting ultrasound tongue imaging for the automatic detection of speech articulation errors
Figure 4 for Exploiting ultrasound tongue imaging for the automatic detection of speech articulation errors
Viaarxiv icon

TaL: a synchronised multi-speaker corpus of ultrasound tongue imaging, audio, and lip videos

Add code
Nov 19, 2020
Figure 1 for TaL: a synchronised multi-speaker corpus of ultrasound tongue imaging, audio, and lip videos
Figure 2 for TaL: a synchronised multi-speaker corpus of ultrasound tongue imaging, audio, and lip videos
Figure 3 for TaL: a synchronised multi-speaker corpus of ultrasound tongue imaging, audio, and lip videos
Figure 4 for TaL: a synchronised multi-speaker corpus of ultrasound tongue imaging, audio, and lip videos
Viaarxiv icon

Ultrasound tongue imaging for diarization and alignment of child speech therapy sessions

Add code
Aug 15, 2019
Figure 1 for Ultrasound tongue imaging for diarization and alignment of child speech therapy sessions
Figure 2 for Ultrasound tongue imaging for diarization and alignment of child speech therapy sessions
Figure 3 for Ultrasound tongue imaging for diarization and alignment of child speech therapy sessions
Figure 4 for Ultrasound tongue imaging for diarization and alignment of child speech therapy sessions
Viaarxiv icon

UltraSuite: A Repository of Ultrasound and Acoustic Data from Child Speech Therapy Sessions

Add code
Jul 01, 2019
Figure 1 for UltraSuite: A Repository of Ultrasound and Acoustic Data from Child Speech Therapy Sessions
Figure 2 for UltraSuite: A Repository of Ultrasound and Acoustic Data from Child Speech Therapy Sessions
Figure 3 for UltraSuite: A Repository of Ultrasound and Acoustic Data from Child Speech Therapy Sessions
Figure 4 for UltraSuite: A Repository of Ultrasound and Acoustic Data from Child Speech Therapy Sessions
Viaarxiv icon

Synchronising audio and ultrasound by learning cross-modal embeddings

Add code
Jul 01, 2019
Figure 1 for Synchronising audio and ultrasound by learning cross-modal embeddings
Figure 2 for Synchronising audio and ultrasound by learning cross-modal embeddings
Figure 3 for Synchronising audio and ultrasound by learning cross-modal embeddings
Figure 4 for Synchronising audio and ultrasound by learning cross-modal embeddings
Viaarxiv icon