Picture for Herman Kamper

Herman Kamper

Keyword localisation in untranscribed speech using visually grounded speech models

Add code
Feb 02, 2022
Figure 1 for Keyword localisation in untranscribed speech using visually grounded speech models
Figure 2 for Keyword localisation in untranscribed speech using visually grounded speech models
Figure 3 for Keyword localisation in untranscribed speech using visually grounded speech models
Figure 4 for Keyword localisation in untranscribed speech using visually grounded speech models
Viaarxiv icon

Towards Learning to Speak and Hear Through Multi-Agent Communication over a Continuous Acoustic Channel

Add code
Nov 04, 2021
Figure 1 for Towards Learning to Speak and Hear Through Multi-Agent Communication over a Continuous Acoustic Channel
Figure 2 for Towards Learning to Speak and Hear Through Multi-Agent Communication over a Continuous Acoustic Channel
Figure 3 for Towards Learning to Speak and Hear Through Multi-Agent Communication over a Continuous Acoustic Channel
Figure 4 for Towards Learning to Speak and Hear Through Multi-Agent Communication over a Continuous Acoustic Channel
Viaarxiv icon

Voice Conversion Can Improve ASR in Very Low-Resource Settings

Add code
Nov 04, 2021
Figure 1 for Voice Conversion Can Improve ASR in Very Low-Resource Settings
Figure 2 for Voice Conversion Can Improve ASR in Very Low-Resource Settings
Figure 3 for Voice Conversion Can Improve ASR in Very Low-Resource Settings
Figure 4 for Voice Conversion Can Improve ASR in Very Low-Resource Settings
Viaarxiv icon

A Comparison of Discrete and Soft Speech Units for Improved Voice Conversion

Add code
Nov 03, 2021
Figure 1 for A Comparison of Discrete and Soft Speech Units for Improved Voice Conversion
Figure 2 for A Comparison of Discrete and Soft Speech Units for Improved Voice Conversion
Figure 3 for A Comparison of Discrete and Soft Speech Units for Improved Voice Conversion
Figure 4 for A Comparison of Discrete and Soft Speech Units for Improved Voice Conversion
Viaarxiv icon

Feature learning for efficient ASR-free keyword spotting in low-resource languages

Add code
Aug 13, 2021
Figure 1 for Feature learning for efficient ASR-free keyword spotting in low-resource languages
Figure 2 for Feature learning for efficient ASR-free keyword spotting in low-resource languages
Figure 3 for Feature learning for efficient ASR-free keyword spotting in low-resource languages
Figure 4 for Feature learning for efficient ASR-free keyword spotting in low-resource languages
Viaarxiv icon

Analyzing Speaker Information in Self-Supervised Models to Improve Zero-Resource Speech Processing

Add code
Aug 02, 2021
Figure 1 for Analyzing Speaker Information in Self-Supervised Models to Improve Zero-Resource Speech Processing
Figure 2 for Analyzing Speaker Information in Self-Supervised Models to Improve Zero-Resource Speech Processing
Figure 3 for Analyzing Speaker Information in Self-Supervised Models to Improve Zero-Resource Speech Processing
Figure 4 for Analyzing Speaker Information in Self-Supervised Models to Improve Zero-Resource Speech Processing
Viaarxiv icon

Mava: a research framework for distributed multi-agent reinforcement learning

Add code
Jul 03, 2021
Figure 1 for Mava: a research framework for distributed multi-agent reinforcement learning
Figure 2 for Mava: a research framework for distributed multi-agent reinforcement learning
Figure 3 for Mava: a research framework for distributed multi-agent reinforcement learning
Figure 4 for Mava: a research framework for distributed multi-agent reinforcement learning
Viaarxiv icon

Multilingual transfer of acoustic word embeddings improves when training on languages related to the target zero-resource language

Add code
Jun 24, 2021
Figure 1 for Multilingual transfer of acoustic word embeddings improves when training on languages related to the target zero-resource language
Figure 2 for Multilingual transfer of acoustic word embeddings improves when training on languages related to the target zero-resource language
Figure 3 for Multilingual transfer of acoustic word embeddings improves when training on languages related to the target zero-resource language
Figure 4 for Multilingual transfer of acoustic word embeddings improves when training on languages related to the target zero-resource language
Viaarxiv icon

Attention-Based Keyword Localisation in Speech using Visual Grounding

Add code
Jun 23, 2021
Figure 1 for Attention-Based Keyword Localisation in Speech using Visual Grounding
Figure 2 for Attention-Based Keyword Localisation in Speech using Visual Grounding
Figure 3 for Attention-Based Keyword Localisation in Speech using Visual Grounding
Figure 4 for Attention-Based Keyword Localisation in Speech using Visual Grounding
Viaarxiv icon

StarGAN-ZSVC: Towards Zero-Shot Voice Conversion in Low-Resource Contexts

Add code
May 31, 2021
Figure 1 for StarGAN-ZSVC: Towards Zero-Shot Voice Conversion in Low-Resource Contexts
Figure 2 for StarGAN-ZSVC: Towards Zero-Shot Voice Conversion in Low-Resource Contexts
Figure 3 for StarGAN-ZSVC: Towards Zero-Shot Voice Conversion in Low-Resource Contexts
Figure 4 for StarGAN-ZSVC: Towards Zero-Shot Voice Conversion in Low-Resource Contexts
Viaarxiv icon