Alert button
Picture for Herman Kamper

Herman Kamper

Alert button

Voice Conversion Can Improve ASR in Very Low-Resource Settings

Nov 04, 2021
Matthew Baas, Herman Kamper

Figure 1 for Voice Conversion Can Improve ASR in Very Low-Resource Settings
Figure 2 for Voice Conversion Can Improve ASR in Very Low-Resource Settings
Figure 3 for Voice Conversion Can Improve ASR in Very Low-Resource Settings
Figure 4 for Voice Conversion Can Improve ASR in Very Low-Resource Settings
Viaarxiv icon

A Comparison of Discrete and Soft Speech Units for Improved Voice Conversion

Nov 03, 2021
Benjamin van Niekerk, Marc-André Carbonneau, Julian Zaïdi, Mathew Baas, Hugo Seuté, Herman Kamper

Figure 1 for A Comparison of Discrete and Soft Speech Units for Improved Voice Conversion
Figure 2 for A Comparison of Discrete and Soft Speech Units for Improved Voice Conversion
Figure 3 for A Comparison of Discrete and Soft Speech Units for Improved Voice Conversion
Figure 4 for A Comparison of Discrete and Soft Speech Units for Improved Voice Conversion
Viaarxiv icon

Feature learning for efficient ASR-free keyword spotting in low-resource languages

Aug 13, 2021
Ewald van der Westhuizen, Herman Kamper, Raghav Menon, John Quinn, Thomas Niesler

Figure 1 for Feature learning for efficient ASR-free keyword spotting in low-resource languages
Figure 2 for Feature learning for efficient ASR-free keyword spotting in low-resource languages
Figure 3 for Feature learning for efficient ASR-free keyword spotting in low-resource languages
Figure 4 for Feature learning for efficient ASR-free keyword spotting in low-resource languages
Viaarxiv icon

Analyzing Speaker Information in Self-Supervised Models to Improve Zero-Resource Speech Processing

Aug 02, 2021
Benjamin van Niekerk, Leanne Nortje, Matthew Baas, Herman Kamper

Figure 1 for Analyzing Speaker Information in Self-Supervised Models to Improve Zero-Resource Speech Processing
Figure 2 for Analyzing Speaker Information in Self-Supervised Models to Improve Zero-Resource Speech Processing
Figure 3 for Analyzing Speaker Information in Self-Supervised Models to Improve Zero-Resource Speech Processing
Figure 4 for Analyzing Speaker Information in Self-Supervised Models to Improve Zero-Resource Speech Processing
Viaarxiv icon

Mava: a research framework for distributed multi-agent reinforcement learning

Jul 03, 2021
Arnu Pretorius, Kale-ab Tessera, Andries P. Smit, Claude Formanek, St John Grimbly, Kevin Eloff, Siphelele Danisa, Lawrence Francis, Jonathan Shock, Herman Kamper, Willie Brink, Herman Engelbrecht, Alexandre Laterre, Karim Beguir

Figure 1 for Mava: a research framework for distributed multi-agent reinforcement learning
Figure 2 for Mava: a research framework for distributed multi-agent reinforcement learning
Figure 3 for Mava: a research framework for distributed multi-agent reinforcement learning
Figure 4 for Mava: a research framework for distributed multi-agent reinforcement learning
Viaarxiv icon

Multilingual transfer of acoustic word embeddings improves when training on languages related to the target zero-resource language

Jun 24, 2021
Christiaan Jacobs, Herman Kamper

Figure 1 for Multilingual transfer of acoustic word embeddings improves when training on languages related to the target zero-resource language
Figure 2 for Multilingual transfer of acoustic word embeddings improves when training on languages related to the target zero-resource language
Figure 3 for Multilingual transfer of acoustic word embeddings improves when training on languages related to the target zero-resource language
Figure 4 for Multilingual transfer of acoustic word embeddings improves when training on languages related to the target zero-resource language
Viaarxiv icon

Attention-Based Keyword Localisation in Speech using Visual Grounding

Jun 23, 2021
Kayode Olaleye, Herman Kamper

Figure 1 for Attention-Based Keyword Localisation in Speech using Visual Grounding
Figure 2 for Attention-Based Keyword Localisation in Speech using Visual Grounding
Figure 3 for Attention-Based Keyword Localisation in Speech using Visual Grounding
Figure 4 for Attention-Based Keyword Localisation in Speech using Visual Grounding
Viaarxiv icon

StarGAN-ZSVC: Towards Zero-Shot Voice Conversion in Low-Resource Contexts

May 31, 2021
Matthew Baas, Herman Kamper

Figure 1 for StarGAN-ZSVC: Towards Zero-Shot Voice Conversion in Low-Resource Contexts
Figure 2 for StarGAN-ZSVC: Towards Zero-Shot Voice Conversion in Low-Resource Contexts
Figure 3 for StarGAN-ZSVC: Towards Zero-Shot Voice Conversion in Low-Resource Contexts
Figure 4 for StarGAN-ZSVC: Towards Zero-Shot Voice Conversion in Low-Resource Contexts
Viaarxiv icon

Acoustic word embeddings for zero-resource languages using self-supervised contrastive learning and multilingual adaptation

Mar 19, 2021
Christiaan Jacobs, Yevgen Matusevych, Herman Kamper

Figure 1 for Acoustic word embeddings for zero-resource languages using self-supervised contrastive learning and multilingual adaptation
Figure 2 for Acoustic word embeddings for zero-resource languages using self-supervised contrastive learning and multilingual adaptation
Figure 3 for Acoustic word embeddings for zero-resource languages using self-supervised contrastive learning and multilingual adaptation
Figure 4 for Acoustic word embeddings for zero-resource languages using self-supervised contrastive learning and multilingual adaptation
Viaarxiv icon