Picture for Junichi Yamagishi

Junichi Yamagishi

Spoof Diarization: "What Spoofed When" in Partially Spoofed Audio

Add code
Jun 12, 2024
Figure 1 for Spoof Diarization: "What Spoofed When" in Partially Spoofed Audio
Figure 2 for Spoof Diarization: "What Spoofed When" in Partially Spoofed Audio
Figure 3 for Spoof Diarization: "What Spoofed When" in Partially Spoofed Audio
Figure 4 for Spoof Diarization: "What Spoofed When" in Partially Spoofed Audio
Viaarxiv icon

To what extent can ASV systems naturally defend against spoofing attacks?

Add code
Jun 08, 2024
Viaarxiv icon

Exploring Self-Supervised Vision Transformers for Deepfake Detection: A Comparative Analysis

Add code
May 01, 2024
Viaarxiv icon

The VoicePrivacy 2024 Challenge Evaluation Plan

Add code
Apr 03, 2024
Figure 1 for The VoicePrivacy 2024 Challenge Evaluation Plan
Figure 2 for The VoicePrivacy 2024 Challenge Evaluation Plan
Figure 3 for The VoicePrivacy 2024 Challenge Evaluation Plan
Figure 4 for The VoicePrivacy 2024 Challenge Evaluation Plan
Viaarxiv icon

Bridging Textual and Tabular Worlds for Fact Verification: A Lightweight, Attention-Based Model

Add code
Mar 26, 2024
Viaarxiv icon

Uncertainty as a Predictor: Leveraging Self-Supervised Learning for Zero-Shot MOS Prediction

Add code
Dec 25, 2023
Figure 1 for Uncertainty as a Predictor: Leveraging Self-Supervised Learning for Zero-Shot MOS Prediction
Figure 2 for Uncertainty as a Predictor: Leveraging Self-Supervised Learning for Zero-Shot MOS Prediction
Figure 3 for Uncertainty as a Predictor: Leveraging Self-Supervised Learning for Zero-Shot MOS Prediction
Figure 4 for Uncertainty as a Predictor: Leveraging Self-Supervised Learning for Zero-Shot MOS Prediction
Viaarxiv icon

ZMM-TTS: Zero-shot Multilingual and Multispeaker Speech Synthesis Conditioned on Self-supervised Discrete Speech Representations

Add code
Dec 22, 2023
Figure 1 for ZMM-TTS: Zero-shot Multilingual and Multispeaker Speech Synthesis Conditioned on Self-supervised Discrete Speech Representations
Figure 2 for ZMM-TTS: Zero-shot Multilingual and Multispeaker Speech Synthesis Conditioned on Self-supervised Discrete Speech Representations
Figure 3 for ZMM-TTS: Zero-shot Multilingual and Multispeaker Speech Synthesis Conditioned on Self-supervised Discrete Speech Representations
Figure 4 for ZMM-TTS: Zero-shot Multilingual and Multispeaker Speech Synthesis Conditioned on Self-supervised Discrete Speech Representations
Viaarxiv icon

Speaker-Text Retrieval via Contrastive Learning

Add code
Dec 11, 2023
Viaarxiv icon

XFEVER: Exploring Fact Verification across Languages

Add code
Oct 25, 2023
Viaarxiv icon

Partial Rank Similarity Minimization Method for Quality MOS Prediction of Unseen Speech Synthesis Systems in Zero-Shot and Semi-supervised setting

Add code
Oct 08, 2023
Figure 1 for Partial Rank Similarity Minimization Method for Quality MOS Prediction of Unseen Speech Synthesis Systems in Zero-Shot and Semi-supervised setting
Figure 2 for Partial Rank Similarity Minimization Method for Quality MOS Prediction of Unseen Speech Synthesis Systems in Zero-Shot and Semi-supervised setting
Figure 3 for Partial Rank Similarity Minimization Method for Quality MOS Prediction of Unseen Speech Synthesis Systems in Zero-Shot and Semi-supervised setting
Figure 4 for Partial Rank Similarity Minimization Method for Quality MOS Prediction of Unseen Speech Synthesis Systems in Zero-Shot and Semi-supervised setting
Viaarxiv icon