Picture for Satoshi Nakamura

Satoshi Nakamura

Subspace-based Set Operations on a Pre-trained Word Embedding Space

Add code
Oct 24, 2022
Viaarxiv icon

Actor-identified Spatiotemporal Action Detection -- Detecting Who Is Doing What in Videos

Add code
Aug 27, 2022
Figure 1 for Actor-identified Spatiotemporal Action Detection -- Detecting Who Is Doing What in Videos
Figure 2 for Actor-identified Spatiotemporal Action Detection -- Detecting Who Is Doing What in Videos
Figure 3 for Actor-identified Spatiotemporal Action Detection -- Detecting Who Is Doing What in Videos
Figure 4 for Actor-identified Spatiotemporal Action Detection -- Detecting Who Is Doing What in Videos
Viaarxiv icon

USB: A Unified Semi-supervised Learning Benchmark

Add code
Aug 12, 2022
Figure 1 for USB: A Unified Semi-supervised Learning Benchmark
Figure 2 for USB: A Unified Semi-supervised Learning Benchmark
Figure 3 for USB: A Unified Semi-supervised Learning Benchmark
Figure 4 for USB: A Unified Semi-supervised Learning Benchmark
Viaarxiv icon

Speech Artifact Removal from EEG Recordings of Spoken Word Production with Tensor Decomposition

Add code
Jun 01, 2022
Figure 1 for Speech Artifact Removal from EEG Recordings of Spoken Word Production with Tensor Decomposition
Figure 2 for Speech Artifact Removal from EEG Recordings of Spoken Word Production with Tensor Decomposition
Figure 3 for Speech Artifact Removal from EEG Recordings of Spoken Word Production with Tensor Decomposition
Figure 4 for Speech Artifact Removal from EEG Recordings of Spoken Word Production with Tensor Decomposition
Viaarxiv icon

Improved Consistency Training for Semi-Supervised Sequence-to-Sequence ASR via Speech Chain Reconstruction and Self-Transcribing

Add code
May 14, 2022
Figure 1 for Improved Consistency Training for Semi-Supervised Sequence-to-Sequence ASR via Speech Chain Reconstruction and Self-Transcribing
Figure 2 for Improved Consistency Training for Semi-Supervised Sequence-to-Sequence ASR via Speech Chain Reconstruction and Self-Transcribing
Figure 3 for Improved Consistency Training for Semi-Supervised Sequence-to-Sequence ASR via Speech Chain Reconstruction and Self-Transcribing
Viaarxiv icon

Representing `how you say' with `what you say': English corpus of focused speech and text reflecting corresponding implications

Add code
Mar 29, 2022
Figure 1 for Representing `how you say' with `what you say': English corpus of focused speech and text reflecting corresponding implications
Figure 2 for Representing `how you say' with `what you say': English corpus of focused speech and text reflecting corresponding implications
Figure 3 for Representing `how you say' with `what you say': English corpus of focused speech and text reflecting corresponding implications
Figure 4 for Representing `how you say' with `what you say': English corpus of focused speech and text reflecting corresponding implications
Viaarxiv icon

Speech Segmentation Optimization using Segmented Bilingual Speech Corpus for End-to-end Speech Translation

Add code
Mar 29, 2022
Figure 1 for Speech Segmentation Optimization using Segmented Bilingual Speech Corpus for End-to-end Speech Translation
Figure 2 for Speech Segmentation Optimization using Segmented Bilingual Speech Corpus for End-to-end Speech Translation
Figure 3 for Speech Segmentation Optimization using Segmented Bilingual Speech Corpus for End-to-end Speech Translation
Figure 4 for Speech Segmentation Optimization using Segmented Bilingual Speech Corpus for End-to-end Speech Translation
Viaarxiv icon

Applying Syntax$\unicode{x2013}$Prosody Mapping Hypothesis and Prosodic Well-Formedness Constraints to Neural Sequence-to-Sequence Speech Synthesis

Add code
Mar 29, 2022
Figure 1 for Applying Syntax$\unicode{x2013}$Prosody Mapping Hypothesis and Prosodic Well-Formedness Constraints to Neural Sequence-to-Sequence Speech Synthesis
Figure 2 for Applying Syntax$\unicode{x2013}$Prosody Mapping Hypothesis and Prosodic Well-Formedness Constraints to Neural Sequence-to-Sequence Speech Synthesis
Figure 3 for Applying Syntax$\unicode{x2013}$Prosody Mapping Hypothesis and Prosodic Well-Formedness Constraints to Neural Sequence-to-Sequence Speech Synthesis
Figure 4 for Applying Syntax$\unicode{x2013}$Prosody Mapping Hypothesis and Prosodic Well-Formedness Constraints to Neural Sequence-to-Sequence Speech Synthesis
Viaarxiv icon

Simultaneous Neural Machine Translation with Constituent Label Prediction

Add code
Oct 26, 2021
Figure 1 for Simultaneous Neural Machine Translation with Constituent Label Prediction
Figure 2 for Simultaneous Neural Machine Translation with Constituent Label Prediction
Figure 3 for Simultaneous Neural Machine Translation with Constituent Label Prediction
Figure 4 for Simultaneous Neural Machine Translation with Constituent Label Prediction
Viaarxiv icon

Using Perturbed Length-aware Positional Encoding for Non-autoregressive Neural Machine Translation

Add code
Jul 29, 2021
Figure 1 for Using Perturbed Length-aware Positional Encoding for Non-autoregressive Neural Machine Translation
Figure 2 for Using Perturbed Length-aware Positional Encoding for Non-autoregressive Neural Machine Translation
Figure 3 for Using Perturbed Length-aware Positional Encoding for Non-autoregressive Neural Machine Translation
Figure 4 for Using Perturbed Length-aware Positional Encoding for Non-autoregressive Neural Machine Translation
Viaarxiv icon