Alert button
Picture for Rio Yokota

Rio Yokota

Alert button

Variational Learning is Effective for Large Deep Networks

Feb 27, 2024
Yuesong Shen, Nico Daheim, Bai Cong, Peter Nickl, Gian Maria Marconi, Clement Bazan, Rio Yokota, Iryna Gurevych, Daniel Cremers, Mohammad Emtiyaz Khan, Thomas Möllenhoff

Viaarxiv icon

SegRCDB: Semantic Segmentation via Formula-Driven Supervised Learning

Sep 29, 2023
Risa Shinoda, Ryo Hayamizu, Kodai Nakashima, Nakamasa Inoue, Rio Yokota, Hirokatsu Kataoka

Viaarxiv icon

Pre-training Vision Transformers with Very Limited Synthesized Images

Jul 31, 2023
Ryo Nakamura, Hirokatsu Kataoka, Sora Takashima, Edgar Josafat Martinez Noriega, Rio Yokota, Nakamasa Inoue

Figure 1 for Pre-training Vision Transformers with Very Limited Synthesized Images
Figure 2 for Pre-training Vision Transformers with Very Limited Synthesized Images
Figure 3 for Pre-training Vision Transformers with Very Limited Synthesized Images
Figure 4 for Pre-training Vision Transformers with Very Limited Synthesized Images
Viaarxiv icon

ASDL: A Unified Interface for Gradient Preconditioning in PyTorch

May 08, 2023
Kazuki Osawa, Satoki Ishikawa, Rio Yokota, Shigang Li, Torsten Hoefler

Figure 1 for ASDL: A Unified Interface for Gradient Preconditioning in PyTorch
Figure 2 for ASDL: A Unified Interface for Gradient Preconditioning in PyTorch
Figure 3 for ASDL: A Unified Interface for Gradient Preconditioning in PyTorch
Figure 4 for ASDL: A Unified Interface for Gradient Preconditioning in PyTorch
Viaarxiv icon

Visual Atoms: Pre-training Vision Transformers with Sinusoidal Waves

Mar 02, 2023
Sora Takashima, Ryo Hayamizu, Nakamasa Inoue, Hirokatsu Kataoka, Rio Yokota

Figure 1 for Visual Atoms: Pre-training Vision Transformers with Sinusoidal Waves
Figure 2 for Visual Atoms: Pre-training Vision Transformers with Sinusoidal Waves
Figure 3 for Visual Atoms: Pre-training Vision Transformers with Sinusoidal Waves
Figure 4 for Visual Atoms: Pre-training Vision Transformers with Sinusoidal Waves
Viaarxiv icon

Empirical Study on Optimizer Selection for Out-of-Distribution Generalization

Nov 18, 2022
Hiroki Naganuma, Kartik Ahuja, Shiro Takagi, Tetsuya Motokawa, Rio Yokota, Kohta Ishikawa, Ikuro Sato, Ioannis Mitliagkas

Figure 1 for Empirical Study on Optimizer Selection for Out-of-Distribution Generalization
Figure 2 for Empirical Study on Optimizer Selection for Out-of-Distribution Generalization
Figure 3 for Empirical Study on Optimizer Selection for Out-of-Distribution Generalization
Figure 4 for Empirical Study on Optimizer Selection for Out-of-Distribution Generalization
Viaarxiv icon

Informative Sample-Aware Proxy for Deep Metric Learning

Nov 18, 2022
Aoyu Li, Ikuro Sato, Kohta Ishikawa, Rei Kawakami, Rio Yokota

Figure 1 for Informative Sample-Aware Proxy for Deep Metric Learning
Figure 2 for Informative Sample-Aware Proxy for Deep Metric Learning
Figure 3 for Informative Sample-Aware Proxy for Deep Metric Learning
Figure 4 for Informative Sample-Aware Proxy for Deep Metric Learning
Viaarxiv icon

Replacing Labeled Real-image Datasets with Auto-generated Contours

Jun 18, 2022
Hirokatsu Kataoka, Ryo Hayamizu, Ryosuke Yamada, Kodai Nakashima, Sora Takashima, Xinyu Zhang, Edgar Josafat Martinez-Noriega, Nakamasa Inoue, Rio Yokota

Figure 1 for Replacing Labeled Real-image Datasets with Auto-generated Contours
Figure 2 for Replacing Labeled Real-image Datasets with Auto-generated Contours
Figure 3 for Replacing Labeled Real-image Datasets with Auto-generated Contours
Figure 4 for Replacing Labeled Real-image Datasets with Auto-generated Contours
Viaarxiv icon

OPIRL: Sample Efficient Off-Policy Inverse Reinforcement Learning via Distribution Matching

Sep 09, 2021
Hana Hoshino, Kei Ota, Asako Kanezaki, Rio Yokota

Figure 1 for OPIRL: Sample Efficient Off-Policy Inverse Reinforcement Learning via Distribution Matching
Figure 2 for OPIRL: Sample Efficient Off-Policy Inverse Reinforcement Learning via Distribution Matching
Figure 3 for OPIRL: Sample Efficient Off-Policy Inverse Reinforcement Learning via Distribution Matching
Figure 4 for OPIRL: Sample Efficient Off-Policy Inverse Reinforcement Learning via Distribution Matching
Viaarxiv icon

RePOSE: Real-Time Iterative Rendering and Refinement for 6D Object Pose Estimation

Apr 01, 2021
Shun Iwase, Xingyu Liu, Rawal Khirodkar, Rio Yokota, Kris M. Kitani

Figure 1 for RePOSE: Real-Time Iterative Rendering and Refinement for 6D Object Pose Estimation
Figure 2 for RePOSE: Real-Time Iterative Rendering and Refinement for 6D Object Pose Estimation
Figure 3 for RePOSE: Real-Time Iterative Rendering and Refinement for 6D Object Pose Estimation
Figure 4 for RePOSE: Real-Time Iterative Rendering and Refinement for 6D Object Pose Estimation
Viaarxiv icon