Picture for Wanli Ouyang

Wanli Ouyang

School of Electrical and Information Engineering, The University of Sydney, Australia

DATAP-SfM: Dynamic-Aware Tracking Any Point for Robust Structure from Motion in the Wild

Add code
Nov 20, 2024
Figure 1 for DATAP-SfM: Dynamic-Aware Tracking Any Point for Robust Structure from Motion in the Wild
Figure 2 for DATAP-SfM: Dynamic-Aware Tracking Any Point for Robust Structure from Motion in the Wild
Figure 3 for DATAP-SfM: Dynamic-Aware Tracking Any Point for Robust Structure from Motion in the Wild
Figure 4 for DATAP-SfM: Dynamic-Aware Tracking Any Point for Robust Structure from Motion in the Wild
Viaarxiv icon

FengWu-W2S: A deep learning model for seamless weather-to-subseasonal forecast of global atmosphere

Add code
Nov 15, 2024
Viaarxiv icon

WeatherGFM: Learning A Weather Generalist Foundation Model via In-context Learning

Add code
Nov 08, 2024
Figure 1 for WeatherGFM: Learning A Weather Generalist Foundation Model via In-context Learning
Figure 2 for WeatherGFM: Learning A Weather Generalist Foundation Model via In-context Learning
Figure 3 for WeatherGFM: Learning A Weather Generalist Foundation Model via In-context Learning
Figure 4 for WeatherGFM: Learning A Weather Generalist Foundation Model via In-context Learning
Viaarxiv icon

DiffPano: Scalable and Consistent Text to Panorama Generation with Spherical Epipolar-Aware Diffusion

Add code
Oct 31, 2024
Figure 1 for DiffPano: Scalable and Consistent Text to Panorama Generation with Spherical Epipolar-Aware Diffusion
Figure 2 for DiffPano: Scalable and Consistent Text to Panorama Generation with Spherical Epipolar-Aware Diffusion
Figure 3 for DiffPano: Scalable and Consistent Text to Panorama Generation with Spherical Epipolar-Aware Diffusion
Figure 4 for DiffPano: Scalable and Consistent Text to Panorama Generation with Spherical Epipolar-Aware Diffusion
Viaarxiv icon

MotionGPT-2: A General-Purpose Motion-Language Model for Motion Generation and Understanding

Add code
Oct 29, 2024
Figure 1 for MotionGPT-2: A General-Purpose Motion-Language Model for Motion Generation and Understanding
Figure 2 for MotionGPT-2: A General-Purpose Motion-Language Model for Motion Generation and Understanding
Figure 3 for MotionGPT-2: A General-Purpose Motion-Language Model for Motion Generation and Understanding
Figure 4 for MotionGPT-2: A General-Purpose Motion-Language Model for Motion Generation and Understanding
Viaarxiv icon

Where Am I and What Will I See: An Auto-Regressive Model for Spatial Localization and View Prediction

Add code
Oct 24, 2024
Figure 1 for Where Am I and What Will I See: An Auto-Regressive Model for Spatial Localization and View Prediction
Figure 2 for Where Am I and What Will I See: An Auto-Regressive Model for Spatial Localization and View Prediction
Figure 3 for Where Am I and What Will I See: An Auto-Regressive Model for Spatial Localization and View Prediction
Figure 4 for Where Am I and What Will I See: An Auto-Regressive Model for Spatial Localization and View Prediction
Viaarxiv icon

WorldSimBench: Towards Video Generation Models as World Simulators

Add code
Oct 23, 2024
Figure 1 for WorldSimBench: Towards Video Generation Models as World Simulators
Figure 2 for WorldSimBench: Towards Video Generation Models as World Simulators
Figure 3 for WorldSimBench: Towards Video Generation Models as World Simulators
Figure 4 for WorldSimBench: Towards Video Generation Models as World Simulators
Viaarxiv icon

CrystalX: Ultra-Precision Crystal Structure Resolution and Error Correction Using Deep Learning

Add code
Oct 17, 2024
Viaarxiv icon

FiTv2: Scalable and Improved Flexible Vision Transformer for Diffusion Model

Add code
Oct 17, 2024
Figure 1 for FiTv2: Scalable and Improved Flexible Vision Transformer for Diffusion Model
Figure 2 for FiTv2: Scalable and Improved Flexible Vision Transformer for Diffusion Model
Figure 3 for FiTv2: Scalable and Improved Flexible Vision Transformer for Diffusion Model
Figure 4 for FiTv2: Scalable and Improved Flexible Vision Transformer for Diffusion Model
Viaarxiv icon

A CLIP-Powered Framework for Robust and Generalizable Data Selection

Add code
Oct 15, 2024
Figure 1 for A CLIP-Powered Framework for Robust and Generalizable Data Selection
Figure 2 for A CLIP-Powered Framework for Robust and Generalizable Data Selection
Figure 3 for A CLIP-Powered Framework for Robust and Generalizable Data Selection
Figure 4 for A CLIP-Powered Framework for Robust and Generalizable Data Selection
Viaarxiv icon