Picture for Xiaohu Qie

Xiaohu Qie

OmniZoomer: Learning to Move and Zoom in on Sphere at High-Resolution

Add code
Aug 19, 2023
Figure 1 for OmniZoomer: Learning to Move and Zoom in on Sphere at High-Resolution
Figure 2 for OmniZoomer: Learning to Move and Zoom in on Sphere at High-Resolution
Figure 3 for OmniZoomer: Learning to Move and Zoom in on Sphere at High-Resolution
Figure 4 for OmniZoomer: Learning to Move and Zoom in on Sphere at High-Resolution
Viaarxiv icon

HOSNeRF: Dynamic Human-Object-Scene Neural Radiance Fields from a Single Video

Add code
Apr 24, 2023
Figure 1 for HOSNeRF: Dynamic Human-Object-Scene Neural Radiance Fields from a Single Video
Figure 2 for HOSNeRF: Dynamic Human-Object-Scene Neural Radiance Fields from a Single Video
Figure 3 for HOSNeRF: Dynamic Human-Object-Scene Neural Radiance Fields from a Single Video
Figure 4 for HOSNeRF: Dynamic Human-Object-Scene Neural Radiance Fields from a Single Video
Viaarxiv icon

MasaCtrl: Tuning-Free Mutual Self-Attention Control for Consistent Image Synthesis and Editing

Add code
Apr 17, 2023
Figure 1 for MasaCtrl: Tuning-Free Mutual Self-Attention Control for Consistent Image Synthesis and Editing
Figure 2 for MasaCtrl: Tuning-Free Mutual Self-Attention Control for Consistent Image Synthesis and Editing
Figure 3 for MasaCtrl: Tuning-Free Mutual Self-Attention Control for Consistent Image Synthesis and Editing
Figure 4 for MasaCtrl: Tuning-Free Mutual Self-Attention Control for Consistent Image Synthesis and Editing
Viaarxiv icon

VMesh: Hybrid Volume-Mesh Representation for Efficient View Synthesis

Add code
Mar 28, 2023
Figure 1 for VMesh: Hybrid Volume-Mesh Representation for Efficient View Synthesis
Figure 2 for VMesh: Hybrid Volume-Mesh Representation for Efficient View Synthesis
Figure 3 for VMesh: Hybrid Volume-Mesh Representation for Efficient View Synthesis
Figure 4 for VMesh: Hybrid Volume-Mesh Representation for Efficient View Synthesis
Viaarxiv icon

Accelerating Vision-Language Pretraining with Free Language Modeling

Add code
Mar 24, 2023
Figure 1 for Accelerating Vision-Language Pretraining with Free Language Modeling
Figure 2 for Accelerating Vision-Language Pretraining with Free Language Modeling
Figure 3 for Accelerating Vision-Language Pretraining with Free Language Modeling
Figure 4 for Accelerating Vision-Language Pretraining with Free Language Modeling
Viaarxiv icon

T2I-Adapter: Learning Adapters to Dig out More Controllable Ability for Text-to-Image Diffusion Models

Add code
Feb 16, 2023
Figure 1 for T2I-Adapter: Learning Adapters to Dig out More Controllable Ability for Text-to-Image Diffusion Models
Figure 2 for T2I-Adapter: Learning Adapters to Dig out More Controllable Ability for Text-to-Image Diffusion Models
Figure 3 for T2I-Adapter: Learning Adapters to Dig out More Controllable Ability for Text-to-Image Diffusion Models
Figure 4 for T2I-Adapter: Learning Adapters to Dig out More Controllable Ability for Text-to-Image Diffusion Models
Viaarxiv icon

Masked Visual Reconstruction in Language Semantic Space

Add code
Jan 17, 2023
Figure 1 for Masked Visual Reconstruction in Language Semantic Space
Figure 2 for Masked Visual Reconstruction in Language Semantic Space
Figure 3 for Masked Visual Reconstruction in Language Semantic Space
Figure 4 for Masked Visual Reconstruction in Language Semantic Space
Viaarxiv icon

Dream3D: Zero-Shot Text-to-3D Synthesis Using 3D Shape Prior and Text-to-Image Diffusion Models

Add code
Dec 28, 2022
Figure 1 for Dream3D: Zero-Shot Text-to-3D Synthesis Using 3D Shape Prior and Text-to-Image Diffusion Models
Figure 2 for Dream3D: Zero-Shot Text-to-3D Synthesis Using 3D Shape Prior and Text-to-Image Diffusion Models
Figure 3 for Dream3D: Zero-Shot Text-to-3D Synthesis Using 3D Shape Prior and Text-to-Image Diffusion Models
Figure 4 for Dream3D: Zero-Shot Text-to-3D Synthesis Using 3D Shape Prior and Text-to-Image Diffusion Models
Viaarxiv icon

Tune-A-Video: One-Shot Tuning of Image Diffusion Models for Text-to-Video Generation

Add code
Dec 22, 2022
Figure 1 for Tune-A-Video: One-Shot Tuning of Image Diffusion Models for Text-to-Video Generation
Figure 2 for Tune-A-Video: One-Shot Tuning of Image Diffusion Models for Text-to-Video Generation
Figure 3 for Tune-A-Video: One-Shot Tuning of Image Diffusion Models for Text-to-Video Generation
Figure 4 for Tune-A-Video: One-Shot Tuning of Image Diffusion Models for Text-to-Video Generation
Viaarxiv icon

Rethinking the Objectives of Vector-Quantized Tokenizers for Image Synthesis

Add code
Dec 06, 2022
Figure 1 for Rethinking the Objectives of Vector-Quantized Tokenizers for Image Synthesis
Figure 2 for Rethinking the Objectives of Vector-Quantized Tokenizers for Image Synthesis
Figure 3 for Rethinking the Objectives of Vector-Quantized Tokenizers for Image Synthesis
Figure 4 for Rethinking the Objectives of Vector-Quantized Tokenizers for Image Synthesis
Viaarxiv icon