Picture for Jizhong Han

Jizhong Han

Customize your NeRF: Adaptive Source Driven 3D Scene Editing via Local-Global Iterative Training

Add code
Dec 04, 2023
Figure 1 for Customize your NeRF: Adaptive Source Driven 3D Scene Editing via Local-Global Iterative Training
Figure 2 for Customize your NeRF: Adaptive Source Driven 3D Scene Editing via Local-Global Iterative Training
Figure 3 for Customize your NeRF: Adaptive Source Driven 3D Scene Editing via Local-Global Iterative Training
Figure 4 for Customize your NeRF: Adaptive Source Driven 3D Scene Editing via Local-Global Iterative Training
Viaarxiv icon

Enriching Phrases with Coupled Pixel and Object Contexts for Panoptic Narrative Grounding

Add code
Nov 02, 2023
Viaarxiv icon

OSM-Net: One-to-Many One-shot Talking Head Generation with Spontaneous Head Motions

Add code
Sep 28, 2023
Figure 1 for OSM-Net: One-to-Many One-shot Talking Head Generation with Spontaneous Head Motions
Figure 2 for OSM-Net: One-to-Many One-shot Talking Head Generation with Spontaneous Head Motions
Figure 3 for OSM-Net: One-to-Many One-shot Talking Head Generation with Spontaneous Head Motions
Figure 4 for OSM-Net: One-to-Many One-shot Talking Head Generation with Spontaneous Head Motions
Viaarxiv icon

Discovering Sounding Objects by Audio Queries for Audio Visual Segmentation

Add code
Sep 18, 2023
Figure 1 for Discovering Sounding Objects by Audio Queries for Audio Visual Segmentation
Figure 2 for Discovering Sounding Objects by Audio Queries for Audio Visual Segmentation
Figure 3 for Discovering Sounding Objects by Audio Queries for Audio Visual Segmentation
Figure 4 for Discovering Sounding Objects by Audio Queries for Audio Visual Segmentation
Viaarxiv icon

MFR-Net: Multi-faceted Responsive Listening Head Generation via Denoising Diffusion Model

Add code
Aug 31, 2023
Figure 1 for MFR-Net: Multi-faceted Responsive Listening Head Generation via Denoising Diffusion Model
Figure 2 for MFR-Net: Multi-faceted Responsive Listening Head Generation via Denoising Diffusion Model
Figure 3 for MFR-Net: Multi-faceted Responsive Listening Head Generation via Denoising Diffusion Model
Figure 4 for MFR-Net: Multi-faceted Responsive Listening Head Generation via Denoising Diffusion Model
Viaarxiv icon

Modality-Agnostic Audio-Visual Deepfake Detection

Add code
Jul 26, 2023
Figure 1 for Modality-Agnostic Audio-Visual Deepfake Detection
Figure 2 for Modality-Agnostic Audio-Visual Deepfake Detection
Figure 3 for Modality-Agnostic Audio-Visual Deepfake Detection
Figure 4 for Modality-Agnostic Audio-Visual Deepfake Detection
Viaarxiv icon

FONT: Flow-guided One-shot Talking Head Generation with Natural Head Motions

Add code
Mar 31, 2023
Figure 1 for FONT: Flow-guided One-shot Talking Head Generation with Natural Head Motions
Figure 2 for FONT: Flow-guided One-shot Talking Head Generation with Natural Head Motions
Figure 3 for FONT: Flow-guided One-shot Talking Head Generation with Natural Head Motions
Figure 4 for FONT: Flow-guided One-shot Talking Head Generation with Natural Head Motions
Viaarxiv icon

OPT: One-shot Pose-Controllable Talking Head Generation

Add code
Feb 16, 2023
Figure 1 for OPT: One-shot Pose-Controllable Talking Head Generation
Figure 2 for OPT: One-shot Pose-Controllable Talking Head Generation
Figure 3 for OPT: One-shot Pose-Controllable Talking Head Generation
Figure 4 for OPT: One-shot Pose-Controllable Talking Head Generation
Viaarxiv icon

Anchor3DLane: Learning to Regress 3D Anchors for Monocular 3D Lane Detection

Add code
Jan 06, 2023
Figure 1 for Anchor3DLane: Learning to Regress 3D Anchors for Monocular 3D Lane Detection
Figure 2 for Anchor3DLane: Learning to Regress 3D Anchors for Monocular 3D Lane Detection
Figure 3 for Anchor3DLane: Learning to Regress 3D Anchors for Monocular 3D Lane Detection
Figure 4 for Anchor3DLane: Learning to Regress 3D Anchors for Monocular 3D Lane Detection
Viaarxiv icon

RaP: Redundancy-aware Video-language Pre-training for Text-Video Retrieval

Add code
Oct 13, 2022
Figure 1 for RaP: Redundancy-aware Video-language Pre-training for Text-Video Retrieval
Figure 2 for RaP: Redundancy-aware Video-language Pre-training for Text-Video Retrieval
Figure 3 for RaP: Redundancy-aware Video-language Pre-training for Text-Video Retrieval
Figure 4 for RaP: Redundancy-aware Video-language Pre-training for Text-Video Retrieval
Viaarxiv icon