Picture for Hongfa Wang

Hongfa Wang

Video-Language Alignment Pre-training via Spatio-Temporal Graph Transformer

Add code
Jul 16, 2024
Viaarxiv icon

Follow-Your-Pose v2: Multiple-Condition Guided Character Image Animation for Stable Pose Control

Add code
Jun 05, 2024
Figure 1 for Follow-Your-Pose v2: Multiple-Condition Guided Character Image Animation for Stable Pose Control
Figure 2 for Follow-Your-Pose v2: Multiple-Condition Guided Character Image Animation for Stable Pose Control
Figure 3 for Follow-Your-Pose v2: Multiple-Condition Guided Character Image Animation for Stable Pose Control
Figure 4 for Follow-Your-Pose v2: Multiple-Condition Guided Character Image Animation for Stable Pose Control
Viaarxiv icon

Follow-Your-Emoji: Fine-Controllable and Expressive Freestyle Portrait Animation

Add code
Jun 04, 2024
Viaarxiv icon

Follow-Your-Click: Open-domain Regional Image Animation via Short Prompts

Add code
Mar 13, 2024
Figure 1 for Follow-Your-Click: Open-domain Regional Image Animation via Short Prompts
Figure 2 for Follow-Your-Click: Open-domain Regional Image Animation via Short Prompts
Figure 3 for Follow-Your-Click: Open-domain Regional Image Animation via Short Prompts
Figure 4 for Follow-Your-Click: Open-domain Regional Image Animation via Short Prompts
Viaarxiv icon

Inverse-like Antagonistic Scene Text Spotting via Reading-Order Estimation and Dynamic Sampling

Add code
Jan 08, 2024
Viaarxiv icon

Global and Local Semantic Completion Learning for Vision-Language Pre-training

Add code
Jun 12, 2023
Figure 1 for Global and Local Semantic Completion Learning for Vision-Language Pre-training
Figure 2 for Global and Local Semantic Completion Learning for Vision-Language Pre-training
Figure 3 for Global and Local Semantic Completion Learning for Vision-Language Pre-training
Figure 4 for Global and Local Semantic Completion Learning for Vision-Language Pre-training
Viaarxiv icon

Img2Vec: A Teacher of High Token-Diversity Helps Masked AutoEncoders

Add code
Apr 25, 2023
Figure 1 for Img2Vec: A Teacher of High Token-Diversity Helps Masked AutoEncoders
Figure 2 for Img2Vec: A Teacher of High Token-Diversity Helps Masked AutoEncoders
Figure 3 for Img2Vec: A Teacher of High Token-Diversity Helps Masked AutoEncoders
Figure 4 for Img2Vec: A Teacher of High Token-Diversity Helps Masked AutoEncoders
Viaarxiv icon

Seeing What You Miss: Vision-Language Pre-training with Semantic Completion Learning

Add code
Nov 24, 2022
Figure 1 for Seeing What You Miss: Vision-Language Pre-training with Semantic Completion Learning
Figure 2 for Seeing What You Miss: Vision-Language Pre-training with Semantic Completion Learning
Figure 3 for Seeing What You Miss: Vision-Language Pre-training with Semantic Completion Learning
Figure 4 for Seeing What You Miss: Vision-Language Pre-training with Semantic Completion Learning
Viaarxiv icon

MAP: Modality-Agnostic Uncertainty-Aware Vision-Language Pre-training Model

Add code
Oct 11, 2022
Figure 1 for MAP: Modality-Agnostic Uncertainty-Aware Vision-Language Pre-training Model
Figure 2 for MAP: Modality-Agnostic Uncertainty-Aware Vision-Language Pre-training Model
Figure 3 for MAP: Modality-Agnostic Uncertainty-Aware Vision-Language Pre-training Model
Figure 4 for MAP: Modality-Agnostic Uncertainty-Aware Vision-Language Pre-training Model
Viaarxiv icon

Unsupervised Hashing with Semantic Concept Mining

Add code
Sep 23, 2022
Figure 1 for Unsupervised Hashing with Semantic Concept Mining
Figure 2 for Unsupervised Hashing with Semantic Concept Mining
Figure 3 for Unsupervised Hashing with Semantic Concept Mining
Figure 4 for Unsupervised Hashing with Semantic Concept Mining
Viaarxiv icon