Picture for Lei Zhu

Lei Zhu

Partially Supervised Unpaired Multi-Modal Learning for Label-Efficient Medical Image Segmentation

Add code
Mar 07, 2025
Viaarxiv icon

MagicInfinite: Generating Infinite Talking Videos with Your Words and Voice

Add code
Mar 07, 2025
Viaarxiv icon

Occlusion-Aware Consistent Model Predictive Control for Robot Navigation in Occluded Obstacle-Dense Environments

Add code
Mar 06, 2025
Figure 1 for Occlusion-Aware Consistent Model Predictive Control for Robot Navigation in Occluded Obstacle-Dense Environments
Figure 2 for Occlusion-Aware Consistent Model Predictive Control for Robot Navigation in Occluded Obstacle-Dense Environments
Figure 3 for Occlusion-Aware Consistent Model Predictive Control for Robot Navigation in Occluded Obstacle-Dense Environments
Figure 4 for Occlusion-Aware Consistent Model Predictive Control for Robot Navigation in Occluded Obstacle-Dense Environments
Viaarxiv icon

Provable Ordering and Continuity in Vision-Language Pretraining for Generalizable Embodied Agents

Add code
Feb 03, 2025
Viaarxiv icon

Learning Semantic Facial Descriptors for Accurate Face Animation

Add code
Jan 29, 2025
Figure 1 for Learning Semantic Facial Descriptors for Accurate Face Animation
Figure 2 for Learning Semantic Facial Descriptors for Accurate Face Animation
Figure 3 for Learning Semantic Facial Descriptors for Accurate Face Animation
Figure 4 for Learning Semantic Facial Descriptors for Accurate Face Animation
Viaarxiv icon

V2C-CBM: Building Concept Bottlenecks with Vision-to-Concept Tokenizer

Add code
Jan 09, 2025
Figure 1 for V2C-CBM: Building Concept Bottlenecks with Vision-to-Concept Tokenizer
Figure 2 for V2C-CBM: Building Concept Bottlenecks with Vision-to-Concept Tokenizer
Figure 3 for V2C-CBM: Building Concept Bottlenecks with Vision-to-Concept Tokenizer
Figure 4 for V2C-CBM: Building Concept Bottlenecks with Vision-to-Concept Tokenizer
Viaarxiv icon

CALMM-Drive: Confidence-Aware Autonomous Driving with Large Multimodal Model

Add code
Dec 05, 2024
Figure 1 for CALMM-Drive: Confidence-Aware Autonomous Driving with Large Multimodal Model
Figure 2 for CALMM-Drive: Confidence-Aware Autonomous Driving with Large Multimodal Model
Figure 3 for CALMM-Drive: Confidence-Aware Autonomous Driving with Large Multimodal Model
Figure 4 for CALMM-Drive: Confidence-Aware Autonomous Driving with Large Multimodal Model
Viaarxiv icon

DRIVE: Dual-Robustness via Information Variability and Entropic Consistency in Source-Free Unsupervised Domain Adaptation

Add code
Nov 24, 2024
Figure 1 for DRIVE: Dual-Robustness via Information Variability and Entropic Consistency in Source-Free Unsupervised Domain Adaptation
Figure 2 for DRIVE: Dual-Robustness via Information Variability and Entropic Consistency in Source-Free Unsupervised Domain Adaptation
Figure 3 for DRIVE: Dual-Robustness via Information Variability and Entropic Consistency in Source-Free Unsupervised Domain Adaptation
Figure 4 for DRIVE: Dual-Robustness via Information Variability and Entropic Consistency in Source-Free Unsupervised Domain Adaptation
Viaarxiv icon

Revisiting the Integration of Convolution and Attention for Vision Backbone

Add code
Nov 21, 2024
Viaarxiv icon

Federated Domain Generalization via Prompt Learning and Aggregation

Add code
Nov 15, 2024
Figure 1 for Federated Domain Generalization via Prompt Learning and Aggregation
Figure 2 for Federated Domain Generalization via Prompt Learning and Aggregation
Figure 3 for Federated Domain Generalization via Prompt Learning and Aggregation
Figure 4 for Federated Domain Generalization via Prompt Learning and Aggregation
Viaarxiv icon