Picture for Yu Tian

Yu Tian

Rutgers University

FairDomain: Achieving Fairness in Cross-Domain Medical Image Segmentation and Classification

Add code
Jul 11, 2024
Viaarxiv icon

Strong and Weak Random Walks on Signed Networks

Add code
Jun 12, 2024
Figure 1 for Strong and Weak Random Walks on Signed Networks
Figure 2 for Strong and Weak Random Walks on Signed Networks
Figure 3 for Strong and Weak Random Walks on Signed Networks
Figure 4 for Strong and Weak Random Walks on Signed Networks
Viaarxiv icon

HPE-CogVLM: New Head Pose Grounding Task Exploration on Vision Language Model

Add code
Jun 04, 2024
Viaarxiv icon

AutoBreach: Universal and Adaptive Jailbreaking with Efficient Wordplay-Guided Optimization

Add code
May 30, 2024
Viaarxiv icon

The RoboDrive Challenge: Drive Anytime Anywhere in Any Condition

Add code
May 14, 2024
Figure 1 for The RoboDrive Challenge: Drive Anytime Anywhere in Any Condition
Figure 2 for The RoboDrive Challenge: Drive Anytime Anywhere in Any Condition
Figure 3 for The RoboDrive Challenge: Drive Anytime Anywhere in Any Condition
Figure 4 for The RoboDrive Challenge: Drive Anytime Anywhere in Any Condition
Viaarxiv icon

Enhancing Multi-modal Learning: Meta-learned Cross-modal Knowledge Distillation for Handling Missing Modalities

Add code
May 12, 2024
Figure 1 for Enhancing Multi-modal Learning: Meta-learned Cross-modal Knowledge Distillation for Handling Missing Modalities
Figure 2 for Enhancing Multi-modal Learning: Meta-learned Cross-modal Knowledge Distillation for Handling Missing Modalities
Figure 3 for Enhancing Multi-modal Learning: Meta-learned Cross-modal Knowledge Distillation for Handling Missing Modalities
Figure 4 for Enhancing Multi-modal Learning: Meta-learned Cross-modal Knowledge Distillation for Handling Missing Modalities
Viaarxiv icon

FairCLIP: Harnessing Fairness in Vision-Language Learning

Add code
Apr 05, 2024
Figure 1 for FairCLIP: Harnessing Fairness in Vision-Language Learning
Figure 2 for FairCLIP: Harnessing Fairness in Vision-Language Learning
Figure 3 for FairCLIP: Harnessing Fairness in Vision-Language Learning
Figure 4 for FairCLIP: Harnessing Fairness in Vision-Language Learning
Viaarxiv icon

VisionGPT-3D: A Generalized Multimodal Agent for Enhanced 3D Vision Understanding

Add code
Mar 22, 2024
Figure 1 for VisionGPT-3D: A Generalized Multimodal Agent for Enhanced 3D Vision Understanding
Figure 2 for VisionGPT-3D: A Generalized Multimodal Agent for Enhanced 3D Vision Understanding
Figure 3 for VisionGPT-3D: A Generalized Multimodal Agent for Enhanced 3D Vision Understanding
Figure 4 for VisionGPT-3D: A Generalized Multimodal Agent for Enhanced 3D Vision Understanding
Viaarxiv icon

VisionGPT: Vision-Language Understanding Agent Using Generalized Multimodal Framework

Add code
Mar 14, 2024
Figure 1 for VisionGPT: Vision-Language Understanding Agent Using Generalized Multimodal Framework
Figure 2 for VisionGPT: Vision-Language Understanding Agent Using Generalized Multimodal Framework
Figure 3 for VisionGPT: Vision-Language Understanding Agent Using Generalized Multimodal Framework
Figure 4 for VisionGPT: Vision-Language Understanding Agent Using Generalized Multimodal Framework
Viaarxiv icon

WorldGPT: A Sora-Inspired Video AI Agent as Rich World Models from Text and Image Inputs

Add code
Mar 10, 2024
Figure 1 for WorldGPT: A Sora-Inspired Video AI Agent as Rich World Models from Text and Image Inputs
Figure 2 for WorldGPT: A Sora-Inspired Video AI Agent as Rich World Models from Text and Image Inputs
Figure 3 for WorldGPT: A Sora-Inspired Video AI Agent as Rich World Models from Text and Image Inputs
Figure 4 for WorldGPT: A Sora-Inspired Video AI Agent as Rich World Models from Text and Image Inputs
Viaarxiv icon