Picture for Can Qin

Can Qin

STLLaVA-Med: Self-Training Large Language and Vision Assistant for Medical

Add code
Jun 28, 2024
Figure 1 for STLLaVA-Med: Self-Training Large Language and Vision Assistant for Medical
Figure 2 for STLLaVA-Med: Self-Training Large Language and Vision Assistant for Medical
Figure 3 for STLLaVA-Med: Self-Training Large Language and Vision Assistant for Medical
Figure 4 for STLLaVA-Med: Self-Training Large Language and Vision Assistant for Medical
Viaarxiv icon

MuseumMaker: Continual Style Customization without Catastrophic Forgetting

Add code
Apr 29, 2024
Viaarxiv icon

SQ-LLaVA: Self-Questioning for Large Vision-Language Assistant

Add code
Mar 17, 2024
Figure 1 for SQ-LLaVA: Self-Questioning for Large Vision-Language Assistant
Figure 2 for SQ-LLaVA: Self-Questioning for Large Vision-Language Assistant
Figure 3 for SQ-LLaVA: Self-Questioning for Large Vision-Language Assistant
Figure 4 for SQ-LLaVA: Self-Questioning for Large Vision-Language Assistant
Viaarxiv icon

M3SOT: Multi-frame, Multi-field, Multi-space 3D Single Object Tracking

Add code
Dec 11, 2023
Viaarxiv icon

Camouflaged Image Synthesis Is All You Need to Boost Camouflaged Detection

Add code
Aug 13, 2023
Figure 1 for Camouflaged Image Synthesis Is All You Need to Boost Camouflaged Detection
Figure 2 for Camouflaged Image Synthesis Is All You Need to Boost Camouflaged Detection
Figure 3 for Camouflaged Image Synthesis Is All You Need to Boost Camouflaged Detection
Figure 4 for Camouflaged Image Synthesis Is All You Need to Boost Camouflaged Detection
Viaarxiv icon

UniControl: A Unified Diffusion Model for Controllable Visual Generation In the Wild

Add code
May 25, 2023
Figure 1 for UniControl: A Unified Diffusion Model for Controllable Visual Generation In the Wild
Figure 2 for UniControl: A Unified Diffusion Model for Controllable Visual Generation In the Wild
Figure 3 for UniControl: A Unified Diffusion Model for Controllable Visual Generation In the Wild
Figure 4 for UniControl: A Unified Diffusion Model for Controllable Visual Generation In the Wild
Viaarxiv icon

Mask-free OVIS: Open-Vocabulary Instance Segmentation without Manual Mask Annotations

Add code
Mar 29, 2023
Figure 1 for Mask-free OVIS: Open-Vocabulary Instance Segmentation without Manual Mask Annotations
Figure 2 for Mask-free OVIS: Open-Vocabulary Instance Segmentation without Manual Mask Annotations
Figure 3 for Mask-free OVIS: Open-Vocabulary Instance Segmentation without Manual Mask Annotations
Figure 4 for Mask-free OVIS: Open-Vocabulary Instance Segmentation without Manual Mask Annotations
Viaarxiv icon

GlueGen: Plug and Play Multi-modal Encoders for X-to-image Generation

Add code
Mar 17, 2023
Figure 1 for GlueGen: Plug and Play Multi-modal Encoders for X-to-image Generation
Figure 2 for GlueGen: Plug and Play Multi-modal Encoders for X-to-image Generation
Figure 3 for GlueGen: Plug and Play Multi-modal Encoders for X-to-image Generation
Figure 4 for GlueGen: Plug and Play Multi-modal Encoders for X-to-image Generation
Viaarxiv icon

HIVE: Harnessing Human Feedback for Instructional Visual Editing

Add code
Mar 16, 2023
Figure 1 for HIVE: Harnessing Human Feedback for Instructional Visual Editing
Figure 2 for HIVE: Harnessing Human Feedback for Instructional Visual Editing
Figure 3 for HIVE: Harnessing Human Feedback for Instructional Visual Editing
Figure 4 for HIVE: Harnessing Human Feedback for Instructional Visual Editing
Viaarxiv icon

Image as Set of Points

Add code
Mar 02, 2023
Figure 1 for Image as Set of Points
Figure 2 for Image as Set of Points
Figure 3 for Image as Set of Points
Figure 4 for Image as Set of Points
Viaarxiv icon