Picture for Xiaodan Liang

Xiaodan Liang

FancyVideo: Towards Dynamic and Consistent Video Generation via Cross-frame Textual Guidance

Add code
Aug 15, 2024
Figure 1 for FancyVideo: Towards Dynamic and Consistent Video Generation via Cross-frame Textual Guidance
Figure 2 for FancyVideo: Towards Dynamic and Consistent Video Generation via Cross-frame Textual Guidance
Figure 3 for FancyVideo: Towards Dynamic and Consistent Video Generation via Cross-frame Textual Guidance
Figure 4 for FancyVideo: Towards Dynamic and Consistent Video Generation via Cross-frame Textual Guidance
Viaarxiv icon

APTNESS: Incorporating Appraisal Theory and Emotion Support Strategies for Empathetic Response Generation

Add code
Jul 23, 2024
Figure 1 for APTNESS: Incorporating Appraisal Theory and Emotion Support Strategies for Empathetic Response Generation
Figure 2 for APTNESS: Incorporating Appraisal Theory and Emotion Support Strategies for Empathetic Response Generation
Figure 3 for APTNESS: Incorporating Appraisal Theory and Emotion Support Strategies for Empathetic Response Generation
Figure 4 for APTNESS: Incorporating Appraisal Theory and Emotion Support Strategies for Empathetic Response Generation
Viaarxiv icon

DreamVTON: Customizing 3D Virtual Try-on with Personalized Diffusion Models

Add code
Jul 23, 2024
Figure 1 for DreamVTON: Customizing 3D Virtual Try-on with Personalized Diffusion Models
Figure 2 for DreamVTON: Customizing 3D Virtual Try-on with Personalized Diffusion Models
Figure 3 for DreamVTON: Customizing 3D Virtual Try-on with Personalized Diffusion Models
Figure 4 for DreamVTON: Customizing 3D Virtual Try-on with Personalized Diffusion Models
Viaarxiv icon

CatVTON: Concatenation Is All You Need for Virtual Try-On with Diffusion Models

Add code
Jul 21, 2024
Figure 1 for CatVTON: Concatenation Is All You Need for Virtual Try-On with Diffusion Models
Figure 2 for CatVTON: Concatenation Is All You Need for Virtual Try-On with Diffusion Models
Figure 3 for CatVTON: Concatenation Is All You Need for Virtual Try-On with Diffusion Models
Figure 4 for CatVTON: Concatenation Is All You Need for Virtual Try-On with Diffusion Models
Viaarxiv icon

Contrastive Learning with Counterfactual Explanations for Radiology Report Generation

Add code
Jul 19, 2024
Figure 1 for Contrastive Learning with Counterfactual Explanations for Radiology Report Generation
Figure 2 for Contrastive Learning with Counterfactual Explanations for Radiology Report Generation
Figure 3 for Contrastive Learning with Counterfactual Explanations for Radiology Report Generation
Figure 4 for Contrastive Learning with Counterfactual Explanations for Radiology Report Generation
Viaarxiv icon

Benchmarking LLMs for Optimization Modeling and Enhancing Reasoning via Reverse Socratic Synthesis

Add code
Jul 13, 2024
Figure 1 for Benchmarking LLMs for Optimization Modeling and Enhancing Reasoning via Reverse Socratic Synthesis
Figure 2 for Benchmarking LLMs for Optimization Modeling and Enhancing Reasoning via Reverse Socratic Synthesis
Figure 3 for Benchmarking LLMs for Optimization Modeling and Enhancing Reasoning via Reverse Socratic Synthesis
Figure 4 for Benchmarking LLMs for Optimization Modeling and Enhancing Reasoning via Reverse Socratic Synthesis
Viaarxiv icon

HiRes-LLaVA: Restoring Fragmentation Input in High-Resolution Large Vision-Language Models

Add code
Jul 11, 2024
Figure 1 for HiRes-LLaVA: Restoring Fragmentation Input in High-Resolution Large Vision-Language Models
Figure 2 for HiRes-LLaVA: Restoring Fragmentation Input in High-Resolution Large Vision-Language Models
Figure 3 for HiRes-LLaVA: Restoring Fragmentation Input in High-Resolution Large Vision-Language Models
Figure 4 for HiRes-LLaVA: Restoring Fragmentation Input in High-Resolution Large Vision-Language Models
Viaarxiv icon

OV-DINO: Unified Open-Vocabulary Detection with Language-Aware Selective Fusion

Add code
Jul 10, 2024
Figure 1 for OV-DINO: Unified Open-Vocabulary Detection with Language-Aware Selective Fusion
Figure 2 for OV-DINO: Unified Open-Vocabulary Detection with Language-Aware Selective Fusion
Figure 3 for OV-DINO: Unified Open-Vocabulary Detection with Language-Aware Selective Fusion
Figure 4 for OV-DINO: Unified Open-Vocabulary Detection with Language-Aware Selective Fusion
Viaarxiv icon

HumanRefiner: Benchmarking Abnormal Human Generation and Refining with Coarse-to-fine Pose-Reversible Guidance

Add code
Jul 09, 2024
Figure 1 for HumanRefiner: Benchmarking Abnormal Human Generation and Refining with Coarse-to-fine Pose-Reversible Guidance
Figure 2 for HumanRefiner: Benchmarking Abnormal Human Generation and Refining with Coarse-to-fine Pose-Reversible Guidance
Figure 3 for HumanRefiner: Benchmarking Abnormal Human Generation and Refining with Coarse-to-fine Pose-Reversible Guidance
Figure 4 for HumanRefiner: Benchmarking Abnormal Human Generation and Refining with Coarse-to-fine Pose-Reversible Guidance
Viaarxiv icon

Affordances-Oriented Planning using Foundation Models for Continuous Vision-Language Navigation

Add code
Jul 08, 2024
Figure 1 for Affordances-Oriented Planning using Foundation Models for Continuous Vision-Language Navigation
Figure 2 for Affordances-Oriented Planning using Foundation Models for Continuous Vision-Language Navigation
Figure 3 for Affordances-Oriented Planning using Foundation Models for Continuous Vision-Language Navigation
Figure 4 for Affordances-Oriented Planning using Foundation Models for Continuous Vision-Language Navigation
Viaarxiv icon