Picture for Hai Huang

Hai Huang

ACE: A Generative Cross-Modal Retrieval Framework with Coarse-To-Fine Semantic Modeling

Add code
Jun 25, 2024
Figure 1 for ACE: A Generative Cross-Modal Retrieval Framework with Coarse-To-Fine Semantic Modeling
Figure 2 for ACE: A Generative Cross-Modal Retrieval Framework with Coarse-To-Fine Semantic Modeling
Figure 3 for ACE: A Generative Cross-Modal Retrieval Framework with Coarse-To-Fine Semantic Modeling
Figure 4 for ACE: A Generative Cross-Modal Retrieval Framework with Coarse-To-Fine Semantic Modeling
Viaarxiv icon

ADBA:Approximation Decision Boundary Approach for Black-Box Adversarial Attacks

Add code
Jun 07, 2024
Viaarxiv icon

Long and Short Guidance in Score identity Distillation for One-Step Text-to-Image Generation

Add code
Jun 03, 2024
Figure 1 for Long and Short Guidance in Score identity Distillation for One-Step Text-to-Image Generation
Figure 2 for Long and Short Guidance in Score identity Distillation for One-Step Text-to-Image Generation
Figure 3 for Long and Short Guidance in Score identity Distillation for One-Step Text-to-Image Generation
Figure 4 for Long and Short Guidance in Score identity Distillation for One-Step Text-to-Image Generation
Viaarxiv icon

ControlSpeech: Towards Simultaneous Zero-shot Speaker Cloning and Zero-shot Language Style Control With Decoupled Codec

Add code
Jun 03, 2024
Viaarxiv icon

Less is More: on the Over-Globalizing Problem in Graph Transformers

Add code
May 02, 2024
Viaarxiv icon

GCC: Generative Calibration Clustering

Add code
Apr 14, 2024
Viaarxiv icon

Score identity Distillation: Exponentially Fast Distillation of Pretrained Diffusion Models for One-Step Generation

Add code
Apr 05, 2024
Figure 1 for Score identity Distillation: Exponentially Fast Distillation of Pretrained Diffusion Models for One-Step Generation
Figure 2 for Score identity Distillation: Exponentially Fast Distillation of Pretrained Diffusion Models for One-Step Generation
Figure 3 for Score identity Distillation: Exponentially Fast Distillation of Pretrained Diffusion Models for One-Step Generation
Figure 4 for Score identity Distillation: Exponentially Fast Distillation of Pretrained Diffusion Models for One-Step Generation
Viaarxiv icon

Unlocking the Potential of Multimodal Unified Discrete Representation through Training-Free Codebook Optimization and Hierarchical Alignment

Add code
Mar 08, 2024
Figure 1 for Unlocking the Potential of Multimodal Unified Discrete Representation through Training-Free Codebook Optimization and Hierarchical Alignment
Figure 2 for Unlocking the Potential of Multimodal Unified Discrete Representation through Training-Free Codebook Optimization and Hierarchical Alignment
Figure 3 for Unlocking the Potential of Multimodal Unified Discrete Representation through Training-Free Codebook Optimization and Hierarchical Alignment
Figure 4 for Unlocking the Potential of Multimodal Unified Discrete Representation through Training-Free Codebook Optimization and Hierarchical Alignment
Viaarxiv icon

Relative Preference Optimization: Enhancing LLM Alignment through Contrasting Responses across Identical and Diverse Prompts

Add code
Feb 12, 2024
Viaarxiv icon

Rethinking Evaluation Metrics of Open-Vocabulary Segmentaion

Add code
Nov 06, 2023
Viaarxiv icon