Picture for Jiaya Jia

Jiaya Jia

LogoSticker: Inserting Logos into Diffusion Models for Customized Generation

Add code
Jul 18, 2024
Figure 1 for LogoSticker: Inserting Logos into Diffusion Models for Customized Generation
Figure 2 for LogoSticker: Inserting Logos into Diffusion Models for Customized Generation
Figure 3 for LogoSticker: Inserting Logos into Diffusion Models for Customized Generation
Figure 4 for LogoSticker: Inserting Logos into Diffusion Models for Customized Generation
Viaarxiv icon

Mind the Interference: Retaining Pre-trained Knowledge in Parameter Efficient Continual Learning of Vision-Language Models

Add code
Jul 07, 2024
Figure 1 for Mind the Interference: Retaining Pre-trained Knowledge in Parameter Efficient Continual Learning of Vision-Language Models
Viaarxiv icon

Step-DPO: Step-wise Preference Optimization for Long-chain Reasoning of LLMs

Add code
Jun 26, 2024
Figure 1 for Step-DPO: Step-wise Preference Optimization for Long-chain Reasoning of LLMs
Figure 2 for Step-DPO: Step-wise Preference Optimization for Long-chain Reasoning of LLMs
Figure 3 for Step-DPO: Step-wise Preference Optimization for Long-chain Reasoning of LLMs
Figure 4 for Step-DPO: Step-wise Preference Optimization for Long-chain Reasoning of LLMs
Viaarxiv icon

MR-BEN: A Comprehensive Meta-Reasoning Benchmark for Large Language Models

Add code
Jun 20, 2024
Figure 1 for MR-BEN: A Comprehensive Meta-Reasoning Benchmark for Large Language Models
Figure 2 for MR-BEN: A Comprehensive Meta-Reasoning Benchmark for Large Language Models
Figure 3 for MR-BEN: A Comprehensive Meta-Reasoning Benchmark for Large Language Models
Figure 4 for MR-BEN: A Comprehensive Meta-Reasoning Benchmark for Large Language Models
Viaarxiv icon

QuickLLaMA: Query-aware Inference Acceleration for Large Language Models

Add code
Jun 11, 2024
Figure 1 for QuickLLaMA: Query-aware Inference Acceleration for Large Language Models
Figure 2 for QuickLLaMA: Query-aware Inference Acceleration for Large Language Models
Figure 3 for QuickLLaMA: Query-aware Inference Acceleration for Large Language Models
Figure 4 for QuickLLaMA: Query-aware Inference Acceleration for Large Language Models
Viaarxiv icon

RoboCoder: Robotic Learning from Basic Skills to General Tasks with Large Language Models

Add code
Jun 06, 2024
Figure 1 for RoboCoder: Robotic Learning from Basic Skills to General Tasks with Large Language Models
Figure 2 for RoboCoder: Robotic Learning from Basic Skills to General Tasks with Large Language Models
Figure 3 for RoboCoder: Robotic Learning from Basic Skills to General Tasks with Large Language Models
Figure 4 for RoboCoder: Robotic Learning from Basic Skills to General Tasks with Large Language Models
Viaarxiv icon

Empowering Visual Creativity: A Vision-Language Assistant to Image Editing Recommendations

Add code
May 31, 2024
Figure 1 for Empowering Visual Creativity: A Vision-Language Assistant to Image Editing Recommendations
Figure 2 for Empowering Visual Creativity: A Vision-Language Assistant to Image Editing Recommendations
Figure 3 for Empowering Visual Creativity: A Vision-Language Assistant to Image Editing Recommendations
Figure 4 for Empowering Visual Creativity: A Vision-Language Assistant to Image Editing Recommendations
Viaarxiv icon

Scalable Language Model with Generalized Continual Learning

Add code
Apr 11, 2024
Figure 1 for Scalable Language Model with Generalized Continual Learning
Figure 2 for Scalable Language Model with Generalized Continual Learning
Figure 3 for Scalable Language Model with Generalized Continual Learning
Figure 4 for Scalable Language Model with Generalized Continual Learning
Viaarxiv icon

Unified Language-driven Zero-shot Domain Adaptation

Add code
Apr 10, 2024
Viaarxiv icon

Mini-Gemini: Mining the Potential of Multi-modality Vision Language Models

Add code
Mar 27, 2024
Figure 1 for Mini-Gemini: Mining the Potential of Multi-modality Vision Language Models
Figure 2 for Mini-Gemini: Mining the Potential of Multi-modality Vision Language Models
Figure 3 for Mini-Gemini: Mining the Potential of Multi-modality Vision Language Models
Figure 4 for Mini-Gemini: Mining the Potential of Multi-modality Vision Language Models
Viaarxiv icon