Picture for Kaikai An

Kaikai An

UltraEdit: Instruction-based Fine-Grained Image Editing at Scale

Add code
Jul 07, 2024
Figure 1 for UltraEdit: Instruction-based Fine-Grained Image Editing at Scale
Figure 2 for UltraEdit: Instruction-based Fine-Grained Image Editing at Scale
Figure 3 for UltraEdit: Instruction-based Fine-Grained Image Editing at Scale
Figure 4 for UltraEdit: Instruction-based Fine-Grained Image Editing at Scale
Viaarxiv icon

Thread: A Logic-Based Data Organization Paradigm for How-To Question Answering with Retrieval Augmented Generation

Add code
Jun 19, 2024
Figure 1 for Thread: A Logic-Based Data Organization Paradigm for How-To Question Answering with Retrieval Augmented Generation
Figure 2 for Thread: A Logic-Based Data Organization Paradigm for How-To Question Answering with Retrieval Augmented Generation
Figure 3 for Thread: A Logic-Based Data Organization Paradigm for How-To Question Answering with Retrieval Augmented Generation
Figure 4 for Thread: A Logic-Based Data Organization Paradigm for How-To Question Answering with Retrieval Augmented Generation
Viaarxiv icon

LaDiC: Are Diffusion Models Really Inferior to Autoregressive Counterparts for Image-to-Text Generation?

Add code
Apr 16, 2024
Viaarxiv icon

Mitigating Language-Level Performance Disparity in mPLMs via Teacher Language Selection and Cross-lingual Self-Distillation

Add code
Apr 12, 2024
Figure 1 for Mitigating Language-Level Performance Disparity in mPLMs via Teacher Language Selection and Cross-lingual Self-Distillation
Figure 2 for Mitigating Language-Level Performance Disparity in mPLMs via Teacher Language Selection and Cross-lingual Self-Distillation
Figure 3 for Mitigating Language-Level Performance Disparity in mPLMs via Teacher Language Selection and Cross-lingual Self-Distillation
Figure 4 for Mitigating Language-Level Performance Disparity in mPLMs via Teacher Language Selection and Cross-lingual Self-Distillation
Viaarxiv icon

Nissist: An Incident Mitigation Copilot based on Troubleshooting Guides

Add code
Feb 27, 2024
Viaarxiv icon

GAIA: Zero-shot Talking Avatar Generation

Add code
Nov 26, 2023
Figure 1 for GAIA: Zero-shot Talking Avatar Generation
Figure 2 for GAIA: Zero-shot Talking Avatar Generation
Figure 3 for GAIA: Zero-shot Talking Avatar Generation
Figure 4 for GAIA: Zero-shot Talking Avatar Generation
Viaarxiv icon

ML-Bench: Large Language Models Leverage Open-source Libraries for Machine Learning Tasks

Add code
Nov 16, 2023
Figure 1 for ML-Bench: Large Language Models Leverage Open-source Libraries for Machine Learning Tasks
Figure 2 for ML-Bench: Large Language Models Leverage Open-source Libraries for Machine Learning Tasks
Figure 3 for ML-Bench: Large Language Models Leverage Open-source Libraries for Machine Learning Tasks
Figure 4 for ML-Bench: Large Language Models Leverage Open-source Libraries for Machine Learning Tasks
Viaarxiv icon

Distantly-Supervised Named Entity Recognition with Uncertainty-aware Teacher Learning and Student-student Collaborative Learning

Add code
Nov 14, 2023
Figure 1 for Distantly-Supervised Named Entity Recognition with Uncertainty-aware Teacher Learning and Student-student Collaborative Learning
Figure 2 for Distantly-Supervised Named Entity Recognition with Uncertainty-aware Teacher Learning and Student-student Collaborative Learning
Figure 3 for Distantly-Supervised Named Entity Recognition with Uncertainty-aware Teacher Learning and Student-student Collaborative Learning
Figure 4 for Distantly-Supervised Named Entity Recognition with Uncertainty-aware Teacher Learning and Student-student Collaborative Learning
Viaarxiv icon

Coarse-to-Fine Dual Encoders are Better Frame Identification Learners

Add code
Oct 20, 2023
Figure 1 for Coarse-to-Fine Dual Encoders are Better Frame Identification Learners
Figure 2 for Coarse-to-Fine Dual Encoders are Better Frame Identification Learners
Figure 3 for Coarse-to-Fine Dual Encoders are Better Frame Identification Learners
Figure 4 for Coarse-to-Fine Dual Encoders are Better Frame Identification Learners
Viaarxiv icon

MMICL: Empowering Vision-language Model with Multi-Modal In-Context Learning

Add code
Oct 02, 2023
Figure 1 for MMICL: Empowering Vision-language Model with Multi-Modal In-Context Learning
Figure 2 for MMICL: Empowering Vision-language Model with Multi-Modal In-Context Learning
Figure 3 for MMICL: Empowering Vision-language Model with Multi-Modal In-Context Learning
Figure 4 for MMICL: Empowering Vision-language Model with Multi-Modal In-Context Learning
Viaarxiv icon