Picture for Tianyu Yu

Tianyu Yu

RLAIF-V: Aligning MLLMs through Open-Source AI Feedback for Super GPT-4V Trustworthiness

Add code
May 27, 2024
Viaarxiv icon

UltraWiki: Ultra-fine-grained Entity Set Expansion with Negative Seed Entities

Add code
Mar 07, 2024
Figure 1 for UltraWiki: Ultra-fine-grained Entity Set Expansion with Negative Seed Entities
Figure 2 for UltraWiki: Ultra-fine-grained Entity Set Expansion with Negative Seed Entities
Figure 3 for UltraWiki: Ultra-fine-grained Entity Set Expansion with Negative Seed Entities
Figure 4 for UltraWiki: Ultra-fine-grained Entity Set Expansion with Negative Seed Entities
Viaarxiv icon

RLHF-V: Towards Trustworthy MLLMs via Behavior Alignment from Fine-grained Correctional Human Feedback

Add code
Dec 01, 2023
Figure 1 for RLHF-V: Towards Trustworthy MLLMs via Behavior Alignment from Fine-grained Correctional Human Feedback
Figure 2 for RLHF-V: Towards Trustworthy MLLMs via Behavior Alignment from Fine-grained Correctional Human Feedback
Figure 3 for RLHF-V: Towards Trustworthy MLLMs via Behavior Alignment from Fine-grained Correctional Human Feedback
Figure 4 for RLHF-V: Towards Trustworthy MLLMs via Behavior Alignment from Fine-grained Correctional Human Feedback
Viaarxiv icon

Reformulating Vision-Language Foundation Models and Datasets Towards Universal Multimodal Assistants

Add code
Oct 01, 2023
Figure 1 for Reformulating Vision-Language Foundation Models and Datasets Towards Universal Multimodal Assistants
Figure 2 for Reformulating Vision-Language Foundation Models and Datasets Towards Universal Multimodal Assistants
Figure 3 for Reformulating Vision-Language Foundation Models and Datasets Towards Universal Multimodal Assistants
Figure 4 for Reformulating Vision-Language Foundation Models and Datasets Towards Universal Multimodal Assistants
Viaarxiv icon

Large Multilingual Models Pivot Zero-Shot Multimodal Learning across Languages

Add code
Aug 23, 2023
Figure 1 for Large Multilingual Models Pivot Zero-Shot Multimodal Learning across Languages
Figure 2 for Large Multilingual Models Pivot Zero-Shot Multimodal Learning across Languages
Figure 3 for Large Multilingual Models Pivot Zero-Shot Multimodal Learning across Languages
Figure 4 for Large Multilingual Models Pivot Zero-Shot Multimodal Learning across Languages
Viaarxiv icon

SeqGPT: An Out-of-the-box Large Language Model for Open Domain Sequence Understanding

Add code
Aug 21, 2023
Figure 1 for SeqGPT: An Out-of-the-box Large Language Model for Open Domain Sequence Understanding
Figure 2 for SeqGPT: An Out-of-the-box Large Language Model for Open Domain Sequence Understanding
Figure 3 for SeqGPT: An Out-of-the-box Large Language Model for Open Domain Sequence Understanding
Figure 4 for SeqGPT: An Out-of-the-box Large Language Model for Open Domain Sequence Understanding
Viaarxiv icon

MESED: A Multi-modal Entity Set Expansion Dataset with Fine-grained Semantic Classes and Hard Negative Entities

Add code
Jul 27, 2023
Figure 1 for MESED: A Multi-modal Entity Set Expansion Dataset with Fine-grained Semantic Classes and Hard Negative Entities
Figure 2 for MESED: A Multi-modal Entity Set Expansion Dataset with Fine-grained Semantic Classes and Hard Negative Entities
Figure 3 for MESED: A Multi-modal Entity Set Expansion Dataset with Fine-grained Semantic Classes and Hard Negative Entities
Figure 4 for MESED: A Multi-modal Entity Set Expansion Dataset with Fine-grained Semantic Classes and Hard Negative Entities
Viaarxiv icon

Knowledge-augmented Few-shot Visual Relation Detection

Add code
Mar 09, 2023
Figure 1 for Knowledge-augmented Few-shot Visual Relation Detection
Figure 2 for Knowledge-augmented Few-shot Visual Relation Detection
Figure 3 for Knowledge-augmented Few-shot Visual Relation Detection
Figure 4 for Knowledge-augmented Few-shot Visual Relation Detection
Viaarxiv icon

Visually Grounded Commonsense Knowledge Acquisition

Add code
Nov 22, 2022
Figure 1 for Visually Grounded Commonsense Knowledge Acquisition
Figure 2 for Visually Grounded Commonsense Knowledge Acquisition
Figure 3 for Visually Grounded Commonsense Knowledge Acquisition
Figure 4 for Visually Grounded Commonsense Knowledge Acquisition
Viaarxiv icon

Embracing Ambiguity: Improving Similarity-oriented Tasks with Contextual Synonym Knowledge

Add code
Nov 20, 2022
Figure 1 for Embracing Ambiguity: Improving Similarity-oriented Tasks with Contextual Synonym Knowledge
Figure 2 for Embracing Ambiguity: Improving Similarity-oriented Tasks with Contextual Synonym Knowledge
Figure 3 for Embracing Ambiguity: Improving Similarity-oriented Tasks with Contextual Synonym Knowledge
Figure 4 for Embracing Ambiguity: Improving Similarity-oriented Tasks with Contextual Synonym Knowledge
Viaarxiv icon