Picture for Ganqu Cui

Ganqu Cui

Zero-Shot Generalization during Instruction Tuning: Insights from Similarity and Granularity

Add code
Jun 17, 2024
Viaarxiv icon

UltraMedical: Building Specialized Generalists in Biomedicine

Add code
Jun 06, 2024
Viaarxiv icon

RLAIF-V: Aligning MLLMs through Open-Source AI Feedback for Super GPT-4V Trustworthiness

Add code
May 27, 2024
Viaarxiv icon

MiniCPM: Unveiling the Potential of Small Language Models with Scalable Training Strategies

Add code
Apr 09, 2024
Figure 1 for MiniCPM: Unveiling the Potential of Small Language Models with Scalable Training Strategies
Figure 2 for MiniCPM: Unveiling the Potential of Small Language Models with Scalable Training Strategies
Figure 3 for MiniCPM: Unveiling the Potential of Small Language Models with Scalable Training Strategies
Figure 4 for MiniCPM: Unveiling the Potential of Small Language Models with Scalable Training Strategies
Viaarxiv icon

Advancing LLM Reasoning Generalists with Preference Trees

Add code
Apr 02, 2024
Viaarxiv icon

Mastering Text, Code and Math Simultaneously via Fusing Highly Specialized Language Models

Add code
Mar 18, 2024
Figure 1 for Mastering Text, Code and Math Simultaneously via Fusing Highly Specialized Language Models
Figure 2 for Mastering Text, Code and Math Simultaneously via Fusing Highly Specialized Language Models
Figure 3 for Mastering Text, Code and Math Simultaneously via Fusing Highly Specialized Language Models
Figure 4 for Mastering Text, Code and Math Simultaneously via Fusing Highly Specialized Language Models
Viaarxiv icon

Controllable Preference Optimization: Toward Controllable Multi-Objective Alignment

Add code
Feb 29, 2024
Figure 1 for Controllable Preference Optimization: Toward Controllable Multi-Objective Alignment
Figure 2 for Controllable Preference Optimization: Toward Controllable Multi-Objective Alignment
Figure 3 for Controllable Preference Optimization: Toward Controllable Multi-Objective Alignment
Figure 4 for Controllable Preference Optimization: Toward Controllable Multi-Objective Alignment
Viaarxiv icon

RLHF-V: Towards Trustworthy MLLMs via Behavior Alignment from Fine-grained Correctional Human Feedback

Add code
Dec 01, 2023
Figure 1 for RLHF-V: Towards Trustworthy MLLMs via Behavior Alignment from Fine-grained Correctional Human Feedback
Figure 2 for RLHF-V: Towards Trustworthy MLLMs via Behavior Alignment from Fine-grained Correctional Human Feedback
Figure 3 for RLHF-V: Towards Trustworthy MLLMs via Behavior Alignment from Fine-grained Correctional Human Feedback
Figure 4 for RLHF-V: Towards Trustworthy MLLMs via Behavior Alignment from Fine-grained Correctional Human Feedback
Viaarxiv icon

INTERVENOR: Prompt the Coding Ability of Large Language Models with the Interactive Chain of Repairing

Add code
Nov 16, 2023
Figure 1 for INTERVENOR: Prompt the Coding Ability of Large Language Models with the Interactive Chain of Repairing
Figure 2 for INTERVENOR: Prompt the Coding Ability of Large Language Models with the Interactive Chain of Repairing
Figure 3 for INTERVENOR: Prompt the Coding Ability of Large Language Models with the Interactive Chain of Repairing
Figure 4 for INTERVENOR: Prompt the Coding Ability of Large Language Models with the Interactive Chain of Repairing
Viaarxiv icon

UltraFeedback: Boosting Language Models with High-quality Feedback

Add code
Oct 02, 2023
Figure 1 for UltraFeedback: Boosting Language Models with High-quality Feedback
Figure 2 for UltraFeedback: Boosting Language Models with High-quality Feedback
Figure 3 for UltraFeedback: Boosting Language Models with High-quality Feedback
Figure 4 for UltraFeedback: Boosting Language Models with High-quality Feedback
Viaarxiv icon