Picture for Yongqi Tong

Yongqi Tong

Understanding and Mitigating Overrefusal in LLMs from an Unveiling Perspective of Safety Decision Boundary

Add code
May 23, 2025
Viaarxiv icon

REWARD CONSISTENCY: Improving Multi-Objective Alignment from a Data-Centric Perspective

Add code
Apr 15, 2025
Figure 1 for REWARD CONSISTENCY: Improving Multi-Objective Alignment from a Data-Centric Perspective
Figure 2 for REWARD CONSISTENCY: Improving Multi-Objective Alignment from a Data-Centric Perspective
Figure 3 for REWARD CONSISTENCY: Improving Multi-Objective Alignment from a Data-Centric Perspective
Figure 4 for REWARD CONSISTENCY: Improving Multi-Objective Alignment from a Data-Centric Perspective
Viaarxiv icon

BPO: Towards Balanced Preference Optimization between Knowledge Breadth and Depth in Alignment

Add code
Nov 16, 2024
Figure 1 for BPO: Towards Balanced Preference Optimization between Knowledge Breadth and Depth in Alignment
Figure 2 for BPO: Towards Balanced Preference Optimization between Knowledge Breadth and Depth in Alignment
Figure 3 for BPO: Towards Balanced Preference Optimization between Knowledge Breadth and Depth in Alignment
Figure 4 for BPO: Towards Balanced Preference Optimization between Knowledge Breadth and Depth in Alignment
Viaarxiv icon

Optimizing Language Model's Reasoning Abilities with Weak Supervision

Add code
May 07, 2024
Figure 1 for Optimizing Language Model's Reasoning Abilities with Weak Supervision
Figure 2 for Optimizing Language Model's Reasoning Abilities with Weak Supervision
Figure 3 for Optimizing Language Model's Reasoning Abilities with Weak Supervision
Figure 4 for Optimizing Language Model's Reasoning Abilities with Weak Supervision
Viaarxiv icon

Can LLMs Learn from Previous Mistakes? Investigating LLMs' Errors to Boost for Reasoning

Add code
Mar 29, 2024
Figure 1 for Can LLMs Learn from Previous Mistakes? Investigating LLMs' Errors to Boost for Reasoning
Figure 2 for Can LLMs Learn from Previous Mistakes? Investigating LLMs' Errors to Boost for Reasoning
Figure 3 for Can LLMs Learn from Previous Mistakes? Investigating LLMs' Errors to Boost for Reasoning
Figure 4 for Can LLMs Learn from Previous Mistakes? Investigating LLMs' Errors to Boost for Reasoning
Viaarxiv icon

ToxicChat: Unveiling Hidden Challenges of Toxicity Detection in Real-World User-AI Conversation

Add code
Oct 26, 2023
Viaarxiv icon

Eliminating Reasoning via Inferring with Planning: A New Framework to Guide LLMs' Non-linear Thinking

Add code
Oct 18, 2023
Viaarxiv icon

Pretrained Language Encoders are Natural Tagging Frameworks for Aspect Sentiment Triplet Extraction

Add code
Aug 20, 2022
Figure 1 for Pretrained Language Encoders are Natural Tagging Frameworks for Aspect Sentiment Triplet Extraction
Figure 2 for Pretrained Language Encoders are Natural Tagging Frameworks for Aspect Sentiment Triplet Extraction
Figure 3 for Pretrained Language Encoders are Natural Tagging Frameworks for Aspect Sentiment Triplet Extraction
Figure 4 for Pretrained Language Encoders are Natural Tagging Frameworks for Aspect Sentiment Triplet Extraction
Viaarxiv icon