Picture for Xiang Kong

Xiang Kong

Checklists Are Better Than Reward Models For Aligning Language Models

Add code
Jul 24, 2025
Viaarxiv icon

Mutual Reinforcement of LLM Dialogue Synthesis and Summarization Capabilities for Few-Shot Dialogue Summarization

Add code
Feb 24, 2025
Viaarxiv icon

TIS-DPO: Token-level Importance Sampling for Direct Preference Optimization With Estimated Weights

Add code
Oct 06, 2024
Figure 1 for TIS-DPO: Token-level Importance Sampling for Direct Preference Optimization With Estimated Weights
Figure 2 for TIS-DPO: Token-level Importance Sampling for Direct Preference Optimization With Estimated Weights
Figure 3 for TIS-DPO: Token-level Importance Sampling for Direct Preference Optimization With Estimated Weights
Figure 4 for TIS-DPO: Token-level Importance Sampling for Direct Preference Optimization With Estimated Weights
Viaarxiv icon

Step-by-Step Reasoning for Math Problems via Twisted Sequential Monte Carlo

Add code
Oct 02, 2024
Figure 1 for Step-by-Step Reasoning for Math Problems via Twisted Sequential Monte Carlo
Figure 2 for Step-by-Step Reasoning for Math Problems via Twisted Sequential Monte Carlo
Figure 3 for Step-by-Step Reasoning for Math Problems via Twisted Sequential Monte Carlo
Figure 4 for Step-by-Step Reasoning for Math Problems via Twisted Sequential Monte Carlo
Viaarxiv icon

Apple Intelligence Foundation Language Models

Add code
Jul 29, 2024
Figure 1 for Apple Intelligence Foundation Language Models
Figure 2 for Apple Intelligence Foundation Language Models
Figure 3 for Apple Intelligence Foundation Language Models
Figure 4 for Apple Intelligence Foundation Language Models
Viaarxiv icon

Large Language Model-guided Document Selection

Add code
Jun 07, 2024
Figure 1 for Large Language Model-guided Document Selection
Figure 2 for Large Language Model-guided Document Selection
Figure 3 for Large Language Model-guided Document Selection
Figure 4 for Large Language Model-guided Document Selection
Viaarxiv icon

Revisiting MoE and Dense Speed-Accuracy Comparisons for LLM Training

Add code
May 23, 2024
Figure 1 for Revisiting MoE and Dense Speed-Accuracy Comparisons for LLM Training
Figure 2 for Revisiting MoE and Dense Speed-Accuracy Comparisons for LLM Training
Figure 3 for Revisiting MoE and Dense Speed-Accuracy Comparisons for LLM Training
Figure 4 for Revisiting MoE and Dense Speed-Accuracy Comparisons for LLM Training
Viaarxiv icon

MM1: Methods, Analysis & Insights from Multimodal LLM Pre-training

Add code
Mar 22, 2024
Figure 1 for MM1: Methods, Analysis & Insights from Multimodal LLM Pre-training
Figure 2 for MM1: Methods, Analysis & Insights from Multimodal LLM Pre-training
Figure 3 for MM1: Methods, Analysis & Insights from Multimodal LLM Pre-training
Figure 4 for MM1: Methods, Analysis & Insights from Multimodal LLM Pre-training
Viaarxiv icon

Direct Large Language Model Alignment Through Self-Rewarding Contrastive Prompt Distillation

Add code
Feb 19, 2024
Figure 1 for Direct Large Language Model Alignment Through Self-Rewarding Contrastive Prompt Distillation
Figure 2 for Direct Large Language Model Alignment Through Self-Rewarding Contrastive Prompt Distillation
Figure 3 for Direct Large Language Model Alignment Through Self-Rewarding Contrastive Prompt Distillation
Figure 4 for Direct Large Language Model Alignment Through Self-Rewarding Contrastive Prompt Distillation
Viaarxiv icon

Mega: Moving Average Equipped Gated Attention

Add code
Sep 26, 2022
Figure 1 for Mega: Moving Average Equipped Gated Attention
Figure 2 for Mega: Moving Average Equipped Gated Attention
Figure 3 for Mega: Moving Average Equipped Gated Attention
Figure 4 for Mega: Moving Average Equipped Gated Attention
Viaarxiv icon