Picture for Hongning Wang

Hongning Wang

User Welfare Optimization in Recommender Systems with Competing Content Creators

Add code
Apr 28, 2024
Figure 1 for User Welfare Optimization in Recommender Systems with Competing Content Creators
Figure 2 for User Welfare Optimization in Recommender Systems with Competing Content Creators
Figure 3 for User Welfare Optimization in Recommender Systems with Competing Content Creators
Figure 4 for User Welfare Optimization in Recommender Systems with Competing Content Creators
Viaarxiv icon

ChatGLM-RLHF: Practices of Aligning Large Language Models with Human Feedback

Add code
Apr 03, 2024
Figure 1 for ChatGLM-RLHF: Practices of Aligning Large Language Models with Human Feedback
Figure 2 for ChatGLM-RLHF: Practices of Aligning Large Language Models with Human Feedback
Figure 3 for ChatGLM-RLHF: Practices of Aligning Large Language Models with Human Feedback
Figure 4 for ChatGLM-RLHF: Practices of Aligning Large Language Models with Human Feedback
Viaarxiv icon

Overcoming Reward Overoptimization via Adversarial Policy Optimization with Lightweight Uncertainty Estimation

Add code
Mar 08, 2024
Figure 1 for Overcoming Reward Overoptimization via Adversarial Policy Optimization with Lightweight Uncertainty Estimation
Figure 2 for Overcoming Reward Overoptimization via Adversarial Policy Optimization with Lightweight Uncertainty Estimation
Figure 3 for Overcoming Reward Overoptimization via Adversarial Policy Optimization with Lightweight Uncertainty Estimation
Viaarxiv icon

Federated Linear Contextual Bandits with Heterogeneous Clients

Add code
Feb 29, 2024
Viaarxiv icon

ShieldLM: Empowering LLMs as Aligned, Customizable and Explainable Safety Detectors

Add code
Feb 26, 2024
Figure 1 for ShieldLM: Empowering LLMs as Aligned, Customizable and Explainable Safety Detectors
Figure 2 for ShieldLM: Empowering LLMs as Aligned, Customizable and Explainable Safety Detectors
Figure 3 for ShieldLM: Empowering LLMs as Aligned, Customizable and Explainable Safety Detectors
Figure 4 for ShieldLM: Empowering LLMs as Aligned, Customizable and Explainable Safety Detectors
Viaarxiv icon

Stealthy Adversarial Attacks on Stochastic Multi-Armed Bandits

Add code
Feb 21, 2024
Viaarxiv icon

Incentivized Truthful Communication for Federated Bandits

Add code
Feb 07, 2024
Viaarxiv icon

Towards Efficient and Exact Optimization of Language Model Alignment

Add code
Feb 02, 2024
Figure 1 for Towards Efficient and Exact Optimization of Language Model Alignment
Figure 2 for Towards Efficient and Exact Optimization of Language Model Alignment
Figure 3 for Towards Efficient and Exact Optimization of Language Model Alignment
Figure 4 for Towards Efficient and Exact Optimization of Language Model Alignment
Viaarxiv icon

AMOR: A Recipe for Building Adaptable Modular Knowledge Agents Through Process Feedback

Add code
Feb 02, 2024
Figure 1 for AMOR: A Recipe for Building Adaptable Modular Knowledge Agents Through Process Feedback
Figure 2 for AMOR: A Recipe for Building Adaptable Modular Knowledge Agents Through Process Feedback
Figure 3 for AMOR: A Recipe for Building Adaptable Modular Knowledge Agents Through Process Feedback
Figure 4 for AMOR: A Recipe for Building Adaptable Modular Knowledge Agents Through Process Feedback
Viaarxiv icon

The Impact of Snippet Reliability on Misinformation in Online Health Search

Add code
Jan 28, 2024
Viaarxiv icon