Picture for Xiaoying Zhang

Xiaoying Zhang

User-Creator Feature Dynamics in Recommender Systems with Dual Influence

Add code
Jul 19, 2024
Figure 1 for User-Creator Feature Dynamics in Recommender Systems with Dual Influence
Figure 2 for User-Creator Feature Dynamics in Recommender Systems with Dual Influence
Figure 3 for User-Creator Feature Dynamics in Recommender Systems with Dual Influence
Figure 4 for User-Creator Feature Dynamics in Recommender Systems with Dual Influence
Viaarxiv icon

Toward Optimal LLM Alignments Using Two-Player Games

Add code
Jun 16, 2024
Figure 1 for Toward Optimal LLM Alignments Using Two-Player Games
Figure 2 for Toward Optimal LLM Alignments Using Two-Player Games
Figure 3 for Toward Optimal LLM Alignments Using Two-Player Games
Figure 4 for Toward Optimal LLM Alignments Using Two-Player Games
Viaarxiv icon

Self-Tuning: Instructing LLMs to Effectively Acquire New Knowledge through Self-Teaching

Add code
Jun 11, 2024
Figure 1 for Self-Tuning: Instructing LLMs to Effectively Acquire New Knowledge through Self-Teaching
Figure 2 for Self-Tuning: Instructing LLMs to Effectively Acquire New Knowledge through Self-Teaching
Figure 3 for Self-Tuning: Instructing LLMs to Effectively Acquire New Knowledge through Self-Teaching
Figure 4 for Self-Tuning: Instructing LLMs to Effectively Acquire New Knowledge through Self-Teaching
Viaarxiv icon

GI-Free Pilot-Aided Channel Estimation for Affine Frequency Division Multiplexing Systems

Add code
Apr 01, 2024
Figure 1 for GI-Free Pilot-Aided Channel Estimation for Affine Frequency Division Multiplexing Systems
Figure 2 for GI-Free Pilot-Aided Channel Estimation for Affine Frequency Division Multiplexing Systems
Figure 3 for GI-Free Pilot-Aided Channel Estimation for Affine Frequency Division Multiplexing Systems
Figure 4 for GI-Free Pilot-Aided Channel Estimation for Affine Frequency Division Multiplexing Systems
Viaarxiv icon

Improving Reinforcement Learning from Human Feedback Using Contrastive Rewards

Add code
Mar 14, 2024
Figure 1 for Improving Reinforcement Learning from Human Feedback Using Contrastive Rewards
Figure 2 for Improving Reinforcement Learning from Human Feedback Using Contrastive Rewards
Figure 3 for Improving Reinforcement Learning from Human Feedback Using Contrastive Rewards
Figure 4 for Improving Reinforcement Learning from Human Feedback Using Contrastive Rewards
Viaarxiv icon

Overcoming Reward Overoptimization via Adversarial Policy Optimization with Lightweight Uncertainty Estimation

Add code
Mar 08, 2024
Figure 1 for Overcoming Reward Overoptimization via Adversarial Policy Optimization with Lightweight Uncertainty Estimation
Figure 2 for Overcoming Reward Overoptimization via Adversarial Policy Optimization with Lightweight Uncertainty Estimation
Figure 3 for Overcoming Reward Overoptimization via Adversarial Policy Optimization with Lightweight Uncertainty Estimation
Viaarxiv icon

Self-Alignment for Factuality: Mitigating Hallucinations in LLMs via Self-Evaluation

Add code
Feb 14, 2024
Viaarxiv icon

Human-Instruction-Free LLM Self-Alignment with Limited Samples

Add code
Jan 06, 2024
Viaarxiv icon

Rethinking Machine Ethics -- Can LLMs Perform Moral Reasoning through the Lens of Moral Theories?

Add code
Aug 29, 2023
Figure 1 for Rethinking Machine Ethics -- Can LLMs Perform Moral Reasoning through the Lens of Moral Theories?
Figure 2 for Rethinking Machine Ethics -- Can LLMs Perform Moral Reasoning through the Lens of Moral Theories?
Figure 3 for Rethinking Machine Ethics -- Can LLMs Perform Moral Reasoning through the Lens of Moral Theories?
Figure 4 for Rethinking Machine Ethics -- Can LLMs Perform Moral Reasoning through the Lens of Moral Theories?
Viaarxiv icon

Trustworthy LLMs: a Survey and Guideline for Evaluating Large Language Models' Alignment

Add code
Aug 10, 2023
Figure 1 for Trustworthy LLMs: a Survey and Guideline for Evaluating Large Language Models' Alignment
Figure 2 for Trustworthy LLMs: a Survey and Guideline for Evaluating Large Language Models' Alignment
Figure 3 for Trustworthy LLMs: a Survey and Guideline for Evaluating Large Language Models' Alignment
Figure 4 for Trustworthy LLMs: a Survey and Guideline for Evaluating Large Language Models' Alignment
Viaarxiv icon