Picture for Zhenghai Xue

Zhenghai Xue

Modeling User Retention through Generative Flow Networks

Add code
Jun 10, 2024
Viaarxiv icon

S$^2$AC: Energy-Based Reinforcement Learning with Stein Soft Actor Critic

Add code
May 02, 2024
Figure 1 for S$^2$AC: Energy-Based Reinforcement Learning with Stein Soft Actor Critic
Figure 2 for S$^2$AC: Energy-Based Reinforcement Learning with Stein Soft Actor Critic
Figure 3 for S$^2$AC: Energy-Based Reinforcement Learning with Stein Soft Actor Critic
Figure 4 for S$^2$AC: Energy-Based Reinforcement Learning with Stein Soft Actor Critic
Viaarxiv icon

AgentStudio: A Toolkit for Building General Virtual Agents

Add code
Mar 26, 2024
Figure 1 for AgentStudio: A Toolkit for Building General Virtual Agents
Figure 2 for AgentStudio: A Toolkit for Building General Virtual Agents
Figure 3 for AgentStudio: A Toolkit for Building General Virtual Agents
Figure 4 for AgentStudio: A Toolkit for Building General Virtual Agents
Viaarxiv icon

AdaRec: Adaptive Sequential Recommendation for Reinforcing Long-term User Engagement

Add code
Oct 06, 2023
Figure 1 for AdaRec: Adaptive Sequential Recommendation for Reinforcing Long-term User Engagement
Figure 2 for AdaRec: Adaptive Sequential Recommendation for Reinforcing Long-term User Engagement
Figure 3 for AdaRec: Adaptive Sequential Recommendation for Reinforcing Long-term User Engagement
Figure 4 for AdaRec: Adaptive Sequential Recommendation for Reinforcing Long-term User Engagement
Viaarxiv icon

A Large Language Model Enhanced Conversational Recommender System

Add code
Aug 11, 2023
Figure 1 for A Large Language Model Enhanced Conversational Recommender System
Figure 2 for A Large Language Model Enhanced Conversational Recommender System
Figure 3 for A Large Language Model Enhanced Conversational Recommender System
Figure 4 for A Large Language Model Enhanced Conversational Recommender System
Viaarxiv icon

State Regularized Policy Optimization on Data with Dynamics Shift

Add code
Jun 06, 2023
Figure 1 for State Regularized Policy Optimization on Data with Dynamics Shift
Figure 2 for State Regularized Policy Optimization on Data with Dynamics Shift
Figure 3 for State Regularized Policy Optimization on Data with Dynamics Shift
Figure 4 for State Regularized Policy Optimization on Data with Dynamics Shift
Viaarxiv icon

Guarded Policy Optimization with Imperfect Online Demonstrations

Add code
Mar 03, 2023
Figure 1 for Guarded Policy Optimization with Imperfect Online Demonstrations
Figure 2 for Guarded Policy Optimization with Imperfect Online Demonstrations
Figure 3 for Guarded Policy Optimization with Imperfect Online Demonstrations
Figure 4 for Guarded Policy Optimization with Imperfect Online Demonstrations
Viaarxiv icon

Two-Stage Constrained Actor-Critic for Short Video Recommendation

Add code
Feb 06, 2023
Figure 1 for Two-Stage Constrained Actor-Critic for Short Video Recommendation
Figure 2 for Two-Stage Constrained Actor-Critic for Short Video Recommendation
Figure 3 for Two-Stage Constrained Actor-Critic for Short Video Recommendation
Figure 4 for Two-Stage Constrained Actor-Critic for Short Video Recommendation
Viaarxiv icon

PrefRec: Preference-based Recommender Systems for Reinforcing Long-term User Engagement

Add code
Dec 06, 2022
Figure 1 for PrefRec: Preference-based Recommender Systems for Reinforcing Long-term User Engagement
Figure 2 for PrefRec: Preference-based Recommender Systems for Reinforcing Long-term User Engagement
Figure 3 for PrefRec: Preference-based Recommender Systems for Reinforcing Long-term User Engagement
Figure 4 for PrefRec: Preference-based Recommender Systems for Reinforcing Long-term User Engagement
Viaarxiv icon

MetaDrive: Composing Diverse Driving Scenarios for Generalizable Reinforcement Learning

Add code
Sep 26, 2021
Figure 1 for MetaDrive: Composing Diverse Driving Scenarios for Generalizable Reinforcement Learning
Figure 2 for MetaDrive: Composing Diverse Driving Scenarios for Generalizable Reinforcement Learning
Figure 3 for MetaDrive: Composing Diverse Driving Scenarios for Generalizable Reinforcement Learning
Figure 4 for MetaDrive: Composing Diverse Driving Scenarios for Generalizable Reinforcement Learning
Viaarxiv icon