Picture for Xutong Liu

Xutong Liu

Combinatorial Multivariant Multi-Armed Bandits with Applications to Episodic Reinforcement Learning and Beyond

Add code
Jun 03, 2024
Viaarxiv icon

Cost-Effective Online Multi-LLM Selection with Versatile Reward Models

Add code
May 26, 2024
Viaarxiv icon

Federated Contextual Cascading Bandits with Asynchronous Communication and Heterogeneous Users

Add code
Feb 26, 2024
Figure 1 for Federated Contextual Cascading Bandits with Asynchronous Communication and Heterogeneous Users
Figure 2 for Federated Contextual Cascading Bandits with Asynchronous Communication and Heterogeneous Users
Figure 3 for Federated Contextual Cascading Bandits with Asynchronous Communication and Heterogeneous Users
Figure 4 for Federated Contextual Cascading Bandits with Asynchronous Communication and Heterogeneous Users
Viaarxiv icon

Online Clustering of Bandits with Misspecified User Models

Add code
Oct 10, 2023
Figure 1 for Online Clustering of Bandits with Misspecified User Models
Figure 2 for Online Clustering of Bandits with Misspecified User Models
Viaarxiv icon

Contextual Combinatorial Bandits with Probabilistically Triggered Arms

Add code
Mar 30, 2023
Figure 1 for Contextual Combinatorial Bandits with Probabilistically Triggered Arms
Figure 2 for Contextual Combinatorial Bandits with Probabilistically Triggered Arms
Figure 3 for Contextual Combinatorial Bandits with Probabilistically Triggered Arms
Figure 4 for Contextual Combinatorial Bandits with Probabilistically Triggered Arms
Viaarxiv icon

Efficient Explorative Key-term Selection Strategies for Conversational Contextual Bandits

Add code
Mar 01, 2023
Figure 1 for Efficient Explorative Key-term Selection Strategies for Conversational Contextual Bandits
Figure 2 for Efficient Explorative Key-term Selection Strategies for Conversational Contextual Bandits
Figure 3 for Efficient Explorative Key-term Selection Strategies for Conversational Contextual Bandits
Viaarxiv icon

On-Demand Communication for Asynchronous Multi-Agent Bandits

Add code
Feb 15, 2023
Figure 1 for On-Demand Communication for Asynchronous Multi-Agent Bandits
Figure 2 for On-Demand Communication for Asynchronous Multi-Agent Bandits
Figure 3 for On-Demand Communication for Asynchronous Multi-Agent Bandits
Figure 4 for On-Demand Communication for Asynchronous Multi-Agent Bandits
Viaarxiv icon

Federated Online Clustering of Bandits

Add code
Aug 31, 2022
Figure 1 for Federated Online Clustering of Bandits
Figure 2 for Federated Online Clustering of Bandits
Figure 3 for Federated Online Clustering of Bandits
Figure 4 for Federated Online Clustering of Bandits
Viaarxiv icon

Batch-Size Independent Regret Bounds for Combinatorial Semi-Bandits with Probabilistically Triggered Arms or Independent Arms

Add code
Aug 31, 2022
Figure 1 for Batch-Size Independent Regret Bounds for Combinatorial Semi-Bandits with Probabilistically Triggered Arms or Independent Arms
Figure 2 for Batch-Size Independent Regret Bounds for Combinatorial Semi-Bandits with Probabilistically Triggered Arms or Independent Arms
Figure 3 for Batch-Size Independent Regret Bounds for Combinatorial Semi-Bandits with Probabilistically Triggered Arms or Independent Arms
Figure 4 for Batch-Size Independent Regret Bounds for Combinatorial Semi-Bandits with Probabilistically Triggered Arms or Independent Arms
Viaarxiv icon

Multi-layered Network Exploration via Random Walks: From Offline Optimization to Online Learning

Add code
Jun 09, 2021
Figure 1 for Multi-layered Network Exploration via Random Walks: From Offline Optimization to Online Learning
Figure 2 for Multi-layered Network Exploration via Random Walks: From Offline Optimization to Online Learning
Figure 3 for Multi-layered Network Exploration via Random Walks: From Offline Optimization to Online Learning
Figure 4 for Multi-layered Network Exploration via Random Walks: From Offline Optimization to Online Learning
Viaarxiv icon