Picture for Vincent Liu

Vincent Liu

Dynamic Rebatching for Efficient Early-Exit Inference with DREX

Add code
Dec 17, 2025
Viaarxiv icon

EgoZero: Robot Learning from Smart Glasses

Add code
May 26, 2025
Figure 1 for EgoZero: Robot Learning from Smart Glasses
Figure 2 for EgoZero: Robot Learning from Smart Glasses
Figure 3 for EgoZero: Robot Learning from Smart Glasses
Figure 4 for EgoZero: Robot Learning from Smart Glasses
Viaarxiv icon

Carbon Connect: An Ecosystem for Sustainable Computing

Add code
May 22, 2024
Figure 1 for Carbon Connect: An Ecosystem for Sustainable Computing
Figure 2 for Carbon Connect: An Ecosystem for Sustainable Computing
Figure 3 for Carbon Connect: An Ecosystem for Sustainable Computing
Figure 4 for Carbon Connect: An Ecosystem for Sustainable Computing
Viaarxiv icon

Switching the Loss Reduces the Cost in Batch Reinforcement Learning

Add code
Mar 12, 2024
Figure 1 for Switching the Loss Reduces the Cost in Batch Reinforcement Learning
Figure 2 for Switching the Loss Reduces the Cost in Batch Reinforcement Learning
Figure 3 for Switching the Loss Reduces the Cost in Batch Reinforcement Learning
Figure 4 for Switching the Loss Reduces the Cost in Batch Reinforcement Learning
Viaarxiv icon

Under the Surface: Tracking the Artifactuality of LLM-Generated Data

Add code
Jan 30, 2024
Figure 1 for Under the Surface: Tracking the Artifactuality of LLM-Generated Data
Figure 2 for Under the Surface: Tracking the Artifactuality of LLM-Generated Data
Figure 3 for Under the Surface: Tracking the Artifactuality of LLM-Generated Data
Figure 4 for Under the Surface: Tracking the Artifactuality of LLM-Generated Data
Viaarxiv icon

When is Offline Policy Selection Sample Efficient for Reinforcement Learning?

Add code
Dec 04, 2023
Figure 1 for When is Offline Policy Selection Sample Efficient for Reinforcement Learning?
Figure 2 for When is Offline Policy Selection Sample Efficient for Reinforcement Learning?
Figure 3 for When is Offline Policy Selection Sample Efficient for Reinforcement Learning?
Figure 4 for When is Offline Policy Selection Sample Efficient for Reinforcement Learning?
Viaarxiv icon

Measuring and Mitigating Interference in Reinforcement Learning

Add code
Jul 10, 2023
Figure 1 for Measuring and Mitigating Interference in Reinforcement Learning
Figure 2 for Measuring and Mitigating Interference in Reinforcement Learning
Figure 3 for Measuring and Mitigating Interference in Reinforcement Learning
Figure 4 for Measuring and Mitigating Interference in Reinforcement Learning
Viaarxiv icon

Asymptotically Unbiased Off-Policy Policy Evaluation when Reusing Old Data in Nonstationary Environments

Add code
Feb 23, 2023
Figure 1 for Asymptotically Unbiased Off-Policy Policy Evaluation when Reusing Old Data in Nonstationary Environments
Figure 2 for Asymptotically Unbiased Off-Policy Policy Evaluation when Reusing Old Data in Nonstationary Environments
Figure 3 for Asymptotically Unbiased Off-Policy Policy Evaluation when Reusing Old Data in Nonstationary Environments
Figure 4 for Asymptotically Unbiased Off-Policy Policy Evaluation when Reusing Old Data in Nonstationary Environments
Viaarxiv icon

AlpaServe: Statistical Multiplexing with Model Parallelism for Deep Learning Serving

Add code
Feb 22, 2023
Viaarxiv icon

No More Pesky Hyperparameters: Offline Hyperparameter Tuning for RL

Add code
May 18, 2022
Figure 1 for No More Pesky Hyperparameters: Offline Hyperparameter Tuning for RL
Figure 2 for No More Pesky Hyperparameters: Offline Hyperparameter Tuning for RL
Figure 3 for No More Pesky Hyperparameters: Offline Hyperparameter Tuning for RL
Figure 4 for No More Pesky Hyperparameters: Offline Hyperparameter Tuning for RL
Viaarxiv icon