Alert button
Picture for Keith Ross

Keith Ross

Alert button

Cross Entropy versus Label Smoothing: A Neural Collapse Perspective

Add code
Bookmark button
Alert button
Feb 07, 2024
Li Guo, Keith Ross, Zifan Zhao, George Andriopoulos, Shuyang Ling, Yufeng Xu, Zixuan Dong

Viaarxiv icon

Pre-training with Synthetic Data Helps Offline Reinforcement Learning

Add code
Bookmark button
Alert button
Oct 06, 2023
Zecheng Wang, Che Wang, Zixuan Dong, Keith Ross

Figure 1 for Pre-training with Synthetic Data Helps Offline Reinforcement Learning
Figure 2 for Pre-training with Synthetic Data Helps Offline Reinforcement Learning
Figure 3 for Pre-training with Synthetic Data Helps Offline Reinforcement Learning
Figure 4 for Pre-training with Synthetic Data Helps Offline Reinforcement Learning
Viaarxiv icon

On the Convergence of Monte Carlo UCB for Random-Length Episodic MDPs

Add code
Bookmark button
Alert button
Sep 07, 2022
Zixuan Dong, Che Wang, Keith Ross

Figure 1 for On the Convergence of Monte Carlo UCB for Random-Length Episodic MDPs
Figure 2 for On the Convergence of Monte Carlo UCB for Random-Length Episodic MDPs
Figure 3 for On the Convergence of Monte Carlo UCB for Random-Length Episodic MDPs
Figure 4 for On the Convergence of Monte Carlo UCB for Random-Length Episodic MDPs
Viaarxiv icon

VRL3: A Data-Driven Framework for Visual Deep Reinforcement Learning

Add code
Bookmark button
Alert button
Feb 17, 2022
Che Wang, Xufang Luo, Keith Ross, Dongsheng Li

Figure 1 for VRL3: A Data-Driven Framework for Visual Deep Reinforcement Learning
Figure 2 for VRL3: A Data-Driven Framework for Visual Deep Reinforcement Learning
Figure 3 for VRL3: A Data-Driven Framework for Visual Deep Reinforcement Learning
Figure 4 for VRL3: A Data-Driven Framework for Visual Deep Reinforcement Learning
Viaarxiv icon

Randomized Ensembled Double Q-Learning: Learning Fast Without a Model

Add code
Bookmark button
Alert button
Jan 15, 2021
Xinyue Chen, Che Wang, Zijian Zhou, Keith Ross

Figure 1 for Randomized Ensembled Double Q-Learning: Learning Fast Without a Model
Figure 2 for Randomized Ensembled Double Q-Learning: Learning Fast Without a Model
Figure 3 for Randomized Ensembled Double Q-Learning: Learning Fast Without a Model
Figure 4 for Randomized Ensembled Double Q-Learning: Learning Fast Without a Model
Viaarxiv icon

On the Convergence of the Monte Carlo Exploring Starts Algorithm for Reinforcement Learning

Add code
Bookmark button
Alert button
Feb 10, 2020
Che Wang, Keith Ross

Figure 1 for On the Convergence of the Monte Carlo Exploring Starts Algorithm for Reinforcement Learning
Viaarxiv icon

BAIL: Best-Action Imitation Learning for Batch Deep Reinforcement Learning

Add code
Bookmark button
Alert button
Oct 27, 2019
Xinyue Chen, Zijian Zhou, Zheng Wang, Che Wang, Yanqiu Wu, Qing Deng, Keith Ross

Figure 1 for BAIL: Best-Action Imitation Learning for Batch Deep Reinforcement Learning
Figure 2 for BAIL: Best-Action Imitation Learning for Batch Deep Reinforcement Learning
Figure 3 for BAIL: Best-Action Imitation Learning for Batch Deep Reinforcement Learning
Figure 4 for BAIL: Best-Action Imitation Learning for Batch Deep Reinforcement Learning
Viaarxiv icon

Towards Simplicity in Deep Reinforcement Learning: Streamlined Off-Policy Learning

Add code
Bookmark button
Alert button
Oct 10, 2019
Che Wang, Yanqiu Wu, Quan Vuong, Keith Ross

Figure 1 for Towards Simplicity in Deep Reinforcement Learning: Streamlined Off-Policy Learning
Figure 2 for Towards Simplicity in Deep Reinforcement Learning: Streamlined Off-Policy Learning
Figure 3 for Towards Simplicity in Deep Reinforcement Learning: Streamlined Off-Policy Learning
Figure 4 for Towards Simplicity in Deep Reinforcement Learning: Streamlined Off-Policy Learning
Viaarxiv icon

Boosting Soft Actor-Critic: Emphasizing Recent Experience without Forgetting the Past

Add code
Bookmark button
Alert button
Jun 10, 2019
Che Wang, Keith Ross

Figure 1 for Boosting Soft Actor-Critic: Emphasizing Recent Experience without Forgetting the Past
Figure 2 for Boosting Soft Actor-Critic: Emphasizing Recent Experience without Forgetting the Past
Figure 3 for Boosting Soft Actor-Critic: Emphasizing Recent Experience without Forgetting the Past
Figure 4 for Boosting Soft Actor-Critic: Emphasizing Recent Experience without Forgetting the Past
Viaarxiv icon