Alert button
Picture for A. Rupam Mahmood

A. Rupam Mahmood

Alert button

Addressing Loss of Plasticity and Catastrophic Forgetting in Continual Learning

Add code
Bookmark button
Alert button
Mar 31, 2024
Mohamed Elsayed, A. Rupam Mahmood

Viaarxiv icon

MaDi: Learning to Mask Distractions for Generalization in Visual Deep Reinforcement Learning

Add code
Bookmark button
Alert button
Dec 23, 2023
Bram Grooten, Tristan Tomilin, Gautham Vasan, Matthew E. Taylor, A. Rupam Mahmood, Meng Fang, Mykola Pechenizkiy, Decebal Constantin Mocanu

Viaarxiv icon

Elephant Neural Networks: Born to Be a Continual Learner

Add code
Bookmark button
Alert button
Oct 02, 2023
Qingfeng Lan, A. Rupam Mahmood

Viaarxiv icon

Maintaining Plasticity in Deep Continual Learning

Add code
Bookmark button
Alert button
Jun 23, 2023
Shibhansh Dohare, J. Fernando Hernandez-Garcia, Parash Rahman, Richard S. Sutton, A. Rupam Mahmood

Figure 1 for Maintaining Plasticity in Deep Continual Learning
Figure 2 for Maintaining Plasticity in Deep Continual Learning
Figure 3 for Maintaining Plasticity in Deep Continual Learning
Figure 4 for Maintaining Plasticity in Deep Continual Learning
Viaarxiv icon

Correcting discount-factor mismatch in on-policy policy gradient methods

Add code
Bookmark button
Alert button
Jun 23, 2023
Fengdi Che, Gautham Vasan, A. Rupam Mahmood

Figure 1 for Correcting discount-factor mismatch in on-policy policy gradient methods
Figure 2 for Correcting discount-factor mismatch in on-policy policy gradient methods
Figure 3 for Correcting discount-factor mismatch in on-policy policy gradient methods
Figure 4 for Correcting discount-factor mismatch in on-policy policy gradient methods
Viaarxiv icon

Provable and Practical: Efficient Exploration in Reinforcement Learning via Langevin Monte Carlo

Add code
Bookmark button
Alert button
May 29, 2023
Haque Ishfaq, Qingfeng Lan, Pan Xu, A. Rupam Mahmood, Doina Precup, Anima Anandkumar, Kamyar Azizzadenesheli

Figure 1 for Provable and Practical: Efficient Exploration in Reinforcement Learning via Langevin Monte Carlo
Figure 2 for Provable and Practical: Efficient Exploration in Reinforcement Learning via Langevin Monte Carlo
Figure 3 for Provable and Practical: Efficient Exploration in Reinforcement Learning via Langevin Monte Carlo
Figure 4 for Provable and Practical: Efficient Exploration in Reinforcement Learning via Langevin Monte Carlo
Viaarxiv icon

Reducing the Cost of Cycle-Time Tuning for Real-World Policy Optimization

Add code
Bookmark button
Alert button
May 09, 2023
Homayoon Farrahi, A. Rupam Mahmood

Figure 1 for Reducing the Cost of Cycle-Time Tuning for Real-World Policy Optimization
Figure 2 for Reducing the Cost of Cycle-Time Tuning for Real-World Policy Optimization
Figure 3 for Reducing the Cost of Cycle-Time Tuning for Real-World Policy Optimization
Figure 4 for Reducing the Cost of Cycle-Time Tuning for Real-World Policy Optimization
Viaarxiv icon

Utility-based Perturbed Gradient Descent: An Optimizer for Continual Learning

Add code
Bookmark button
Alert button
Feb 07, 2023
Mohamed Elsayed, A. Rupam Mahmood

Figure 1 for Utility-based Perturbed Gradient Descent: An Optimizer for Continual Learning
Figure 2 for Utility-based Perturbed Gradient Descent: An Optimizer for Continual Learning
Figure 3 for Utility-based Perturbed Gradient Descent: An Optimizer for Continual Learning
Figure 4 for Utility-based Perturbed Gradient Descent: An Optimizer for Continual Learning
Viaarxiv icon

Learning to Optimize for Reinforcement Learning

Add code
Bookmark button
Alert button
Feb 03, 2023
Qingfeng Lan, A. Rupam Mahmood, Shuicheng Yan, Zhongwen Xu

Figure 1 for Learning to Optimize for Reinforcement Learning
Figure 2 for Learning to Optimize for Reinforcement Learning
Figure 3 for Learning to Optimize for Reinforcement Learning
Figure 4 for Learning to Optimize for Reinforcement Learning
Viaarxiv icon

Variable-Decision Frequency Option Critic

Add code
Bookmark button
Alert button
Dec 11, 2022
Amirmohammad Karimi, Jun Jin, Jun Luo, A. Rupam Mahmood, Martin Jagersand, Samuele Tosatto

Figure 1 for Variable-Decision Frequency Option Critic
Figure 2 for Variable-Decision Frequency Option Critic
Figure 3 for Variable-Decision Frequency Option Critic
Viaarxiv icon