Picture for Jian Peng

Jian Peng

School of Information Engineering, Jiangxi Vocational College of Finance & Economics, Jiujiang, China

Characterizing Attacks on Deep Reinforcement Learning

Add code
Jul 24, 2019
Figure 1 for Characterizing Attacks on Deep Reinforcement Learning
Figure 2 for Characterizing Attacks on Deep Reinforcement Learning
Figure 3 for Characterizing Attacks on Deep Reinforcement Learning
Figure 4 for Characterizing Attacks on Deep Reinforcement Learning
Viaarxiv icon

Learning Belief Representations for Imitation Learning in POMDPs

Add code
Jun 22, 2019
Figure 1 for Learning Belief Representations for Imitation Learning in POMDPs
Figure 2 for Learning Belief Representations for Imitation Learning in POMDPs
Figure 3 for Learning Belief Representations for Imitation Learning in POMDPs
Figure 4 for Learning Belief Representations for Imitation Learning in POMDPs
Viaarxiv icon

A gradual, semi-discrete approach to generative network training via explicit Wasserstein minimization

Add code
Jun 11, 2019
Figure 1 for A gradual, semi-discrete approach to generative network training via explicit Wasserstein minimization
Figure 2 for A gradual, semi-discrete approach to generative network training via explicit Wasserstein minimization
Figure 3 for A gradual, semi-discrete approach to generative network training via explicit Wasserstein minimization
Viaarxiv icon

Exploration via Hindsight Goal Generation

Add code
Jun 10, 2019
Figure 1 for Exploration via Hindsight Goal Generation
Figure 2 for Exploration via Hindsight Goal Generation
Figure 3 for Exploration via Hindsight Goal Generation
Figure 4 for Exploration via Hindsight Goal Generation
Viaarxiv icon

Sequence Modeling of Temporal Credit Assignment for Episodic Reinforcement Learning

Add code
May 31, 2019
Figure 1 for Sequence Modeling of Temporal Credit Assignment for Episodic Reinforcement Learning
Figure 2 for Sequence Modeling of Temporal Credit Assignment for Episodic Reinforcement Learning
Figure 3 for Sequence Modeling of Temporal Credit Assignment for Episodic Reinforcement Learning
Viaarxiv icon

Thresholding Bandit with Optimal Aggregate Regret

Add code
May 27, 2019
Figure 1 for Thresholding Bandit with Optimal Aggregate Regret
Figure 2 for Thresholding Bandit with Optimal Aggregate Regret
Figure 3 for Thresholding Bandit with Optimal Aggregate Regret
Figure 4 for Thresholding Bandit with Optimal Aggregate Regret
Viaarxiv icon

Stochastic Variance Reduction for Deep Q-learning

Add code
May 20, 2019
Figure 1 for Stochastic Variance Reduction for Deep Q-learning
Figure 2 for Stochastic Variance Reduction for Deep Q-learning
Figure 3 for Stochastic Variance Reduction for Deep Q-learning
Figure 4 for Stochastic Variance Reduction for Deep Q-learning
Viaarxiv icon

Knowledge Flow: Improve Upon Your Teachers

Add code
Apr 11, 2019
Figure 1 for Knowledge Flow: Improve Upon Your Teachers
Figure 2 for Knowledge Flow: Improve Upon Your Teachers
Figure 3 for Knowledge Flow: Improve Upon Your Teachers
Figure 4 for Knowledge Flow: Improve Upon Your Teachers
Viaarxiv icon

Understanding the Importance of Single Directions via Representative Substitution

Add code
Dec 06, 2018
Figure 1 for Understanding the Importance of Single Directions via Representative Substitution
Figure 2 for Understanding the Importance of Single Directions via Representative Substitution
Figure 3 for Understanding the Importance of Single Directions via Representative Substitution
Figure 4 for Understanding the Importance of Single Directions via Representative Substitution
Viaarxiv icon

Overcoming Catastrophic Forgetting by Soft Parameter Pruning

Add code
Dec 04, 2018
Figure 1 for Overcoming Catastrophic Forgetting by Soft Parameter Pruning
Figure 2 for Overcoming Catastrophic Forgetting by Soft Parameter Pruning
Figure 3 for Overcoming Catastrophic Forgetting by Soft Parameter Pruning
Figure 4 for Overcoming Catastrophic Forgetting by Soft Parameter Pruning
Viaarxiv icon