Picture for Junfeng Wen

Junfeng Wen

An MRP Formulation for Supervised Learning: Generalized Temporal Difference Learning Models

Add code
Apr 23, 2024
Figure 1 for An MRP Formulation for Supervised Learning: Generalized Temporal Difference Learning Models
Figure 2 for An MRP Formulation for Supervised Learning: Generalized Temporal Difference Learning Models
Figure 3 for An MRP Formulation for Supervised Learning: Generalized Temporal Difference Learning Models
Figure 4 for An MRP Formulation for Supervised Learning: Generalized Temporal Difference Learning Models
Viaarxiv icon

Find Your Friends: Personalized Federated Learning with the Right Collaborators

Add code
Oct 14, 2022
Figure 1 for Find Your Friends: Personalized Federated Learning with the Right Collaborators
Figure 2 for Find Your Friends: Personalized Federated Learning with the Right Collaborators
Figure 3 for Find Your Friends: Personalized Federated Learning with the Right Collaborators
Figure 4 for Find Your Friends: Personalized Federated Learning with the Right Collaborators
Viaarxiv icon

A Parametric Class of Approximate Gradient Updates for Policy Optimization

Add code
Jun 17, 2022
Figure 1 for A Parametric Class of Approximate Gradient Updates for Policy Optimization
Figure 2 for A Parametric Class of Approximate Gradient Updates for Policy Optimization
Figure 3 for A Parametric Class of Approximate Gradient Updates for Policy Optimization
Figure 4 for A Parametric Class of Approximate Gradient Updates for Policy Optimization
Viaarxiv icon

ProxyFL: Decentralized Federated Learning through Proxy Model Sharing

Add code
Nov 22, 2021
Figure 1 for ProxyFL: Decentralized Federated Learning through Proxy Model Sharing
Figure 2 for ProxyFL: Decentralized Federated Learning through Proxy Model Sharing
Figure 3 for ProxyFL: Decentralized Federated Learning through Proxy Model Sharing
Figure 4 for ProxyFL: Decentralized Federated Learning through Proxy Model Sharing
Viaarxiv icon

Characterizing the Gap Between Actor-Critic and Policy Gradient

Add code
Jun 13, 2021
Figure 1 for Characterizing the Gap Between Actor-Critic and Policy Gradient
Figure 2 for Characterizing the Gap Between Actor-Critic and Policy Gradient
Figure 3 for Characterizing the Gap Between Actor-Critic and Policy Gradient
Figure 4 for Characterizing the Gap Between Actor-Critic and Policy Gradient
Viaarxiv icon

Batch Stationary Distribution Estimation

Add code
Mar 02, 2020
Figure 1 for Batch Stationary Distribution Estimation
Figure 2 for Batch Stationary Distribution Estimation
Figure 3 for Batch Stationary Distribution Estimation
Figure 4 for Batch Stationary Distribution Estimation
Viaarxiv icon

Universal Successor Features for Transfer Reinforcement Learning

Add code
Jan 05, 2020
Figure 1 for Universal Successor Features for Transfer Reinforcement Learning
Figure 2 for Universal Successor Features for Transfer Reinforcement Learning
Figure 3 for Universal Successor Features for Transfer Reinforcement Learning
Figure 4 for Universal Successor Features for Transfer Reinforcement Learning
Viaarxiv icon

Domain Aggregation Networks for Multi-Source Domain Adaptation

Add code
Sep 25, 2019
Figure 1 for Domain Aggregation Networks for Multi-Source Domain Adaptation
Figure 2 for Domain Aggregation Networks for Multi-Source Domain Adaptation
Figure 3 for Domain Aggregation Networks for Multi-Source Domain Adaptation
Figure 4 for Domain Aggregation Networks for Multi-Source Domain Adaptation
Viaarxiv icon

Few-Shot Self Reminder to Overcome Catastrophic Forgetting

Add code
Dec 03, 2018
Figure 1 for Few-Shot Self Reminder to Overcome Catastrophic Forgetting
Figure 2 for Few-Shot Self Reminder to Overcome Catastrophic Forgetting
Viaarxiv icon

Universal Successor Representations for Transfer Reinforcement Learning

Add code
Apr 11, 2018
Figure 1 for Universal Successor Representations for Transfer Reinforcement Learning
Figure 2 for Universal Successor Representations for Transfer Reinforcement Learning
Viaarxiv icon