Alert button
Picture for Silviu Pitis

Silviu Pitis

Alert button

Consistent Aggregation of Objectives with Diverse Time Preferences Requires Non-Markovian Rewards

Add code
Bookmark button
Alert button
Sep 30, 2023
Silviu Pitis

Viaarxiv icon

Identifying the Risks of LM Agents with an LM-Emulated Sandbox

Add code
Bookmark button
Alert button
Sep 25, 2023
Yangjun Ruan, Honghua Dong, Andrew Wang, Silviu Pitis, Yongchao Zhou, Jimmy Ba, Yann Dubois, Chris J. Maddison, Tatsunori Hashimoto

Figure 1 for Identifying the Risks of LM Agents with an LM-Emulated Sandbox
Figure 2 for Identifying the Risks of LM Agents with an LM-Emulated Sandbox
Figure 3 for Identifying the Risks of LM Agents with an LM-Emulated Sandbox
Figure 4 for Identifying the Risks of LM Agents with an LM-Emulated Sandbox
Viaarxiv icon

Boosted Prompt Ensembles for Large Language Models

Add code
Bookmark button
Alert button
Apr 12, 2023
Silviu Pitis, Michael R. Zhang, Andrew Wang, Jimmy Ba

Figure 1 for Boosted Prompt Ensembles for Large Language Models
Figure 2 for Boosted Prompt Ensembles for Large Language Models
Figure 3 for Boosted Prompt Ensembles for Large Language Models
Figure 4 for Boosted Prompt Ensembles for Large Language Models
Viaarxiv icon

Large Language Models Are Human-Level Prompt Engineers

Add code
Bookmark button
Alert button
Nov 03, 2022
Yongchao Zhou, Andrei Ioan Muresanu, Ziwen Han, Keiran Paster, Silviu Pitis, Harris Chan, Jimmy Ba

Figure 1 for Large Language Models Are Human-Level Prompt Engineers
Figure 2 for Large Language Models Are Human-Level Prompt Engineers
Figure 3 for Large Language Models Are Human-Level Prompt Engineers
Figure 4 for Large Language Models Are Human-Level Prompt Engineers
Viaarxiv icon

MoCoDA: Model-based Counterfactual Data Augmentation

Add code
Bookmark button
Alert button
Oct 20, 2022
Silviu Pitis, Elliot Creager, Ajay Mandlekar, Animesh Garg

Figure 1 for MoCoDA: Model-based Counterfactual Data Augmentation
Figure 2 for MoCoDA: Model-based Counterfactual Data Augmentation
Figure 3 for MoCoDA: Model-based Counterfactual Data Augmentation
Figure 4 for MoCoDA: Model-based Counterfactual Data Augmentation
Viaarxiv icon

Counterfactual Data Augmentation using Locally Factored Dynamics

Add code
Bookmark button
Alert button
Jul 06, 2020
Silviu Pitis, Elliot Creager, Animesh Garg

Figure 1 for Counterfactual Data Augmentation using Locally Factored Dynamics
Figure 2 for Counterfactual Data Augmentation using Locally Factored Dynamics
Figure 3 for Counterfactual Data Augmentation using Locally Factored Dynamics
Figure 4 for Counterfactual Data Augmentation using Locally Factored Dynamics
Viaarxiv icon

Maximum Entropy Gain Exploration for Long Horizon Multi-goal Reinforcement Learning

Add code
Bookmark button
Alert button
Jul 06, 2020
Silviu Pitis, Harris Chan, Stephen Zhao, Bradly Stadie, Jimmy Ba

Figure 1 for Maximum Entropy Gain Exploration for Long Horizon Multi-goal Reinforcement Learning
Figure 2 for Maximum Entropy Gain Exploration for Long Horizon Multi-goal Reinforcement Learning
Figure 3 for Maximum Entropy Gain Exploration for Long Horizon Multi-goal Reinforcement Learning
Figure 4 for Maximum Entropy Gain Exploration for Long Horizon Multi-goal Reinforcement Learning
Viaarxiv icon

An Inductive Bias for Distances: Neural Nets that Respect the Triangle Inequality

Add code
Bookmark button
Alert button
Feb 14, 2020
Silviu Pitis, Harris Chan, Kiarash Jamali, Jimmy Ba

Figure 1 for An Inductive Bias for Distances: Neural Nets that Respect the Triangle Inequality
Figure 2 for An Inductive Bias for Distances: Neural Nets that Respect the Triangle Inequality
Figure 3 for An Inductive Bias for Distances: Neural Nets that Respect the Triangle Inequality
Figure 4 for An Inductive Bias for Distances: Neural Nets that Respect the Triangle Inequality
Viaarxiv icon

Objective Social Choice: Using Auxiliary Information to Improve Voting Outcomes

Add code
Bookmark button
Alert button
Jan 27, 2020
Silviu Pitis, Michael R. Zhang

Figure 1 for Objective Social Choice: Using Auxiliary Information to Improve Voting Outcomes
Figure 2 for Objective Social Choice: Using Auxiliary Information to Improve Voting Outcomes
Figure 3 for Objective Social Choice: Using Auxiliary Information to Improve Voting Outcomes
Figure 4 for Objective Social Choice: Using Auxiliary Information to Improve Voting Outcomes
Viaarxiv icon

Fixed-Horizon Temporal Difference Methods for Stable Reinforcement Learning

Add code
Bookmark button
Alert button
Sep 09, 2019
Kristopher De Asis, Alan Chan, Silviu Pitis, Richard S. Sutton, Daniel Graves

Figure 1 for Fixed-Horizon Temporal Difference Methods for Stable Reinforcement Learning
Figure 2 for Fixed-Horizon Temporal Difference Methods for Stable Reinforcement Learning
Figure 3 for Fixed-Horizon Temporal Difference Methods for Stable Reinforcement Learning
Figure 4 for Fixed-Horizon Temporal Difference Methods for Stable Reinforcement Learning
Viaarxiv icon