Picture for David Brandfonbrener

David Brandfonbrener

Deconstructing What Makes a Good Optimizer for Language Models

Add code
Jul 10, 2024
Viaarxiv icon

Universal Length Generalization with Turing Programs

Add code
Jul 03, 2024
Figure 1 for Universal Length Generalization with Turing Programs
Figure 2 for Universal Length Generalization with Turing Programs
Figure 3 for Universal Length Generalization with Turing Programs
Figure 4 for Universal Length Generalization with Turing Programs
Viaarxiv icon

CoLoR-Filter: Conditional Loss Reduction Filtering for Targeted Language Model Pre-training

Add code
Jun 15, 2024
Viaarxiv icon

Q-Probe: A Lightweight Approach to Reward Maximization for Language Models

Add code
Feb 22, 2024
Figure 1 for Q-Probe: A Lightweight Approach to Reward Maximization for Language Models
Figure 2 for Q-Probe: A Lightweight Approach to Reward Maximization for Language Models
Figure 3 for Q-Probe: A Lightweight Approach to Reward Maximization for Language Models
Figure 4 for Q-Probe: A Lightweight Approach to Reward Maximization for Language Models
Viaarxiv icon

Verified Multi-Step Synthesis using Large Language Models and Monte Carlo Tree Search

Add code
Feb 13, 2024
Viaarxiv icon

Repeat After Me: Transformers are Better than State Space Models at Copying

Add code
Feb 01, 2024
Figure 1 for Repeat After Me: Transformers are Better than State Space Models at Copying
Figure 2 for Repeat After Me: Transformers are Better than State Space Models at Copying
Figure 3 for Repeat After Me: Transformers are Better than State Space Models at Copying
Figure 4 for Repeat After Me: Transformers are Better than State Space Models at Copying
Viaarxiv icon

Inverse Dynamics Pretraining Learns Good Representations for Multitask Imitation

Add code
May 26, 2023
Figure 1 for Inverse Dynamics Pretraining Learns Good Representations for Multitask Imitation
Figure 2 for Inverse Dynamics Pretraining Learns Good Representations for Multitask Imitation
Figure 3 for Inverse Dynamics Pretraining Learns Good Representations for Multitask Imitation
Figure 4 for Inverse Dynamics Pretraining Learns Good Representations for Multitask Imitation
Viaarxiv icon

Visual Backtracking Teleoperation: A Data Collection Protocol for Offline Image-Based Reinforcement Learning

Add code
Oct 05, 2022
Figure 1 for Visual Backtracking Teleoperation: A Data Collection Protocol for Offline Image-Based Reinforcement Learning
Figure 2 for Visual Backtracking Teleoperation: A Data Collection Protocol for Offline Image-Based Reinforcement Learning
Figure 3 for Visual Backtracking Teleoperation: A Data Collection Protocol for Offline Image-Based Reinforcement Learning
Figure 4 for Visual Backtracking Teleoperation: A Data Collection Protocol for Offline Image-Based Reinforcement Learning
Viaarxiv icon

Incorporating Explicit Uncertainty Estimates into Deep Offline Reinforcement Learning

Add code
Jun 02, 2022
Figure 1 for Incorporating Explicit Uncertainty Estimates into Deep Offline Reinforcement Learning
Figure 2 for Incorporating Explicit Uncertainty Estimates into Deep Offline Reinforcement Learning
Figure 3 for Incorporating Explicit Uncertainty Estimates into Deep Offline Reinforcement Learning
Figure 4 for Incorporating Explicit Uncertainty Estimates into Deep Offline Reinforcement Learning
Viaarxiv icon

When does return-conditioned supervised learning work for offline reinforcement learning?

Add code
Jun 02, 2022
Figure 1 for When does return-conditioned supervised learning work for offline reinforcement learning?
Figure 2 for When does return-conditioned supervised learning work for offline reinforcement learning?
Figure 3 for When does return-conditioned supervised learning work for offline reinforcement learning?
Figure 4 for When does return-conditioned supervised learning work for offline reinforcement learning?
Viaarxiv icon