Alert button
Picture for Guy Tennenholtz

Guy Tennenholtz

Alert button

DynaMITE-RL: A Dynamic Model for Improved Temporal Meta-Reinforcement Learning

Add code
Bookmark button
Alert button
Feb 25, 2024
Anthony Liang, Guy Tennenholtz, Chih-wei Hsu, Yinlam Chow, Erdem Bıyık, Craig Boutilier

Viaarxiv icon

Ever Evolving Evaluator (EV3): Towards Flexible and Reliable Meta-Optimization for Knowledge Distillation

Add code
Bookmark button
Alert button
Oct 29, 2023
Li Ding, Masrour Zoghi, Guy Tennenholtz, Maryam Karimzadehgan

Figure 1 for Ever Evolving Evaluator (EV3): Towards Flexible and Reliable Meta-Optimization for Knowledge Distillation
Viaarxiv icon

Factual and Personalized Recommendations using Language Models and Reinforcement Learning

Add code
Bookmark button
Alert button
Oct 09, 2023
Jihwan Jeong, Yinlam Chow, Guy Tennenholtz, Chih-Wei Hsu, Azamat Tulepbergenov, Mohammad Ghavamzadeh, Craig Boutilier

Figure 1 for Factual and Personalized Recommendations using Language Models and Reinforcement Learning
Figure 2 for Factual and Personalized Recommendations using Language Models and Reinforcement Learning
Figure 3 for Factual and Personalized Recommendations using Language Models and Reinforcement Learning
Figure 4 for Factual and Personalized Recommendations using Language Models and Reinforcement Learning
Viaarxiv icon

Demystifying Embedding Spaces using Large Language Models

Add code
Bookmark button
Alert button
Oct 06, 2023
Guy Tennenholtz, Yinlam Chow, Chih-Wei Hsu, Jihwan Jeong, Lior Shani, Azamat Tulepbergenov, Deepak Ramachandran, Martin Mladenov, Craig Boutilier

Figure 1 for Demystifying Embedding Spaces using Large Language Models
Figure 2 for Demystifying Embedding Spaces using Large Language Models
Figure 3 for Demystifying Embedding Spaces using Large Language Models
Figure 4 for Demystifying Embedding Spaces using Large Language Models
Viaarxiv icon

Modeling Recommender Ecosystems: Research Challenges at the Intersection of Mechanism Design, Reinforcement Learning and Generative Models

Add code
Bookmark button
Alert button
Sep 22, 2023
Craig Boutilier, Martin Mladenov, Guy Tennenholtz

Figure 1 for Modeling Recommender Ecosystems: Research Challenges at the Intersection of Mechanism Design, Reinforcement Learning and Generative Models
Viaarxiv icon

A Convex Relaxation Approach to Bayesian Regret Minimization in Offline Bandits

Add code
Bookmark button
Alert button
Jun 02, 2023
Mohammad Ghavamzadeh, Marek Petrik, Guy Tennenholtz

Figure 1 for A Convex Relaxation Approach to Bayesian Regret Minimization in Offline Bandits
Figure 2 for A Convex Relaxation Approach to Bayesian Regret Minimization in Offline Bandits
Figure 3 for A Convex Relaxation Approach to Bayesian Regret Minimization in Offline Bandits
Figure 4 for A Convex Relaxation Approach to Bayesian Regret Minimization in Offline Bandits
Viaarxiv icon

Delphic Offline Reinforcement Learning under Nonidentifiable Hidden Confounding

Add code
Bookmark button
Alert button
Jun 01, 2023
Alizée Pace, Hugo Yèche, Bernhard Schölkopf, Gunnar Rätsch, Guy Tennenholtz

Figure 1 for Delphic Offline Reinforcement Learning under Nonidentifiable Hidden Confounding
Figure 2 for Delphic Offline Reinforcement Learning under Nonidentifiable Hidden Confounding
Figure 3 for Delphic Offline Reinforcement Learning under Nonidentifiable Hidden Confounding
Figure 4 for Delphic Offline Reinforcement Learning under Nonidentifiable Hidden Confounding
Viaarxiv icon

Representation-Driven Reinforcement Learning

Add code
Bookmark button
Alert button
May 31, 2023
Ofir Nabati, Guy Tennenholtz, Shie Mannor

Figure 1 for Representation-Driven Reinforcement Learning
Figure 2 for Representation-Driven Reinforcement Learning
Figure 3 for Representation-Driven Reinforcement Learning
Figure 4 for Representation-Driven Reinforcement Learning
Viaarxiv icon

Reinforcement Learning with History-Dependent Dynamic Contexts

Add code
Bookmark button
Alert button
Feb 04, 2023
Guy Tennenholtz, Nadav Merlis, Lior Shani, Martin Mladenov, Craig Boutilier

Figure 1 for Reinforcement Learning with History-Dependent Dynamic Contexts
Figure 2 for Reinforcement Learning with History-Dependent Dynamic Contexts
Viaarxiv icon