Alert button
Picture for Yinlam Chow

Yinlam Chow

Alert button

DynaMITE-RL: A Dynamic Model for Improved Temporal Meta-Reinforcement Learning

Add code
Bookmark button
Alert button
Feb 25, 2024
Anthony Liang, Guy Tennenholtz, Chih-wei Hsu, Yinlam Chow, Erdem Bıyık, Craig Boutilier

Viaarxiv icon

Preference Elicitation with Soft Attributes in Interactive Recommendation

Add code
Bookmark button
Alert button
Oct 22, 2023
Erdem Biyik, Fan Yao, Yinlam Chow, Alex Haig, Chih-wei Hsu, Mohammad Ghavamzadeh, Craig Boutilier

Viaarxiv icon

Factual and Personalized Recommendations using Language Models and Reinforcement Learning

Add code
Bookmark button
Alert button
Oct 09, 2023
Jihwan Jeong, Yinlam Chow, Guy Tennenholtz, Chih-Wei Hsu, Azamat Tulepbergenov, Mohammad Ghavamzadeh, Craig Boutilier

Figure 1 for Factual and Personalized Recommendations using Language Models and Reinforcement Learning
Figure 2 for Factual and Personalized Recommendations using Language Models and Reinforcement Learning
Figure 3 for Factual and Personalized Recommendations using Language Models and Reinforcement Learning
Figure 4 for Factual and Personalized Recommendations using Language Models and Reinforcement Learning
Viaarxiv icon

Demystifying Embedding Spaces using Large Language Models

Add code
Bookmark button
Alert button
Oct 06, 2023
Guy Tennenholtz, Yinlam Chow, Chih-Wei Hsu, Jihwan Jeong, Lior Shani, Azamat Tulepbergenov, Deepak Ramachandran, Martin Mladenov, Craig Boutilier

Figure 1 for Demystifying Embedding Spaces using Large Language Models
Figure 2 for Demystifying Embedding Spaces using Large Language Models
Figure 3 for Demystifying Embedding Spaces using Large Language Models
Figure 4 for Demystifying Embedding Spaces using Large Language Models
Viaarxiv icon

Offline Reinforcement Learning for Mixture-of-Expert Dialogue Management

Add code
Bookmark button
Alert button
Feb 21, 2023
Dhawal Gupta, Yinlam Chow, Mohammad Ghavamzadeh, Craig Boutilier

Figure 1 for Offline Reinforcement Learning for Mixture-of-Expert Dialogue Management
Figure 2 for Offline Reinforcement Learning for Mixture-of-Expert Dialogue Management
Figure 3 for Offline Reinforcement Learning for Mixture-of-Expert Dialogue Management
Figure 4 for Offline Reinforcement Learning for Mixture-of-Expert Dialogue Management
Viaarxiv icon

Dynamic Planning in Open-Ended Dialogue using Reinforcement Learning

Add code
Bookmark button
Alert button
Jul 25, 2022
Deborah Cohen, Moonkyung Ryu, Yinlam Chow, Orgad Keller, Ido Greenberg, Avinatan Hassidim, Michael Fink, Yossi Matias, Idan Szpektor, Craig Boutilier, Gal Elidan

Figure 1 for Dynamic Planning in Open-Ended Dialogue using Reinforcement Learning
Figure 2 for Dynamic Planning in Open-Ended Dialogue using Reinforcement Learning
Figure 3 for Dynamic Planning in Open-Ended Dialogue using Reinforcement Learning
Figure 4 for Dynamic Planning in Open-Ended Dialogue using Reinforcement Learning
Viaarxiv icon

A Mixture-of-Expert Approach to RL-based Dialogue Management

Add code
Bookmark button
Alert button
May 31, 2022
Yinlam Chow, Aza Tulepbergenov, Ofir Nachum, MoonKyung Ryu, Mohammad Ghavamzadeh, Craig Boutilier

Figure 1 for A Mixture-of-Expert Approach to RL-based Dialogue Management
Figure 2 for A Mixture-of-Expert Approach to RL-based Dialogue Management
Figure 3 for A Mixture-of-Expert Approach to RL-based Dialogue Management
Figure 4 for A Mixture-of-Expert Approach to RL-based Dialogue Management
Viaarxiv icon

Efficient Risk-Averse Reinforcement Learning

Add code
Bookmark button
Alert button
May 10, 2022
Ido Greenberg, Yinlam Chow, Mohammad Ghavamzadeh, Shie Mannor

Figure 1 for Efficient Risk-Averse Reinforcement Learning
Figure 2 for Efficient Risk-Averse Reinforcement Learning
Figure 3 for Efficient Risk-Averse Reinforcement Learning
Figure 4 for Efficient Risk-Averse Reinforcement Learning
Viaarxiv icon

SAFER: Data-Efficient and Safe Reinforcement Learning via Skill Acquisition

Add code
Bookmark button
Alert button
Feb 10, 2022
Dylan Slack, Yinlam Chow, Bo Dai, Nevan Wichers

Figure 1 for SAFER: Data-Efficient and Safe Reinforcement Learning via Skill Acquisition
Figure 2 for SAFER: Data-Efficient and Safe Reinforcement Learning via Skill Acquisition
Figure 3 for SAFER: Data-Efficient and Safe Reinforcement Learning via Skill Acquisition
Figure 4 for SAFER: Data-Efficient and Safe Reinforcement Learning via Skill Acquisition
Viaarxiv icon

Discovering Personalized Semantics for Soft Attributes in Recommender Systems using Concept Activation Vectors

Add code
Bookmark button
Alert button
Feb 06, 2022
Christina Göpfert, Yinlam Chow, Chih-wei Hsu, Ivan Vendrov, Tyler Lu, Deepak Ramachandran, Craig Boutilier

Figure 1 for Discovering Personalized Semantics for Soft Attributes in Recommender Systems using Concept Activation Vectors
Figure 2 for Discovering Personalized Semantics for Soft Attributes in Recommender Systems using Concept Activation Vectors
Figure 3 for Discovering Personalized Semantics for Soft Attributes in Recommender Systems using Concept Activation Vectors
Figure 4 for Discovering Personalized Semantics for Soft Attributes in Recommender Systems using Concept Activation Vectors
Viaarxiv icon