Alert button
Picture for Ed Chi

Ed Chi

Alert button

Self-Consistency Improves Chain of Thought Reasoning in Language Models

Apr 06, 2022
Xuezhi Wang, Jason Wei, Dale Schuurmans, Quoc Le, Ed Chi, Sharan Narang, Aakanksha Chowdhery, Denny Zhou

Figure 1 for Self-Consistency Improves Chain of Thought Reasoning in Language Models
Figure 2 for Self-Consistency Improves Chain of Thought Reasoning in Language Models
Figure 3 for Self-Consistency Improves Chain of Thought Reasoning in Language Models
Figure 4 for Self-Consistency Improves Chain of Thought Reasoning in Language Models
Viaarxiv icon

LaMDA: Language Models for Dialog Applications

Feb 10, 2022
Romal Thoppilan, Daniel De Freitas, Jamie Hall, Noam Shazeer, Apoorv Kulshreshtha, Heng-Tze Cheng, Alicia Jin, Taylor Bos, Leslie Baker, Yu Du, YaGuang Li, Hongrae Lee, Huaixiu Steven Zheng, Amin Ghafouri, Marcelo Menegali, Yanping Huang, Maxim Krikun, Dmitry Lepikhin, James Qin, Dehao Chen, Yuanzhong Xu, Zhifeng Chen, Adam Roberts, Maarten Bosma, Vincent Zhao, Yanqi Zhou, Chung-Ching Chang, Igor Krivokon, Will Rusch, Marc Pickett, Pranesh Srinivasan, Laichee Man, Kathleen Meier-Hellstern, Meredith Ringel Morris, Tulsee Doshi, Renelito Delos Santos, Toju Duke, Johnny Soraker, Ben Zevenbergen, Vinodkumar Prabhakaran, Mark Diaz, Ben Hutchinson, Kristen Olson, Alejandra Molina, Erin Hoffman-John, Josh Lee, Lora Aroyo, Ravi Rajakumar, Alena Butryna, Matthew Lamm, Viktoriya Kuzmina, Joe Fenton, Aaron Cohen, Rachel Bernstein, Ray Kurzweil, Blaise Aguera-Arcas, Claire Cui, Marian Croak, Ed Chi, Quoc Le

Figure 1 for LaMDA: Language Models for Dialog Applications
Figure 2 for LaMDA: Language Models for Dialog Applications
Figure 3 for LaMDA: Language Models for Dialog Applications
Figure 4 for LaMDA: Language Models for Dialog Applications
Viaarxiv icon

Chain of Thought Prompting Elicits Reasoning in Large Language Models

Jan 28, 2022
Jason Wei, Xuezhi Wang, Dale Schuurmans, Maarten Bosma, Ed Chi, Quoc Le, Denny Zhou

Figure 1 for Chain of Thought Prompting Elicits Reasoning in Large Language Models
Figure 2 for Chain of Thought Prompting Elicits Reasoning in Large Language Models
Figure 3 for Chain of Thought Prompting Elicits Reasoning in Large Language Models
Figure 4 for Chain of Thought Prompting Elicits Reasoning in Large Language Models
Viaarxiv icon

Recency Dropout for Recurrent Recommender Systems

Jan 26, 2022
Bo Chang, Can Xu, Matthieu Lê, Jingchen Feng, Ya Le, Sriraj Badam, Ed Chi, Minmin Chen

Figure 1 for Recency Dropout for Recurrent Recommender Systems
Figure 2 for Recency Dropout for Recurrent Recommender Systems
Figure 3 for Recency Dropout for Recurrent Recommender Systems
Figure 4 for Recency Dropout for Recurrent Recommender Systems
Viaarxiv icon

CAT-Gen: Improving Robustness in NLP Models via Controlled Adversarial Text Generation

Oct 05, 2020
Tianlu Wang, Xuezhi Wang, Yao Qin, Ben Packer, Kang Li, Jilin Chen, Alex Beutel, Ed Chi

Figure 1 for CAT-Gen: Improving Robustness in NLP Models via Controlled Adversarial Text Generation
Figure 2 for CAT-Gen: Improving Robustness in NLP Models via Controlled Adversarial Text Generation
Figure 3 for CAT-Gen: Improving Robustness in NLP Models via Controlled Adversarial Text Generation
Figure 4 for CAT-Gen: Improving Robustness in NLP Models via Controlled Adversarial Text Generation
Viaarxiv icon

Developing a Recommendation Benchmark for MLPerf Training and Inference

Mar 16, 2020
Carole-Jean Wu, Robin Burke, Ed Chi, Joseph Konstan, Julian McAuley, Yves Raimond, Hao Zhang

Viaarxiv icon

Data Efficient Training for Reinforcement Learning with Adaptive Behavior Policy Sharing

Feb 12, 2020
Ge Liu, Rui Wu, Heng-Tze Cheng, Jing Wang, Jayden Ooi, Lihong Li, Ang Li, Wai Lok Sibon Li, Craig Boutilier, Ed Chi

Figure 1 for Data Efficient Training for Reinforcement Learning with Adaptive Behavior Policy Sharing
Figure 2 for Data Efficient Training for Reinforcement Learning with Adaptive Behavior Policy Sharing
Figure 3 for Data Efficient Training for Reinforcement Learning with Adaptive Behavior Policy Sharing
Figure 4 for Data Efficient Training for Reinforcement Learning with Adaptive Behavior Policy Sharing
Viaarxiv icon

BRPO: Batch Residual Policy Optimization

Feb 08, 2020
Sungryull Sohn, Yinlam Chow, Jayden Ooi, Ofir Nachum, Honglak Lee, Ed Chi, Craig Boutilier

Figure 1 for BRPO: Batch Residual Policy Optimization
Figure 2 for BRPO: Batch Residual Policy Optimization
Figure 3 for BRPO: Batch Residual Policy Optimization
Figure 4 for BRPO: Batch Residual Policy Optimization
Viaarxiv icon

Top-K Off-Policy Correction for a REINFORCE Recommender System

Dec 06, 2018
Minmin Chen, Alex Beutel, Paul Covington, Sagar Jain, Francois Belletti, Ed Chi

Figure 1 for Top-K Off-Policy Correction for a REINFORCE Recommender System
Figure 2 for Top-K Off-Policy Correction for a REINFORCE Recommender System
Figure 3 for Top-K Off-Policy Correction for a REINFORCE Recommender System
Figure 4 for Top-K Off-Policy Correction for a REINFORCE Recommender System
Viaarxiv icon