Alert button
Picture for Percy Liang

Percy Liang

Alert button

Just One Byte (per gradient): A Note on Low-Bandwidth Decentralized Language Model Finetuning Using Shared Randomness

Add code
Bookmark button
Alert button
Jun 16, 2023
Eric Zelikman, Qian Huang, Percy Liang, Nick Haber, Noah D. Goodman

Figure 1 for Just One Byte (per gradient): A Note on Low-Bandwidth Decentralized Language Model Finetuning Using Shared Randomness
Figure 2 for Just One Byte (per gradient): A Note on Low-Bandwidth Decentralized Language Model Finetuning Using Shared Randomness
Figure 3 for Just One Byte (per gradient): A Note on Low-Bandwidth Decentralized Language Model Finetuning Using Shared Randomness
Viaarxiv icon

Anticipatory Music Transformer

Add code
Bookmark button
Alert button
Jun 14, 2023
John Thickstun, David Hall, Chris Donahue, Percy Liang

Figure 1 for Anticipatory Music Transformer
Figure 2 for Anticipatory Music Transformer
Figure 3 for Anticipatory Music Transformer
Figure 4 for Anticipatory Music Transformer
Viaarxiv icon

One-sided Matrix Completion from Two Observations Per Row

Add code
Bookmark button
Alert button
Jun 06, 2023
Steven Cao, Percy Liang, Gregory Valiant

Figure 1 for One-sided Matrix Completion from Two Observations Per Row
Figure 2 for One-sided Matrix Completion from Two Observations Per Row
Figure 3 for One-sided Matrix Completion from Two Observations Per Row
Figure 4 for One-sided Matrix Completion from Two Observations Per Row
Viaarxiv icon

Has the Machine Learning Review Process Become More Arbitrary as the Field Has Grown? The NeurIPS 2021 Consistency Experiment

Add code
Bookmark button
Alert button
Jun 05, 2023
Alina Beygelzimer, Yann N. Dauphin, Percy Liang, Jennifer Wortman Vaughan

Figure 1 for Has the Machine Learning Review Process Become More Arbitrary as the Field Has Grown? The NeurIPS 2021 Consistency Experiment
Figure 2 for Has the Machine Learning Review Process Become More Arbitrary as the Field Has Grown? The NeurIPS 2021 Consistency Experiment
Figure 3 for Has the Machine Learning Review Process Become More Arbitrary as the Field Has Grown? The NeurIPS 2021 Consistency Experiment
Figure 4 for Has the Machine Learning Review Process Become More Arbitrary as the Field Has Grown? The NeurIPS 2021 Consistency Experiment
Viaarxiv icon

Beyond Positive Scaling: How Negation Impacts Scaling Trends of Language Models

Add code
Bookmark button
Alert button
May 27, 2023
Yuhui Zhang, Michihiro Yasunaga, Zhengping Zhou, Jeff Z. HaoChen, James Zou, Percy Liang, Serena Yeung

Figure 1 for Beyond Positive Scaling: How Negation Impacts Scaling Trends of Language Models
Figure 2 for Beyond Positive Scaling: How Negation Impacts Scaling Trends of Language Models
Figure 3 for Beyond Positive Scaling: How Negation Impacts Scaling Trends of Language Models
Figure 4 for Beyond Positive Scaling: How Negation Impacts Scaling Trends of Language Models
Viaarxiv icon

Backpack Language Models

Add code
Bookmark button
Alert button
May 26, 2023
John Hewitt, John Thickstun, Christopher D. Manning, Percy Liang

Figure 1 for Backpack Language Models
Figure 2 for Backpack Language Models
Figure 3 for Backpack Language Models
Figure 4 for Backpack Language Models
Viaarxiv icon

Lexinvariant Language Models

Add code
Bookmark button
Alert button
May 24, 2023
Qian Huang, Eric Zelikman, Sarah Li Chen, Yuhuai Wu, Gregory Valiant, Percy Liang

Figure 1 for Lexinvariant Language Models
Figure 2 for Lexinvariant Language Models
Figure 3 for Lexinvariant Language Models
Figure 4 for Lexinvariant Language Models
Viaarxiv icon

DoReMi: Optimizing Data Mixtures Speeds Up Language Model Pretraining

Add code
Bookmark button
Alert button
May 24, 2023
Sang Michael Xie, Hieu Pham, Xuanyi Dong, Nan Du, Hanxiao Liu, Yifeng Lu, Percy Liang, Quoc V. Le, Tengyu Ma, Adams Wei Yu

Figure 1 for DoReMi: Optimizing Data Mixtures Speeds Up Language Model Pretraining
Figure 2 for DoReMi: Optimizing Data Mixtures Speeds Up Language Model Pretraining
Figure 3 for DoReMi: Optimizing Data Mixtures Speeds Up Language Model Pretraining
Figure 4 for DoReMi: Optimizing Data Mixtures Speeds Up Language Model Pretraining
Viaarxiv icon

Sophia: A Scalable Stochastic Second-order Optimizer for Language Model Pre-training

Add code
Bookmark button
Alert button
May 23, 2023
Hong Liu, Zhiyuan Li, David Hall, Percy Liang, Tengyu Ma

Figure 1 for Sophia: A Scalable Stochastic Second-order Optimizer for Language Model Pre-training
Figure 2 for Sophia: A Scalable Stochastic Second-order Optimizer for Language Model Pre-training
Figure 3 for Sophia: A Scalable Stochastic Second-order Optimizer for Language Model Pre-training
Figure 4 for Sophia: A Scalable Stochastic Second-order Optimizer for Language Model Pre-training
Viaarxiv icon

AlpacaFarm: A Simulation Framework for Methods that Learn from Human Feedback

Add code
Bookmark button
Alert button
May 22, 2023
Yann Dubois, Xuechen Li, Rohan Taori, Tianyi Zhang, Ishaan Gulrajani, Jimmy Ba, Carlos Guestrin, Percy Liang, Tatsunori B. Hashimoto

Figure 1 for AlpacaFarm: A Simulation Framework for Methods that Learn from Human Feedback
Figure 2 for AlpacaFarm: A Simulation Framework for Methods that Learn from Human Feedback
Figure 3 for AlpacaFarm: A Simulation Framework for Methods that Learn from Human Feedback
Figure 4 for AlpacaFarm: A Simulation Framework for Methods that Learn from Human Feedback
Viaarxiv icon