Alert button
Picture for Taylor Sorensen

Taylor Sorensen

Alert button

A Roadmap to Pluralistic Alignment

Add code
Bookmark button
Alert button
Feb 07, 2024
Taylor Sorensen, Jared Moore, Jillian Fisher, Mitchell Gordon, Niloofar Mireshghallah, Christopher Michael Rytting, Andre Ye, Liwei Jiang, Ximing Lu, Nouha Dziri, Tim Althoff, Yejin Choi

Viaarxiv icon

NovaCOMET: Open Commonsense Foundation Models with Symbolic Knowledge Distillation

Add code
Bookmark button
Alert button
Dec 10, 2023
Peter West, Ronan Le Bras, Taylor Sorensen, Bill Yuchen Lin, Liwei Jiang, Ximing Lu, Khyathi Chandu, Jack Hessel, Ashutosh Baheti, Chandra Bhagavatula, Yejin Choi

Figure 1 for NovaCOMET: Open Commonsense Foundation Models with Symbolic Knowledge Distillation
Figure 2 for NovaCOMET: Open Commonsense Foundation Models with Symbolic Knowledge Distillation
Figure 3 for NovaCOMET: Open Commonsense Foundation Models with Symbolic Knowledge Distillation
Figure 4 for NovaCOMET: Open Commonsense Foundation Models with Symbolic Knowledge Distillation
Viaarxiv icon

Value Kaleidoscope: Engaging AI with Pluralistic Human Values, Rights, and Duties

Add code
Bookmark button
Alert button
Sep 02, 2023
Taylor Sorensen, Liwei Jiang, Jena Hwang, Sydney Levine, Valentina Pyatkin, Peter West, Nouha Dziri, Ximing Lu, Kavel Rao, Chandra Bhagavatula, Maarten Sap, John Tasioulas, Yejin Choi

Figure 1 for Value Kaleidoscope: Engaging AI with Pluralistic Human Values, Rights, and Duties
Figure 2 for Value Kaleidoscope: Engaging AI with Pluralistic Human Values, Rights, and Duties
Figure 3 for Value Kaleidoscope: Engaging AI with Pluralistic Human Values, Rights, and Duties
Figure 4 for Value Kaleidoscope: Engaging AI with Pluralistic Human Values, Rights, and Duties
Viaarxiv icon

Towards Coding Social Science Datasets with Language Models

Add code
Bookmark button
Alert button
Jun 03, 2023
Christopher Michael Rytting, Taylor Sorensen, Lisa Argyle, Ethan Busby, Nancy Fulda, Joshua Gubler, David Wingate

Figure 1 for Towards Coding Social Science Datasets with Language Models
Figure 2 for Towards Coding Social Science Datasets with Language Models
Figure 3 for Towards Coding Social Science Datasets with Language Models
Figure 4 for Towards Coding Social Science Datasets with Language Models
Viaarxiv icon

Impossible Distillation: from Low-Quality Model to High-Quality Dataset & Model for Summarization and Paraphrasing

Add code
Bookmark button
Alert button
May 26, 2023
Jaehun Jung, Peter West, Liwei Jiang, Faeze Brahman, Ximing Lu, Jillian Fisher, Taylor Sorensen, Yejin Choi

Figure 1 for Impossible Distillation: from Low-Quality Model to High-Quality Dataset & Model for Summarization and Paraphrasing
Figure 2 for Impossible Distillation: from Low-Quality Model to High-Quality Dataset & Model for Summarization and Paraphrasing
Figure 3 for Impossible Distillation: from Low-Quality Model to High-Quality Dataset & Model for Summarization and Paraphrasing
Figure 4 for Impossible Distillation: from Low-Quality Model to High-Quality Dataset & Model for Summarization and Paraphrasing
Viaarxiv icon

Prompt Compression and Contrastive Conditioning for Controllability and Toxicity Reduction in Language Models

Add code
Bookmark button
Alert button
Oct 06, 2022
David Wingate, Mohammad Shoeybi, Taylor Sorensen

Figure 1 for Prompt Compression and Contrastive Conditioning for Controllability and Toxicity Reduction in Language Models
Figure 2 for Prompt Compression and Contrastive Conditioning for Controllability and Toxicity Reduction in Language Models
Figure 3 for Prompt Compression and Contrastive Conditioning for Controllability and Toxicity Reduction in Language Models
Figure 4 for Prompt Compression and Contrastive Conditioning for Controllability and Toxicity Reduction in Language Models
Viaarxiv icon

An Information-theoretic Approach to Prompt Engineering Without Ground Truth Labels

Add code
Bookmark button
Alert button
Mar 21, 2022
Taylor Sorensen, Joshua Robinson, Christopher Michael Rytting, Alexander Glenn Shaw, Kyle Jeffrey Rogers, Alexia Pauline Delorey, Mahmoud Khalil, Nancy Fulda, David Wingate

Figure 1 for An Information-theoretic Approach to Prompt Engineering Without Ground Truth Labels
Figure 2 for An Information-theoretic Approach to Prompt Engineering Without Ground Truth Labels
Figure 3 for An Information-theoretic Approach to Prompt Engineering Without Ground Truth Labels
Figure 4 for An Information-theoretic Approach to Prompt Engineering Without Ground Truth Labels
Viaarxiv icon

NL-Augmenter: A Framework for Task-Sensitive Natural Language Augmentation

Add code
Bookmark button
Alert button
Dec 06, 2021
Kaustubh D. Dhole, Varun Gangal, Sebastian Gehrmann, Aadesh Gupta, Zhenhao Li, Saad Mahamood, Abinaya Mahendiran, Simon Mille, Ashish Srivastava, Samson Tan, Tongshuang Wu, Jascha Sohl-Dickstein, Jinho D. Choi, Eduard Hovy, Ondrej Dusek, Sebastian Ruder, Sajant Anand, Nagender Aneja, Rabin Banjade, Lisa Barthe, Hanna Behnke, Ian Berlot-Attwell, Connor Boyle, Caroline Brun, Marco Antonio Sobrevilla Cabezudo, Samuel Cahyawijaya, Emile Chapuis, Wanxiang Che, Mukund Choudhary, Christian Clauss, Pierre Colombo, Filip Cornell, Gautier Dagan, Mayukh Das, Tanay Dixit, Thomas Dopierre, Paul-Alexis Dray, Suchitra Dubey, Tatiana Ekeinhor, Marco Di Giovanni, Rishabh Gupta, Rishabh Gupta, Louanes Hamla, Sang Han, Fabrice Harel-Canada, Antoine Honore, Ishan Jindal, Przemyslaw K. Joniak, Denis Kleyko, Venelin Kovatchev, Kalpesh Krishna, Ashutosh Kumar, Stefan Langer, Seungjae Ryan Lee, Corey James Levinson, Hualou Liang, Kaizhao Liang, Zhexiong Liu, Andrey Lukyanenko, Vukosi Marivate, Gerard de Melo, Simon Meoni, Maxime Meyer, Afnan Mir, Nafise Sadat Moosavi, Niklas Muennighoff, Timothy Sum Hon Mun, Kenton Murray, Marcin Namysl, Maria Obedkova, Priti Oli, Nivranshu Pasricha, Jan Pfister, Richard Plant, Vinay Prabhu, Vasile Pais, Libo Qin, Shahab Raji, Pawan Kumar Rajpoot, Vikas Raunak, Roy Rinberg, Nicolas Roberts, Juan Diego Rodriguez, Claude Roux, Vasconcellos P. H. S., Ananya B. Sai, Robin M. Schmidt, Thomas Scialom, Tshephisho Sefara, Saqib N. Shamsi, Xudong Shen, Haoyue Shi, Yiwen Shi, Anna Shvets, Nick Siegel, Damien Sileo, Jamie Simon, Chandan Singh, Roman Sitelew, Priyank Soni, Taylor Sorensen, William Soto, Aman Srivastava, KV Aditya Srivatsa, Tony Sun, Mukund Varma T, A Tabassum, Fiona Anting Tan, Ryan Teehan, Mo Tiwari, Marie Tolkiehn, Athena Wang, Zijian Wang, Gloria Wang, Zijie J. Wang, Fuxuan Wei, Bryan Wilie, Genta Indra Winata, Xinyi Wu, Witold Wydmański, Tianbao Xie, Usama Yaseen, M. Yee, Jing Zhang, Yue Zhang

Figure 1 for NL-Augmenter: A Framework for Task-Sensitive Natural Language Augmentation
Figure 2 for NL-Augmenter: A Framework for Task-Sensitive Natural Language Augmentation
Figure 3 for NL-Augmenter: A Framework for Task-Sensitive Natural Language Augmentation
Figure 4 for NL-Augmenter: A Framework for Task-Sensitive Natural Language Augmentation
Viaarxiv icon