Alert button
Picture for Tatsunori Hashimoto

Tatsunori Hashimoto

Alert button

Benchmarking and Improving Generator-Validator Consistency of Language Models

Oct 03, 2023
Xiang Lisa Li, Vaishnavi Shrivastava, Siyan Li, Tatsunori Hashimoto, Percy Liang

Figure 1 for Benchmarking and Improving Generator-Validator Consistency of Language Models
Figure 2 for Benchmarking and Improving Generator-Validator Consistency of Language Models
Figure 3 for Benchmarking and Improving Generator-Validator Consistency of Language Models
Figure 4 for Benchmarking and Improving Generator-Validator Consistency of Language Models
Viaarxiv icon

Identifying the Risks of LM Agents with an LM-Emulated Sandbox

Sep 25, 2023
Yangjun Ruan, Honghua Dong, Andrew Wang, Silviu Pitis, Yongchao Zhou, Jimmy Ba, Yann Dubois, Chris J. Maddison, Tatsunori Hashimoto

Figure 1 for Identifying the Risks of LM Agents with an LM-Emulated Sandbox
Figure 2 for Identifying the Risks of LM Agents with an LM-Emulated Sandbox
Figure 3 for Identifying the Risks of LM Agents with an LM-Emulated Sandbox
Figure 4 for Identifying the Risks of LM Agents with an LM-Emulated Sandbox
Viaarxiv icon

Safety-Tuned LLaMAs: Lessons From Improving the Safety of Large Language Models that Follow Instructions

Sep 25, 2023
Federico Bianchi, Mirac Suzgun, Giuseppe Attanasio, Paul Röttger, Dan Jurafsky, Tatsunori Hashimoto, James Zou

Figure 1 for Safety-Tuned LLaMAs: Lessons From Improving the Safety of Large Language Models that Follow Instructions
Figure 2 for Safety-Tuned LLaMAs: Lessons From Improving the Safety of Large Language Models that Follow Instructions
Figure 3 for Safety-Tuned LLaMAs: Lessons From Improving the Safety of Large Language Models that Follow Instructions
Figure 4 for Safety-Tuned LLaMAs: Lessons From Improving the Safety of Large Language Models that Follow Instructions
Viaarxiv icon

Identifying and Mitigating the Security Risks of Generative AI

Aug 28, 2023
Clark Barrett, Brad Boyd, Ellie Burzstein, Nicholas Carlini, Brad Chen, Jihye Choi, Amrita Roy Chowdhury, Mihai Christodorescu, Anupam Datta, Soheil Feizi, Kathleen Fisher, Tatsunori Hashimoto, Dan Hendrycks, Somesh Jha, Daniel Kang, Florian Kerschbaum, Eric Mitchell, John Mitchell, Zulfikar Ramzan, Khawaja Shams, Dawn Song, Ankur Taly, Diyi Yang

Figure 1 for Identifying and Mitigating the Security Risks of Generative AI
Viaarxiv icon

Where's the Liability in Harmful AI Speech?

Aug 16, 2023
Peter Henderson, Tatsunori Hashimoto, Mark Lemley

Figure 1 for Where's the Liability in Harmful AI Speech?
Figure 2 for Where's the Liability in Harmful AI Speech?
Figure 3 for Where's the Liability in Harmful AI Speech?
Figure 4 for Where's the Liability in Harmful AI Speech?
Viaarxiv icon

Robust Distortion-free Watermarks for Language Models

Jul 28, 2023
Rohith Kuditipudi, John Thickstun, Tatsunori Hashimoto, Percy Liang

Figure 1 for Robust Distortion-free Watermarks for Language Models
Figure 2 for Robust Distortion-free Watermarks for Language Models
Figure 3 for Robust Distortion-free Watermarks for Language Models
Figure 4 for Robust Distortion-free Watermarks for Language Models
Viaarxiv icon

Whose Opinions Do Language Models Reflect?

Mar 30, 2023
Shibani Santurkar, Esin Durmus, Faisal Ladhak, Cinoo Lee, Percy Liang, Tatsunori Hashimoto

Figure 1 for Whose Opinions Do Language Models Reflect?
Figure 2 for Whose Opinions Do Language Models Reflect?
Figure 3 for Whose Opinions Do Language Models Reflect?
Figure 4 for Whose Opinions Do Language Models Reflect?
Viaarxiv icon

Foundation Models and Fair Use

Mar 28, 2023
Peter Henderson, Xuechen Li, Dan Jurafsky, Tatsunori Hashimoto, Mark A. Lemley, Percy Liang

Figure 1 for Foundation Models and Fair Use
Figure 2 for Foundation Models and Fair Use
Figure 3 for Foundation Models and Fair Use
Figure 4 for Foundation Models and Fair Use
Viaarxiv icon