Alert button
Picture for Niloofar Mireshghallah

Niloofar Mireshghallah

Alert button

Alpaca against Vicuna: Using LLMs to Uncover Memorization of LLMs

Add code
Bookmark button
Alert button
Mar 05, 2024
Aly M. Kassem, Omar Mahmoud, Niloofar Mireshghallah, Hyunwoo Kim, Yulia Tsvetkov, Yejin Choi, Sherif Saad, Santu Rana

Figure 1 for Alpaca against Vicuna: Using LLMs to Uncover Memorization of LLMs
Figure 2 for Alpaca against Vicuna: Using LLMs to Uncover Memorization of LLMs
Figure 3 for Alpaca against Vicuna: Using LLMs to Uncover Memorization of LLMs
Figure 4 for Alpaca against Vicuna: Using LLMs to Uncover Memorization of LLMs
Viaarxiv icon

Do Membership Inference Attacks Work on Large Language Models?

Add code
Bookmark button
Alert button
Feb 12, 2024
Michael Duan, Anshuman Suri, Niloofar Mireshghallah, Sewon Min, Weijia Shi, Luke Zettlemoyer, Yulia Tsvetkov, Yejin Choi, David Evans, Hannaneh Hajishirzi

Viaarxiv icon

A Roadmap to Pluralistic Alignment

Add code
Bookmark button
Alert button
Feb 07, 2024
Taylor Sorensen, Jared Moore, Jillian Fisher, Mitchell Gordon, Niloofar Mireshghallah, Christopher Michael Rytting, Andre Ye, Liwei Jiang, Ximing Lu, Nouha Dziri, Tim Althoff, Yejin Choi

Viaarxiv icon

A Block Metropolis-Hastings Sampler for Controllable Energy-based Text Generation

Add code
Bookmark button
Alert button
Dec 07, 2023
Jarad Forristal, Niloofar Mireshghallah, Greg Durrett, Taylor Berg-Kirkpatrick

Viaarxiv icon

Can LLMs Keep a Secret? Testing Privacy Implications of Language Models via Contextual Integrity Theory

Add code
Bookmark button
Alert button
Oct 27, 2023
Niloofar Mireshghallah, Hyunwoo Kim, Xuhui Zhou, Yulia Tsvetkov, Maarten Sap, Reza Shokri, Yejin Choi

Figure 1 for Can LLMs Keep a Secret? Testing Privacy Implications of Language Models via Contextual Integrity Theory
Figure 2 for Can LLMs Keep a Secret? Testing Privacy Implications of Language Models via Contextual Integrity Theory
Figure 3 for Can LLMs Keep a Secret? Testing Privacy Implications of Language Models via Contextual Integrity Theory
Figure 4 for Can LLMs Keep a Secret? Testing Privacy Implications of Language Models via Contextual Integrity Theory
Viaarxiv icon

Misusing Tools in Large Language Models With Visual Adversarial Examples

Add code
Bookmark button
Alert button
Oct 04, 2023
Xiaohan Fu, Zihan Wang, Shuheng Li, Rajesh K. Gupta, Niloofar Mireshghallah, Taylor Berg-Kirkpatrick, Earlence Fernandes

Figure 1 for Misusing Tools in Large Language Models With Visual Adversarial Examples
Figure 2 for Misusing Tools in Large Language Models With Visual Adversarial Examples
Figure 3 for Misusing Tools in Large Language Models With Visual Adversarial Examples
Figure 4 for Misusing Tools in Large Language Models With Visual Adversarial Examples
Viaarxiv icon