Alert button
Picture for Christopher Potts

Christopher Potts

Alert button

pyvene: A Library for Understanding and Improving PyTorch Models via Interventions

Mar 12, 2024
Zhengxuan Wu, Atticus Geiger, Aryaman Arora, Jing Huang, Zheng Wang, Noah D. Goodman, Christopher D. Manning, Christopher Potts

Viaarxiv icon

RAVEL: Evaluating Interpretability Methods on Disentangling Language Model Representations

Feb 27, 2024
Jing Huang, Zhengxuan Wu, Christopher Potts, Mor Geva, Atticus Geiger

Viaarxiv icon

CommVQA: Situating Visual Question Answering in Communicative Contexts

Feb 22, 2024
Nandita Shankar Naik, Christopher Potts, Elisa Kreiss

Viaarxiv icon

CausalGym: Benchmarking causal interpretability methods on linguistic tasks

Feb 19, 2024
Aryaman Arora, Dan Jurafsky, Christopher Potts

Viaarxiv icon

A Reply to Makelov et al. (2023)'s "Interpretability Illusion" Arguments

Jan 23, 2024
Zhengxuan Wu, Atticus Geiger, Jing Huang, Aryaman Arora, Thomas Icard, Christopher Potts, Noah D. Goodman

Viaarxiv icon

In-Context Learning for Extreme Multi-Label Classification

Jan 22, 2024
Karel D'Oosterlinck, Omar Khattab, François Remy, Thomas Demeester, Chris Develder, Christopher Potts

Viaarxiv icon

Mission: Impossible Language Models

Jan 12, 2024
Julie Kallini, Isabel Papadimitriou, Richard Futrell, Kyle Mahowald, Christopher Potts

Viaarxiv icon

I am a Strange Dataset: Metalinguistic Tests for Language Models

Jan 10, 2024
Tristan Thrush, Jared Moore, Miguel Monares, Christopher Potts, Douwe Kiela

Viaarxiv icon

Building Efficient and Effective OpenQA Systems for Low-Resource Languages

Jan 07, 2024
Emrah Budur, Rıza Özçelik, Dilara Soylu, Omar Khattab, Tunga Güngör, Christopher Potts

Viaarxiv icon

DSPy Assertions: Computational Constraints for Self-Refining Language Model Pipelines

Dec 20, 2023
Arnav Singhvi, Manish Shetty, Shangyin Tan, Christopher Potts, Koushik Sen, Matei Zaharia, Omar Khattab

Viaarxiv icon