Picture for Ari Holtzman

Ari Holtzman

Benchmarks as Microscopes: A Call for Model Metrology

Add code
Jul 22, 2024
Viaarxiv icon

MUSE: Machine Unlearning Six-Way Evaluation for Language Models

Add code
Jul 08, 2024
Figure 1 for MUSE: Machine Unlearning Six-Way Evaluation for Language Models
Figure 2 for MUSE: Machine Unlearning Six-Way Evaluation for Language Models
Figure 3 for MUSE: Machine Unlearning Six-Way Evaluation for Language Models
Figure 4 for MUSE: Machine Unlearning Six-Way Evaluation for Language Models
Viaarxiv icon

Predicting vs. Acting: A Trade-off Between World Modeling & Agent Modeling

Add code
Jul 02, 2024
Figure 1 for Predicting vs. Acting: A Trade-off Between World Modeling & Agent Modeling
Figure 2 for Predicting vs. Acting: A Trade-off Between World Modeling & Agent Modeling
Figure 3 for Predicting vs. Acting: A Trade-off Between World Modeling & Agent Modeling
Figure 4 for Predicting vs. Acting: A Trade-off Between World Modeling & Agent Modeling
Viaarxiv icon

Nearest Neighbor Speculative Decoding for LLM Generation and Attribution

Add code
May 29, 2024
Figure 1 for Nearest Neighbor Speculative Decoding for LLM Generation and Attribution
Figure 2 for Nearest Neighbor Speculative Decoding for LLM Generation and Attribution
Figure 3 for Nearest Neighbor Speculative Decoding for LLM Generation and Attribution
Figure 4 for Nearest Neighbor Speculative Decoding for LLM Generation and Attribution
Viaarxiv icon

Superposed Decoding: Multiple Generations from a Single Autoregressive Inference Pass

Add code
May 29, 2024
Figure 1 for Superposed Decoding: Multiple Generations from a Single Autoregressive Inference Pass
Figure 2 for Superposed Decoding: Multiple Generations from a Single Autoregressive Inference Pass
Figure 3 for Superposed Decoding: Multiple Generations from a Single Autoregressive Inference Pass
Figure 4 for Superposed Decoding: Multiple Generations from a Single Autoregressive Inference Pass
Viaarxiv icon

CacheGen: Fast Context Loading for Language Model Applications

Add code
Oct 11, 2023
Figure 1 for CacheGen: Fast Context Loading for Language Model Applications
Figure 2 for CacheGen: Fast Context Loading for Language Model Applications
Figure 3 for CacheGen: Fast Context Loading for Language Model Applications
Figure 4 for CacheGen: Fast Context Loading for Language Model Applications
Viaarxiv icon

How FaR Are Large Language Models From Agents with Theory-of-Mind?

Add code
Oct 04, 2023
Figure 1 for How FaR Are Large Language Models From Agents with Theory-of-Mind?
Figure 2 for How FaR Are Large Language Models From Agents with Theory-of-Mind?
Figure 3 for How FaR Are Large Language Models From Agents with Theory-of-Mind?
Figure 4 for How FaR Are Large Language Models From Agents with Theory-of-Mind?
Viaarxiv icon

Generative Models as a Complex Systems Science: How can we make sense of large language model behavior?

Add code
Jul 31, 2023
Viaarxiv icon

QLoRA: Efficient Finetuning of Quantized LLMs

Add code
May 23, 2023
Figure 1 for QLoRA: Efficient Finetuning of Quantized LLMs
Figure 2 for QLoRA: Efficient Finetuning of Quantized LLMs
Figure 3 for QLoRA: Efficient Finetuning of Quantized LLMs
Figure 4 for QLoRA: Efficient Finetuning of Quantized LLMs
Viaarxiv icon

Toward Human Readable Prompt Tuning: Kubrick's The Shining is a good movie, and a good prompt too?

Add code
Dec 20, 2022
Figure 1 for Toward Human Readable Prompt Tuning: Kubrick's The Shining is a good movie, and a good prompt too?
Figure 2 for Toward Human Readable Prompt Tuning: Kubrick's The Shining is a good movie, and a good prompt too?
Figure 3 for Toward Human Readable Prompt Tuning: Kubrick's The Shining is a good movie, and a good prompt too?
Figure 4 for Toward Human Readable Prompt Tuning: Kubrick's The Shining is a good movie, and a good prompt too?
Viaarxiv icon