Picture for Ari Holtzman

Ari Holtzman

Nearest Neighbor Speculative Decoding for LLM Generation and Attribution

Add code
May 29, 2024
Figure 1 for Nearest Neighbor Speculative Decoding for LLM Generation and Attribution
Figure 2 for Nearest Neighbor Speculative Decoding for LLM Generation and Attribution
Figure 3 for Nearest Neighbor Speculative Decoding for LLM Generation and Attribution
Figure 4 for Nearest Neighbor Speculative Decoding for LLM Generation and Attribution
Viaarxiv icon

CacheGen: Fast Context Loading for Language Model Applications

Add code
Oct 11, 2023
Figure 1 for CacheGen: Fast Context Loading for Language Model Applications
Figure 2 for CacheGen: Fast Context Loading for Language Model Applications
Figure 3 for CacheGen: Fast Context Loading for Language Model Applications
Figure 4 for CacheGen: Fast Context Loading for Language Model Applications
Viaarxiv icon

How FaR Are Large Language Models From Agents with Theory-of-Mind?

Add code
Oct 04, 2023
Figure 1 for How FaR Are Large Language Models From Agents with Theory-of-Mind?
Figure 2 for How FaR Are Large Language Models From Agents with Theory-of-Mind?
Figure 3 for How FaR Are Large Language Models From Agents with Theory-of-Mind?
Figure 4 for How FaR Are Large Language Models From Agents with Theory-of-Mind?
Viaarxiv icon

Generative Models as a Complex Systems Science: How can we make sense of large language model behavior?

Add code
Jul 31, 2023
Viaarxiv icon

QLoRA: Efficient Finetuning of Quantized LLMs

Add code
May 23, 2023
Figure 1 for QLoRA: Efficient Finetuning of Quantized LLMs
Figure 2 for QLoRA: Efficient Finetuning of Quantized LLMs
Figure 3 for QLoRA: Efficient Finetuning of Quantized LLMs
Figure 4 for QLoRA: Efficient Finetuning of Quantized LLMs
Viaarxiv icon

Toward Human Readable Prompt Tuning: Kubrick's The Shining is a good movie, and a good prompt too?

Add code
Dec 20, 2022
Figure 1 for Toward Human Readable Prompt Tuning: Kubrick's The Shining is a good movie, and a good prompt too?
Figure 2 for Toward Human Readable Prompt Tuning: Kubrick's The Shining is a good movie, and a good prompt too?
Figure 3 for Toward Human Readable Prompt Tuning: Kubrick's The Shining is a good movie, and a good prompt too?
Figure 4 for Toward Human Readable Prompt Tuning: Kubrick's The Shining is a good movie, and a good prompt too?
Viaarxiv icon

Contrastive Decoding: Open-ended Text Generation as Optimization

Add code
Oct 27, 2022
Viaarxiv icon

What Do NLP Researchers Believe? Results of the NLP Community Metasurvey

Add code
Aug 26, 2022
Figure 1 for What Do NLP Researchers Believe? Results of the NLP Community Metasurvey
Figure 2 for What Do NLP Researchers Believe? Results of the NLP Community Metasurvey
Figure 3 for What Do NLP Researchers Believe? Results of the NLP Community Metasurvey
Figure 4 for What Do NLP Researchers Believe? Results of the NLP Community Metasurvey
Viaarxiv icon

Rethinking the Role of Demonstrations: What Makes In-Context Learning Work?

Add code
Feb 25, 2022
Figure 1 for Rethinking the Role of Demonstrations: What Makes In-Context Learning Work?
Figure 2 for Rethinking the Role of Demonstrations: What Makes In-Context Learning Work?
Figure 3 for Rethinking the Role of Demonstrations: What Makes In-Context Learning Work?
Figure 4 for Rethinking the Role of Demonstrations: What Makes In-Context Learning Work?
Viaarxiv icon

DEMix Layers: Disentangling Domains for Modular Language Modeling

Add code
Aug 20, 2021
Figure 1 for DEMix Layers: Disentangling Domains for Modular Language Modeling
Figure 2 for DEMix Layers: Disentangling Domains for Modular Language Modeling
Figure 3 for DEMix Layers: Disentangling Domains for Modular Language Modeling
Figure 4 for DEMix Layers: Disentangling Domains for Modular Language Modeling
Viaarxiv icon