Alert button
Picture for Guangxuan Xiao

Guangxuan Xiao

Alert button

BitDelta: Your Fine-Tune May Only Be Worth One Bit

Add code
Bookmark button
Alert button
Feb 28, 2024
James Liu, Guangxuan Xiao, Kai Li, Jason D. Lee, Song Han, Tri Dao, Tianle Cai

Viaarxiv icon

InfLLM: Unveiling the Intrinsic Capacity of LLMs for Understanding Extremely Long Sequences with Training-Free Memory

Add code
Bookmark button
Alert button
Feb 07, 2024
Chaojun Xiao, Pengle Zhang, Xu Han, Guangxuan Xiao, Yankai Lin, Zhengyan Zhang, Zhiyuan Liu, Song Han, Maosong Sun

Viaarxiv icon

Efficient Streaming Language Models with Attention Sinks

Add code
Bookmark button
Alert button
Sep 29, 2023
Guangxuan Xiao, Yuandong Tian, Beidi Chen, Song Han, Mike Lewis

Figure 1 for Efficient Streaming Language Models with Attention Sinks
Figure 2 for Efficient Streaming Language Models with Attention Sinks
Figure 3 for Efficient Streaming Language Models with Attention Sinks
Figure 4 for Efficient Streaming Language Models with Attention Sinks
Viaarxiv icon

FastComposer: Tuning-Free Multi-Subject Image Generation with Localized Attention

Add code
Bookmark button
Alert button
May 21, 2023
Guangxuan Xiao, Tianwei Yin, William T. Freeman, Frédo Durand, Song Han

Figure 1 for FastComposer: Tuning-Free Multi-Subject Image Generation with Localized Attention
Figure 2 for FastComposer: Tuning-Free Multi-Subject Image Generation with Localized Attention
Figure 3 for FastComposer: Tuning-Free Multi-Subject Image Generation with Localized Attention
Figure 4 for FastComposer: Tuning-Free Multi-Subject Image Generation with Localized Attention
Viaarxiv icon

Sparse and Local Networks for Hypergraph Reasoning

Add code
Bookmark button
Alert button
Mar 09, 2023
Guangxuan Xiao, Leslie Pack Kaelbling, Jiajun Wu, Jiayuan Mao

Figure 1 for Sparse and Local Networks for Hypergraph Reasoning
Figure 2 for Sparse and Local Networks for Hypergraph Reasoning
Figure 3 for Sparse and Local Networks for Hypergraph Reasoning
Figure 4 for Sparse and Local Networks for Hypergraph Reasoning
Viaarxiv icon

Offsite-Tuning: Transfer Learning without Full Model

Add code
Bookmark button
Alert button
Feb 09, 2023
Guangxuan Xiao, Ji Lin, Song Han

Figure 1 for Offsite-Tuning: Transfer Learning without Full Model
Figure 2 for Offsite-Tuning: Transfer Learning without Full Model
Figure 3 for Offsite-Tuning: Transfer Learning without Full Model
Figure 4 for Offsite-Tuning: Transfer Learning without Full Model
Viaarxiv icon

ReFresh: Reducing Memory Access from Exploiting Stable Historical Embeddings for Graph Neural Network Training

Add code
Bookmark button
Alert button
Jan 19, 2023
Kezhao Huang, Haitian Jiang, Minjie Wang, Guangxuan Xiao, David Wipf, Xiang Song, Quan Gan, Zengfeng Huang, Jidong Zhai, Zheng Zhang

Figure 1 for ReFresh: Reducing Memory Access from Exploiting Stable Historical Embeddings for Graph Neural Network Training
Figure 2 for ReFresh: Reducing Memory Access from Exploiting Stable Historical Embeddings for Graph Neural Network Training
Figure 3 for ReFresh: Reducing Memory Access from Exploiting Stable Historical Embeddings for Graph Neural Network Training
Figure 4 for ReFresh: Reducing Memory Access from Exploiting Stable Historical Embeddings for Graph Neural Network Training
Viaarxiv icon

SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models

Add code
Bookmark button
Alert button
Nov 28, 2022
Guangxuan Xiao, Ji Lin, Mickael Seznec, Julien Demouth, Song Han

Figure 1 for SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models
Figure 2 for SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models
Figure 3 for SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models
Figure 4 for SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models
Viaarxiv icon