Picture for Zichang Liu

Zichang Liu

Wisdom of Committee: Distilling from Foundation Model to Specialized Application Model

Add code
Feb 27, 2024
Figure 1 for Wisdom of Committee: Distilling from Foundation Model to Specialized Application Model
Figure 2 for Wisdom of Committee: Distilling from Foundation Model to Specialized Application Model
Figure 3 for Wisdom of Committee: Distilling from Foundation Model to Specialized Application Model
Figure 4 for Wisdom of Committee: Distilling from Foundation Model to Specialized Application Model
Viaarxiv icon

Heterogeneous federated collaborative filtering using FAIR: Federated Averaging in Random Subspaces

Add code
Nov 03, 2023
Figure 1 for Heterogeneous federated collaborative filtering using FAIR: Federated Averaging in Random Subspaces
Figure 2 for Heterogeneous federated collaborative filtering using FAIR: Federated Averaging in Random Subspaces
Figure 3 for Heterogeneous federated collaborative filtering using FAIR: Federated Averaging in Random Subspaces
Figure 4 for Heterogeneous federated collaborative filtering using FAIR: Federated Averaging in Random Subspaces
Viaarxiv icon

Deja Vu: Contextual Sparsity for Efficient LLMs at Inference Time

Add code
Oct 26, 2023
Figure 1 for Deja Vu: Contextual Sparsity for Efficient LLMs at Inference Time
Figure 2 for Deja Vu: Contextual Sparsity for Efficient LLMs at Inference Time
Figure 3 for Deja Vu: Contextual Sparsity for Efficient LLMs at Inference Time
Figure 4 for Deja Vu: Contextual Sparsity for Efficient LLMs at Inference Time
Viaarxiv icon

Scissorhands: Exploiting the Persistence of Importance Hypothesis for LLM KV Cache Compression at Test Time

Add code
May 26, 2023
Figure 1 for Scissorhands: Exploiting the Persistence of Importance Hypothesis for LLM KV Cache Compression at Test Time
Figure 2 for Scissorhands: Exploiting the Persistence of Importance Hypothesis for LLM KV Cache Compression at Test Time
Figure 3 for Scissorhands: Exploiting the Persistence of Importance Hypothesis for LLM KV Cache Compression at Test Time
Figure 4 for Scissorhands: Exploiting the Persistence of Importance Hypothesis for LLM KV Cache Compression at Test Time
Viaarxiv icon

Learning Multimodal Data Augmentation in Feature Space

Add code
Dec 29, 2022
Figure 1 for Learning Multimodal Data Augmentation in Feature Space
Figure 2 for Learning Multimodal Data Augmentation in Feature Space
Figure 3 for Learning Multimodal Data Augmentation in Feature Space
Figure 4 for Learning Multimodal Data Augmentation in Feature Space
Viaarxiv icon

DONet: Learning Category-Level 6D Object Pose and Size Estimation from Depth Observation

Add code
Jun 27, 2021
Figure 1 for DONet: Learning Category-Level 6D Object Pose and Size Estimation from Depth Observation
Figure 2 for DONet: Learning Category-Level 6D Object Pose and Size Estimation from Depth Observation
Figure 3 for DONet: Learning Category-Level 6D Object Pose and Size Estimation from Depth Observation
Figure 4 for DONet: Learning Category-Level 6D Object Pose and Size Estimation from Depth Observation
Viaarxiv icon

Efficient Inference via Universal LSH Kernel

Add code
Jun 21, 2021
Figure 1 for Efficient Inference via Universal LSH Kernel
Figure 2 for Efficient Inference via Universal LSH Kernel
Figure 3 for Efficient Inference via Universal LSH Kernel
Figure 4 for Efficient Inference via Universal LSH Kernel
Viaarxiv icon

Neighbor Oblivious Learning (NObLe) for Device Localization and Tracking

Add code
Nov 23, 2020
Figure 1 for Neighbor Oblivious Learning (NObLe) for Device Localization and Tracking
Figure 2 for Neighbor Oblivious Learning (NObLe) for Device Localization and Tracking
Figure 3 for Neighbor Oblivious Learning (NObLe) for Device Localization and Tracking
Figure 4 for Neighbor Oblivious Learning (NObLe) for Device Localization and Tracking
Viaarxiv icon

Conditional Automated Channel Pruning for Deep Neural Networks

Add code
Sep 27, 2020
Figure 1 for Conditional Automated Channel Pruning for Deep Neural Networks
Figure 2 for Conditional Automated Channel Pruning for Deep Neural Networks
Viaarxiv icon

Climbing the WOL: Training for Cheaper Inference

Add code
Jul 03, 2020
Figure 1 for Climbing the WOL: Training for Cheaper Inference
Figure 2 for Climbing the WOL: Training for Cheaper Inference
Figure 3 for Climbing the WOL: Training for Cheaper Inference
Figure 4 for Climbing the WOL: Training for Cheaper Inference
Viaarxiv icon