Picture for Chaojun Xiao

Chaojun Xiao

Sparsing Law: Towards Large Language Models with Greater Activation Sparsity

Add code
Nov 04, 2024
Figure 1 for Sparsing Law: Towards Large Language Models with Greater Activation Sparsity
Figure 2 for Sparsing Law: Towards Large Language Models with Greater Activation Sparsity
Figure 3 for Sparsing Law: Towards Large Language Models with Greater Activation Sparsity
Figure 4 for Sparsing Law: Towards Large Language Models with Greater Activation Sparsity
Viaarxiv icon

Enhancing Legal Case Retrieval via Scaling High-quality Synthetic Query-Candidate Pairs

Add code
Oct 09, 2024
Figure 1 for Enhancing Legal Case Retrieval via Scaling High-quality Synthetic Query-Candidate Pairs
Figure 2 for Enhancing Legal Case Retrieval via Scaling High-quality Synthetic Query-Candidate Pairs
Figure 3 for Enhancing Legal Case Retrieval via Scaling High-quality Synthetic Query-Candidate Pairs
Figure 4 for Enhancing Legal Case Retrieval via Scaling High-quality Synthetic Query-Candidate Pairs
Viaarxiv icon

Exploring the Benefit of Activation Sparsity in Pre-training

Add code
Oct 04, 2024
Figure 1 for Exploring the Benefit of Activation Sparsity in Pre-training
Figure 2 for Exploring the Benefit of Activation Sparsity in Pre-training
Figure 3 for Exploring the Benefit of Activation Sparsity in Pre-training
Figure 4 for Exploring the Benefit of Activation Sparsity in Pre-training
Viaarxiv icon

Locret: Enhancing Eviction in Long-Context LLM Inference with Trained Retaining Heads

Add code
Oct 02, 2024
Figure 1 for Locret: Enhancing Eviction in Long-Context LLM Inference with Trained Retaining Heads
Figure 2 for Locret: Enhancing Eviction in Long-Context LLM Inference with Trained Retaining Heads
Figure 3 for Locret: Enhancing Eviction in Long-Context LLM Inference with Trained Retaining Heads
Figure 4 for Locret: Enhancing Eviction in Long-Context LLM Inference with Trained Retaining Heads
Viaarxiv icon

From MOOC to MAIC: Reshaping Online Teaching and Learning through LLM-driven Agents

Add code
Sep 05, 2024
Figure 1 for From MOOC to MAIC: Reshaping Online Teaching and Learning through LLM-driven Agents
Figure 2 for From MOOC to MAIC: Reshaping Online Teaching and Learning through LLM-driven Agents
Figure 3 for From MOOC to MAIC: Reshaping Online Teaching and Learning through LLM-driven Agents
Figure 4 for From MOOC to MAIC: Reshaping Online Teaching and Learning through LLM-driven Agents
Viaarxiv icon

Configurable Foundation Models: Building LLMs from a Modular Perspective

Add code
Sep 04, 2024
Figure 1 for Configurable Foundation Models: Building LLMs from a Modular Perspective
Figure 2 for Configurable Foundation Models: Building LLMs from a Modular Perspective
Figure 3 for Configurable Foundation Models: Building LLMs from a Modular Perspective
Figure 4 for Configurable Foundation Models: Building LLMs from a Modular Perspective
Viaarxiv icon

The Elephant in the Room: Rethinking the Usage of Pre-trained Language Model in Sequential Recommendation

Add code
Apr 12, 2024
Viaarxiv icon

Robust and Scalable Model Editing for Large Language Models

Add code
Mar 26, 2024
Figure 1 for Robust and Scalable Model Editing for Large Language Models
Figure 2 for Robust and Scalable Model Editing for Large Language Models
Figure 3 for Robust and Scalable Model Editing for Large Language Models
Figure 4 for Robust and Scalable Model Editing for Large Language Models
Viaarxiv icon

Ouroboros: Speculative Decoding with Large Model Enhanced Drafting

Add code
Feb 21, 2024
Viaarxiv icon

InfLLM: Unveiling the Intrinsic Capacity of LLMs for Understanding Extremely Long Sequences with Training-Free Memory

Add code
Feb 07, 2024
Figure 1 for InfLLM: Unveiling the Intrinsic Capacity of LLMs for Understanding Extremely Long Sequences with Training-Free Memory
Figure 2 for InfLLM: Unveiling the Intrinsic Capacity of LLMs for Understanding Extremely Long Sequences with Training-Free Memory
Figure 3 for InfLLM: Unveiling the Intrinsic Capacity of LLMs for Understanding Extremely Long Sequences with Training-Free Memory
Figure 4 for InfLLM: Unveiling the Intrinsic Capacity of LLMs for Understanding Extremely Long Sequences with Training-Free Memory
Viaarxiv icon