Picture for Li Shen

Li Shen

Mix Data or Merge Models? Balancing the Helpfulness, Honesty, and Harmlessness of Large Language Model via Model Merging

Add code
Feb 13, 2025
Figure 1 for Mix Data or Merge Models? Balancing the Helpfulness, Honesty, and Harmlessness of Large Language Model via Model Merging
Figure 2 for Mix Data or Merge Models? Balancing the Helpfulness, Honesty, and Harmlessness of Large Language Model via Model Merging
Figure 3 for Mix Data or Merge Models? Balancing the Helpfulness, Honesty, and Harmlessness of Large Language Model via Model Merging
Figure 4 for Mix Data or Merge Models? Balancing the Helpfulness, Honesty, and Harmlessness of Large Language Model via Model Merging
Viaarxiv icon

HRP: High-Rank Preheating for Superior LoRA Initialization

Add code
Feb 11, 2025
Viaarxiv icon

Mask-Enhanced Autoregressive Prediction: Pay Less Attention to Learn More

Add code
Feb 11, 2025
Figure 1 for Mask-Enhanced Autoregressive Prediction: Pay Less Attention to Learn More
Figure 2 for Mask-Enhanced Autoregressive Prediction: Pay Less Attention to Learn More
Figure 3 for Mask-Enhanced Autoregressive Prediction: Pay Less Attention to Learn More
Figure 4 for Mask-Enhanced Autoregressive Prediction: Pay Less Attention to Learn More
Viaarxiv icon

Near-Optimal Online Learning for Multi-Agent Submodular Coordination: Tight Approximation and Communication Efficiency

Add code
Feb 07, 2025
Figure 1 for Near-Optimal Online Learning for Multi-Agent Submodular Coordination: Tight Approximation and Communication Efficiency
Figure 2 for Near-Optimal Online Learning for Multi-Agent Submodular Coordination: Tight Approximation and Communication Efficiency
Figure 3 for Near-Optimal Online Learning for Multi-Agent Submodular Coordination: Tight Approximation and Communication Efficiency
Figure 4 for Near-Optimal Online Learning for Multi-Agent Submodular Coordination: Tight Approximation and Communication Efficiency
Viaarxiv icon

Mediator: Memory-efficient LLM Merging with Less Parameter Conflicts and Uncertainty Based Routing

Add code
Feb 06, 2025
Viaarxiv icon

Leveraging Reasoning with Guidelines to Elicit and Utilize Knowledge for Enhancing Safety Alignment

Add code
Feb 06, 2025
Figure 1 for Leveraging Reasoning with Guidelines to Elicit and Utilize Knowledge for Enhancing Safety Alignment
Figure 2 for Leveraging Reasoning with Guidelines to Elicit and Utilize Knowledge for Enhancing Safety Alignment
Figure 3 for Leveraging Reasoning with Guidelines to Elicit and Utilize Knowledge for Enhancing Safety Alignment
Figure 4 for Leveraging Reasoning with Guidelines to Elicit and Utilize Knowledge for Enhancing Safety Alignment
Viaarxiv icon

TeZO: Empowering the Low-Rankness on the Temporal Dimension in the Zeroth-Order Optimization for Fine-tuning LLMs

Add code
Jan 31, 2025
Figure 1 for TeZO: Empowering the Low-Rankness on the Temporal Dimension in the Zeroth-Order Optimization for Fine-tuning LLMs
Figure 2 for TeZO: Empowering the Low-Rankness on the Temporal Dimension in the Zeroth-Order Optimization for Fine-tuning LLMs
Figure 3 for TeZO: Empowering the Low-Rankness on the Temporal Dimension in the Zeroth-Order Optimization for Fine-tuning LLMs
Figure 4 for TeZO: Empowering the Low-Rankness on the Temporal Dimension in the Zeroth-Order Optimization for Fine-tuning LLMs
Viaarxiv icon

Panacea: Mitigating Harmful Fine-tuning for Large Language Models via Post-fine-tuning Perturbation

Add code
Jan 30, 2025
Figure 1 for Panacea: Mitigating Harmful Fine-tuning for Large Language Models via Post-fine-tuning Perturbation
Figure 2 for Panacea: Mitigating Harmful Fine-tuning for Large Language Models via Post-fine-tuning Perturbation
Figure 3 for Panacea: Mitigating Harmful Fine-tuning for Large Language Models via Post-fine-tuning Perturbation
Figure 4 for Panacea: Mitigating Harmful Fine-tuning for Large Language Models via Post-fine-tuning Perturbation
Viaarxiv icon

Merging Models on the Fly Without Retraining: A Sequential Approach to Scalable Continual Model Merging

Add code
Jan 16, 2025
Figure 1 for Merging Models on the Fly Without Retraining: A Sequential Approach to Scalable Continual Model Merging
Figure 2 for Merging Models on the Fly Without Retraining: A Sequential Approach to Scalable Continual Model Merging
Figure 3 for Merging Models on the Fly Without Retraining: A Sequential Approach to Scalable Continual Model Merging
Figure 4 for Merging Models on the Fly Without Retraining: A Sequential Approach to Scalable Continual Model Merging
Viaarxiv icon

Modeling Multi-Task Model Merging as Adaptive Projective Gradient Descent

Add code
Jan 02, 2025
Figure 1 for Modeling Multi-Task Model Merging as Adaptive Projective Gradient Descent
Figure 2 for Modeling Multi-Task Model Merging as Adaptive Projective Gradient Descent
Figure 3 for Modeling Multi-Task Model Merging as Adaptive Projective Gradient Descent
Figure 4 for Modeling Multi-Task Model Merging as Adaptive Projective Gradient Descent
Viaarxiv icon