Picture for Sijia Liu

Sijia Liu

N3C Natural Language Processing

WAGLE: Strategic Weight Attribution for Effective and Modular Unlearning in Large Language Models

Add code
Oct 23, 2024
Viaarxiv icon

LLM Self-Correction with DeCRIM: Decompose, Critique, and Refine for Enhanced Following of Instructions with Multiple Constraints

Add code
Oct 09, 2024
Figure 1 for LLM Self-Correction with DeCRIM: Decompose, Critique, and Refine for Enhanced Following of Instructions with Multiple Constraints
Figure 2 for LLM Self-Correction with DeCRIM: Decompose, Critique, and Refine for Enhanced Following of Instructions with Multiple Constraints
Figure 3 for LLM Self-Correction with DeCRIM: Decompose, Critique, and Refine for Enhanced Following of Instructions with Multiple Constraints
Figure 4 for LLM Self-Correction with DeCRIM: Decompose, Critique, and Refine for Enhanced Following of Instructions with Multiple Constraints
Viaarxiv icon

Simplicity Prevails: Rethinking Negative Preference Optimization for LLM Unlearning

Add code
Oct 09, 2024
Figure 1 for Simplicity Prevails: Rethinking Negative Preference Optimization for LLM Unlearning
Figure 2 for Simplicity Prevails: Rethinking Negative Preference Optimization for LLM Unlearning
Figure 3 for Simplicity Prevails: Rethinking Negative Preference Optimization for LLM Unlearning
Figure 4 for Simplicity Prevails: Rethinking Negative Preference Optimization for LLM Unlearning
Viaarxiv icon

Pruning then Reweighting: Towards Data-Efficient Training of Diffusion Models

Add code
Sep 27, 2024
Viaarxiv icon

Transient Adversarial 3D Projection Attacks on Object Detection in Autonomous Driving

Add code
Sep 25, 2024
Viaarxiv icon

Adversarial Watermarking for Face Recognition

Add code
Sep 24, 2024
Viaarxiv icon

Learning on Transformers is Provable Low-Rank and Sparse: A One-layer Analysis

Add code
Jun 24, 2024
Figure 1 for Learning on Transformers is Provable Low-Rank and Sparse: A One-layer Analysis
Figure 2 for Learning on Transformers is Provable Low-Rank and Sparse: A One-layer Analysis
Figure 3 for Learning on Transformers is Provable Low-Rank and Sparse: A One-layer Analysis
Figure 4 for Learning on Transformers is Provable Low-Rank and Sparse: A One-layer Analysis
Viaarxiv icon

Reversing the Forget-Retain Objectives: An Efficient LLM Unlearning Framework from Logit Difference

Add code
Jun 12, 2024
Viaarxiv icon

Label Smoothing Improves Machine Unlearning

Add code
Jun 11, 2024
Figure 1 for Label Smoothing Improves Machine Unlearning
Figure 2 for Label Smoothing Improves Machine Unlearning
Figure 3 for Label Smoothing Improves Machine Unlearning
Figure 4 for Label Smoothing Improves Machine Unlearning
Viaarxiv icon

PSBD: Prediction Shift Uncertainty Unlocks Backdoor Detection

Add code
Jun 09, 2024
Viaarxiv icon