Picture for Linlin Wang

Linlin Wang

Unified Defense for Large Language Models against Jailbreak and Fine-Tuning Attacks in Education

Add code
Nov 18, 2025
Figure 1 for Unified Defense for Large Language Models against Jailbreak and Fine-Tuning Attacks in Education
Figure 2 for Unified Defense for Large Language Models against Jailbreak and Fine-Tuning Attacks in Education
Figure 3 for Unified Defense for Large Language Models against Jailbreak and Fine-Tuning Attacks in Education
Figure 4 for Unified Defense for Large Language Models against Jailbreak and Fine-Tuning Attacks in Education
Viaarxiv icon

Iterative pseudo-labeling based adaptive copy-paste supervision for semi-supervised tumor segmentation

Add code
Aug 06, 2025
Viaarxiv icon

Fiber Signal Denoising Algorithm using Hybrid Deep Learning Networks

Add code
Jun 18, 2025
Viaarxiv icon

Bias Amplification in RAG: Poisoning Knowledge Retrieval to Steer LLMs

Add code
Jun 13, 2025
Figure 1 for Bias Amplification in RAG: Poisoning Knowledge Retrieval to Steer LLMs
Figure 2 for Bias Amplification in RAG: Poisoning Knowledge Retrieval to Steer LLMs
Figure 3 for Bias Amplification in RAG: Poisoning Knowledge Retrieval to Steer LLMs
Figure 4 for Bias Amplification in RAG: Poisoning Knowledge Retrieval to Steer LLMs
Viaarxiv icon

Hierarchical Safety Realignment: Lightweight Restoration of Safety in Pruned Large Vision-Language Models

Add code
May 22, 2025
Viaarxiv icon

AutoMedEval: Harnessing Language Models for Automatic Medical Capability Evaluation

Add code
May 17, 2025
Viaarxiv icon

Safe and Reliable Diffusion Models via Subspace Projection

Add code
Mar 21, 2025
Figure 1 for Safe and Reliable Diffusion Models via Subspace Projection
Figure 2 for Safe and Reliable Diffusion Models via Subspace Projection
Figure 3 for Safe and Reliable Diffusion Models via Subspace Projection
Figure 4 for Safe and Reliable Diffusion Models via Subspace Projection
Viaarxiv icon

Latent-space adversarial training with post-aware calibration for defending large language models against jailbreak attacks

Add code
Jan 18, 2025
Figure 1 for Latent-space adversarial training with post-aware calibration for defending large language models against jailbreak attacks
Figure 2 for Latent-space adversarial training with post-aware calibration for defending large language models against jailbreak attacks
Figure 3 for Latent-space adversarial training with post-aware calibration for defending large language models against jailbreak attacks
Figure 4 for Latent-space adversarial training with post-aware calibration for defending large language models against jailbreak attacks
Viaarxiv icon

Hierarchical Divide-and-Conquer for Fine-Grained Alignment in LLM-Based Medical Evaluation

Add code
Jan 12, 2025
Viaarxiv icon

NLSR: Neuron-Level Safety Realignment of Large Language Models Against Harmful Fine-Tuning

Add code
Dec 17, 2024
Viaarxiv icon