Picture for XiaoFeng Wang

XiaoFeng Wang

DP-MGTD: Privacy-Preserving Machine-Generated Text Detection via Adaptive Differentially Private Entity Sanitization

Add code
Jan 08, 2026
Viaarxiv icon

AudioTrust: Benchmarking the Multifaceted Trustworthiness of Audio Large Language Models

Add code
May 22, 2025
Figure 1 for AudioTrust: Benchmarking the Multifaceted Trustworthiness of Audio Large Language Models
Figure 2 for AudioTrust: Benchmarking the Multifaceted Trustworthiness of Audio Large Language Models
Figure 3 for AudioTrust: Benchmarking the Multifaceted Trustworthiness of Audio Large Language Models
Figure 4 for AudioTrust: Benchmarking the Multifaceted Trustworthiness of Audio Large Language Models
Viaarxiv icon

PromptGuard: Soft Prompt-Guided Unsafe Content Moderation for Text-to-Image Models

Add code
Jan 07, 2025
Figure 1 for PromptGuard: Soft Prompt-Guided Unsafe Content Moderation for Text-to-Image Models
Figure 2 for PromptGuard: Soft Prompt-Guided Unsafe Content Moderation for Text-to-Image Models
Figure 3 for PromptGuard: Soft Prompt-Guided Unsafe Content Moderation for Text-to-Image Models
Figure 4 for PromptGuard: Soft Prompt-Guided Unsafe Content Moderation for Text-to-Image Models
Viaarxiv icon

DPAdapter: Improving Differentially Private Deep Learning through Noise Tolerance Pre-training

Add code
Mar 05, 2024
Figure 1 for DPAdapter: Improving Differentially Private Deep Learning through Noise Tolerance Pre-training
Figure 2 for DPAdapter: Improving Differentially Private Deep Learning through Noise Tolerance Pre-training
Figure 3 for DPAdapter: Improving Differentially Private Deep Learning through Noise Tolerance Pre-training
Figure 4 for DPAdapter: Improving Differentially Private Deep Learning through Noise Tolerance Pre-training
Viaarxiv icon

Malla: Demystifying Real-world Large Language Model Integrated Malicious Services

Add code
Jan 06, 2024
Viaarxiv icon

The Janus Interface: How Fine-Tuning in Large Language Models Amplifies the Privacy Risks

Add code
Oct 24, 2023
Figure 1 for The Janus Interface: How Fine-Tuning in Large Language Models Amplifies the Privacy Risks
Figure 2 for The Janus Interface: How Fine-Tuning in Large Language Models Amplifies the Privacy Risks
Figure 3 for The Janus Interface: How Fine-Tuning in Large Language Models Amplifies the Privacy Risks
Figure 4 for The Janus Interface: How Fine-Tuning in Large Language Models Amplifies the Privacy Risks
Viaarxiv icon

MAWSEO: Adversarial Wiki Search Poisoning for Illicit Online Promotion

Add code
Apr 22, 2023
Viaarxiv icon

Selective Amnesia: On Efficient, High-Fidelity and Blind Suppression of Backdoor Effects in Trojaned Machine Learning Models

Add code
Dec 09, 2022
Viaarxiv icon

Understanding Impacts of Task Similarity on Backdoor Attack and Detection

Add code
Oct 12, 2022
Figure 1 for Understanding Impacts of Task Similarity on Backdoor Attack and Detection
Figure 2 for Understanding Impacts of Task Similarity on Backdoor Attack and Detection
Figure 3 for Understanding Impacts of Task Similarity on Backdoor Attack and Detection
Figure 4 for Understanding Impacts of Task Similarity on Backdoor Attack and Detection
Viaarxiv icon

Towards Dark Jargon Interpretation in Underground Forums

Add code
Nov 05, 2020
Figure 1 for Towards Dark Jargon Interpretation in Underground Forums
Figure 2 for Towards Dark Jargon Interpretation in Underground Forums
Viaarxiv icon