Picture for Zonghao Ying

Zonghao Ying

RoboSafe: Safeguarding Embodied Agents via Executable Safety Logic

Add code
Dec 24, 2025
Viaarxiv icon

Disentangling Fact from Sentiment: A Dynamic Conflict-Consensus Framework for Multimodal Fake News Detection

Add code
Dec 19, 2025
Viaarxiv icon

VEIL: Jailbreaking Text-to-Video Models via Visual Exploitation from Implicit Language

Add code
Nov 17, 2025
Viaarxiv icon

AGENTSAFE: Benchmarking the Safety of Embodied Agents on Hazardous Instructions

Add code
Jun 17, 2025
Viaarxiv icon

Pushing the Limits of Safety: A Technical Report on the ATLAS Challenge 2025

Add code
Jun 14, 2025
Viaarxiv icon

Manipulating Multimodal Agents via Cross-Modal Prompt Injection

Add code
Apr 22, 2025
Viaarxiv icon

Towards Understanding the Safety Boundaries of DeepSeek Models: Evaluation and Findings

Add code
Mar 19, 2025
Viaarxiv icon

CogMorph: Cognitive Morphing Attacks for Text-to-Image Models

Add code
Jan 21, 2025
Viaarxiv icon

Unveiling the Safety of GPT-4o: An Empirical Study using Jailbreak Attacks

Add code
Jun 10, 2024
Figure 1 for Unveiling the Safety of GPT-4o: An Empirical Study using Jailbreak Attacks
Figure 2 for Unveiling the Safety of GPT-4o: An Empirical Study using Jailbreak Attacks
Figure 3 for Unveiling the Safety of GPT-4o: An Empirical Study using Jailbreak Attacks
Figure 4 for Unveiling the Safety of GPT-4o: An Empirical Study using Jailbreak Attacks
Viaarxiv icon

Jailbreak Vision Language Models via Bi-Modal Adversarial Prompt

Add code
Jun 06, 2024
Viaarxiv icon