Picture for Xun Zhao

Xun Zhao

JailbreakHunter: A Visual Analytics Approach for Jailbreak Prompts Discovery from Large-Scale Human-LLM Conversational Datasets

Add code
Jul 03, 2024
Figure 1 for JailbreakHunter: A Visual Analytics Approach for Jailbreak Prompts Discovery from Large-Scale Human-LLM Conversational Datasets
Figure 2 for JailbreakHunter: A Visual Analytics Approach for Jailbreak Prompts Discovery from Large-Scale Human-LLM Conversational Datasets
Figure 3 for JailbreakHunter: A Visual Analytics Approach for Jailbreak Prompts Discovery from Large-Scale Human-LLM Conversational Datasets
Figure 4 for JailbreakHunter: A Visual Analytics Approach for Jailbreak Prompts Discovery from Large-Scale Human-LLM Conversational Datasets
Viaarxiv icon

Unveiling the Misuse Potential of Base Large Language Models via In-Context Learning

Add code
Apr 16, 2024
Figure 1 for Unveiling the Misuse Potential of Base Large Language Models via In-Context Learning
Figure 2 for Unveiling the Misuse Potential of Base Large Language Models via In-Context Learning
Figure 3 for Unveiling the Misuse Potential of Base Large Language Models via In-Context Learning
Figure 4 for Unveiling the Misuse Potential of Base Large Language Models via In-Context Learning
Viaarxiv icon

Navigating the OverKill in Large Language Models

Add code
Jan 31, 2024
Viaarxiv icon

XIMAGENET-12: An Explainable AI Benchmark Dataset for Model Robustness Evaluation

Add code
Oct 12, 2023
Figure 1 for XIMAGENET-12: An Explainable AI Benchmark Dataset for Model Robustness Evaluation
Figure 2 for XIMAGENET-12: An Explainable AI Benchmark Dataset for Model Robustness Evaluation
Figure 3 for XIMAGENET-12: An Explainable AI Benchmark Dataset for Model Robustness Evaluation
Figure 4 for XIMAGENET-12: An Explainable AI Benchmark Dataset for Model Robustness Evaluation
Viaarxiv icon

Shadow Alignment: The Ease of Subverting Safely-Aligned Language Models

Add code
Oct 04, 2023
Figure 1 for Shadow Alignment: The Ease of Subverting Safely-Aligned Language Models
Figure 2 for Shadow Alignment: The Ease of Subverting Safely-Aligned Language Models
Figure 3 for Shadow Alignment: The Ease of Subverting Safely-Aligned Language Models
Figure 4 for Shadow Alignment: The Ease of Subverting Safely-Aligned Language Models
Viaarxiv icon

Not All Models Are Equal: Predicting Model Transferability in a Self-challenging Fisher Space

Add code
Jul 19, 2022
Figure 1 for Not All Models Are Equal: Predicting Model Transferability in a Self-challenging Fisher Space
Figure 2 for Not All Models Are Equal: Predicting Model Transferability in a Self-challenging Fisher Space
Figure 3 for Not All Models Are Equal: Predicting Model Transferability in a Self-challenging Fisher Space
Figure 4 for Not All Models Are Equal: Predicting Model Transferability in a Self-challenging Fisher Space
Viaarxiv icon

Temporally Efficient Vision Transformer for Video Instance Segmentation

Add code
Apr 18, 2022
Figure 1 for Temporally Efficient Vision Transformer for Video Instance Segmentation
Figure 2 for Temporally Efficient Vision Transformer for Video Instance Segmentation
Figure 3 for Temporally Efficient Vision Transformer for Video Instance Segmentation
Figure 4 for Temporally Efficient Vision Transformer for Video Instance Segmentation
Viaarxiv icon

Active Learning for Open-set Annotation

Add code
Jan 18, 2022
Figure 1 for Active Learning for Open-set Annotation
Figure 2 for Active Learning for Open-set Annotation
Figure 3 for Active Learning for Open-set Annotation
Figure 4 for Active Learning for Open-set Annotation
Viaarxiv icon

Towards Vivid and Diverse Image Colorization with Generative Color Prior

Add code
Aug 19, 2021
Figure 1 for Towards Vivid and Diverse Image Colorization with Generative Color Prior
Figure 2 for Towards Vivid and Diverse Image Colorization with Generative Color Prior
Figure 3 for Towards Vivid and Diverse Image Colorization with Generative Color Prior
Figure 4 for Towards Vivid and Diverse Image Colorization with Generative Color Prior
Viaarxiv icon