Picture for Alex Mei

Alex Mei

ASSERT: Automated Safety Scenario Red Teaming for Evaluating the Robustness of Large Language Models

Add code
Oct 14, 2023
Figure 1 for ASSERT: Automated Safety Scenario Red Teaming for Evaluating the Robustness of Large Language Models
Figure 2 for ASSERT: Automated Safety Scenario Red Teaming for Evaluating the Robustness of Large Language Models
Figure 3 for ASSERT: Automated Safety Scenario Red Teaming for Evaluating the Robustness of Large Language Models
Figure 4 for ASSERT: Automated Safety Scenario Red Teaming for Evaluating the Robustness of Large Language Models
Viaarxiv icon

Let's Think Frame by Frame: Evaluating Video Chain of Thought with Video Infilling and Prediction

Add code
May 23, 2023
Viaarxiv icon

Visual Chain of Thought: Bridging Logical Gaps with Multimodal Infillings

Add code
May 03, 2023
Figure 1 for Visual Chain of Thought: Bridging Logical Gaps with Multimodal Infillings
Figure 2 for Visual Chain of Thought: Bridging Logical Gaps with Multimodal Infillings
Figure 3 for Visual Chain of Thought: Bridging Logical Gaps with Multimodal Infillings
Figure 4 for Visual Chain of Thought: Bridging Logical Gaps with Multimodal Infillings
Viaarxiv icon

Users are the North Star for AI Transparency

Add code
Mar 09, 2023
Viaarxiv icon

Foveate, Attribute, and Rationalize: Towards Safe and Trustworthy AI

Add code
Dec 19, 2022
Viaarxiv icon

Mitigating Covertly Unsafe Text within Natural Language Systems

Add code
Oct 17, 2022
Figure 1 for Mitigating Covertly Unsafe Text within Natural Language Systems
Figure 2 for Mitigating Covertly Unsafe Text within Natural Language Systems
Figure 3 for Mitigating Covertly Unsafe Text within Natural Language Systems
Figure 4 for Mitigating Covertly Unsafe Text within Natural Language Systems
Viaarxiv icon