Picture for Robin Jia

Robin Jia

Hubble: a Model Suite to Advance the Study of LLM Memorization

Add code
Oct 22, 2025
Viaarxiv icon

Zebra-CoT: A Dataset for Interleaved Vision Language Reasoning

Add code
Jul 22, 2025
Viaarxiv icon

When Do LLMs Admit Their Mistakes? Understanding the Role of Model Belief in Retraction

Add code
May 22, 2025
Viaarxiv icon

Textual Steering Vectors Can Improve Visual Understanding in Multimodal Large Language Models

Add code
May 20, 2025
Figure 1 for Textual Steering Vectors Can Improve Visual Understanding in Multimodal Large Language Models
Figure 2 for Textual Steering Vectors Can Improve Visual Understanding in Multimodal Large Language Models
Figure 3 for Textual Steering Vectors Can Improve Visual Understanding in Multimodal Large Language Models
Figure 4 for Textual Steering Vectors Can Improve Visual Understanding in Multimodal Large Language Models
Viaarxiv icon

Teaching Models to Understand (but not Generate) High-risk Data

Add code
May 05, 2025
Viaarxiv icon

Cancer-Myth: Evaluating AI Chatbot on Patient Questions with False Presuppositions

Add code
Apr 15, 2025
Figure 1 for Cancer-Myth: Evaluating AI Chatbot on Patient Questions with False Presuppositions
Figure 2 for Cancer-Myth: Evaluating AI Chatbot on Patient Questions with False Presuppositions
Figure 3 for Cancer-Myth: Evaluating AI Chatbot on Patient Questions with False Presuppositions
Figure 4 for Cancer-Myth: Evaluating AI Chatbot on Patient Questions with False Presuppositions
Viaarxiv icon

Robust Data Watermarking in Language Models by Injecting Fictitious Knowledge

Add code
Mar 06, 2025
Figure 1 for Robust Data Watermarking in Language Models by Injecting Fictitious Knowledge
Figure 2 for Robust Data Watermarking in Language Models by Injecting Fictitious Knowledge
Figure 3 for Robust Data Watermarking in Language Models by Injecting Fictitious Knowledge
Figure 4 for Robust Data Watermarking in Language Models by Injecting Fictitious Knowledge
Viaarxiv icon

Promote, Suppress, Iterate: How Language Models Answer One-to-Many Factual Queries

Add code
Feb 27, 2025
Viaarxiv icon

Interrogating LLM design under a fair learning doctrine

Add code
Feb 22, 2025
Figure 1 for Interrogating LLM design under a fair learning doctrine
Figure 2 for Interrogating LLM design under a fair learning doctrine
Figure 3 for Interrogating LLM design under a fair learning doctrine
Figure 4 for Interrogating LLM design under a fair learning doctrine
Viaarxiv icon

Mechanistic Interpretability of Emotion Inference in Large Language Models

Add code
Feb 08, 2025
Viaarxiv icon