Picture for Zihao Lin

Zihao Lin

R2I-Bench: Benchmarking Reasoning-Driven Text-to-Image Generation

Add code
May 29, 2025
Viaarxiv icon

Localizing Knowledge in Diffusion Transformers

Add code
May 24, 2025
Viaarxiv icon

A Survey on Mechanistic Interpretability for Multi-Modal Foundation Models

Add code
Feb 22, 2025
Figure 1 for A Survey on Mechanistic Interpretability for Multi-Modal Foundation Models
Figure 2 for A Survey on Mechanistic Interpretability for Multi-Modal Foundation Models
Figure 3 for A Survey on Mechanistic Interpretability for Multi-Modal Foundation Models
Figure 4 for A Survey on Mechanistic Interpretability for Multi-Modal Foundation Models
Viaarxiv icon

Knowing When to Stop: Dynamic Context Cutoff for Large Language Models

Add code
Feb 03, 2025
Figure 1 for Knowing When to Stop: Dynamic Context Cutoff for Large Language Models
Figure 2 for Knowing When to Stop: Dynamic Context Cutoff for Large Language Models
Figure 3 for Knowing When to Stop: Dynamic Context Cutoff for Large Language Models
Figure 4 for Knowing When to Stop: Dynamic Context Cutoff for Large Language Models
Viaarxiv icon

Persona-SQ: A Personalized Suggested Question Generation Framework For Real-world Documents

Add code
Dec 17, 2024
Figure 1 for Persona-SQ: A Personalized Suggested Question Generation Framework For Real-world Documents
Figure 2 for Persona-SQ: A Personalized Suggested Question Generation Framework For Real-world Documents
Figure 3 for Persona-SQ: A Personalized Suggested Question Generation Framework For Real-world Documents
Figure 4 for Persona-SQ: A Personalized Suggested Question Generation Framework For Real-world Documents
Viaarxiv icon

Rethinking the Uncertainty: A Critical Review and Analysis in the Era of Large Language Models

Add code
Oct 26, 2024
Figure 1 for Rethinking the Uncertainty: A Critical Review and Analysis in the Era of Large Language Models
Figure 2 for Rethinking the Uncertainty: A Critical Review and Analysis in the Era of Large Language Models
Figure 3 for Rethinking the Uncertainty: A Critical Review and Analysis in the Era of Large Language Models
Viaarxiv icon

MMedAgent: Learning to Use Medical Tools with Multi-modal Agent

Add code
Jul 02, 2024
Figure 1 for MMedAgent: Learning to Use Medical Tools with Multi-modal Agent
Figure 2 for MMedAgent: Learning to Use Medical Tools with Multi-modal Agent
Figure 3 for MMedAgent: Learning to Use Medical Tools with Multi-modal Agent
Figure 4 for MMedAgent: Learning to Use Medical Tools with Multi-modal Agent
Viaarxiv icon

ToolBeHonest: A Multi-level Hallucination Diagnostic Benchmark for Tool-Augmented Large Language Models

Add code
Jun 28, 2024
Figure 1 for ToolBeHonest: A Multi-level Hallucination Diagnostic Benchmark for Tool-Augmented Large Language Models
Figure 2 for ToolBeHonest: A Multi-level Hallucination Diagnostic Benchmark for Tool-Augmented Large Language Models
Figure 3 for ToolBeHonest: A Multi-level Hallucination Diagnostic Benchmark for Tool-Augmented Large Language Models
Figure 4 for ToolBeHonest: A Multi-level Hallucination Diagnostic Benchmark for Tool-Augmented Large Language Models
Viaarxiv icon

Holistic Evaluation for Interleaved Text-and-Image Generation

Add code
Jun 20, 2024
Viaarxiv icon

InternalInspector $I^2$: Robust Confidence Estimation in LLMs through Internal States

Add code
Jun 17, 2024
Figure 1 for InternalInspector $I^2$: Robust Confidence Estimation in LLMs through Internal States
Figure 2 for InternalInspector $I^2$: Robust Confidence Estimation in LLMs through Internal States
Figure 3 for InternalInspector $I^2$: Robust Confidence Estimation in LLMs through Internal States
Figure 4 for InternalInspector $I^2$: Robust Confidence Estimation in LLMs through Internal States
Viaarxiv icon