Picture for Tongshuang Wu

Tongshuang Wu

The Ramon Llull's Thinking Machine for Automated Ideation

Add code
Aug 28, 2025
Viaarxiv icon

Checklists Are Better Than Reward Models For Aligning Language Models

Add code
Jul 24, 2025
Viaarxiv icon

What Prompts Don't Say: Understanding and Managing Underspecification in LLM Prompts

Add code
May 19, 2025
Viaarxiv icon

Orbit: A Framework for Designing and Evaluating Multi-objective Rankers

Add code
Nov 07, 2024
Figure 1 for Orbit: A Framework for Designing and Evaluating Multi-objective Rankers
Figure 2 for Orbit: A Framework for Designing and Evaluating Multi-objective Rankers
Figure 3 for Orbit: A Framework for Designing and Evaluating Multi-objective Rankers
Figure 4 for Orbit: A Framework for Designing and Evaluating Multi-objective Rankers
Viaarxiv icon

HiMemFormer: Hierarchical Memory-Aware Transformer for Multi-Agent Action Anticipation

Add code
Nov 03, 2024
Figure 1 for HiMemFormer: Hierarchical Memory-Aware Transformer for Multi-Agent Action Anticipation
Figure 2 for HiMemFormer: Hierarchical Memory-Aware Transformer for Multi-Agent Action Anticipation
Figure 3 for HiMemFormer: Hierarchical Memory-Aware Transformer for Multi-Agent Action Anticipation
Figure 4 for HiMemFormer: Hierarchical Memory-Aware Transformer for Multi-Agent Action Anticipation
Viaarxiv icon

What Is Wrong with My Model? Identifying Systematic Problems with Semantic Data Slicing

Add code
Sep 14, 2024
Figure 1 for What Is Wrong with My Model? Identifying Systematic Problems with Semantic Data Slicing
Figure 2 for What Is Wrong with My Model? Identifying Systematic Problems with Semantic Data Slicing
Figure 3 for What Is Wrong with My Model? Identifying Systematic Problems with Semantic Data Slicing
Figure 4 for What Is Wrong with My Model? Identifying Systematic Problems with Semantic Data Slicing
Viaarxiv icon

What You Say = What You Want? Teaching Humans to Articulate Requirements for LLMs

Add code
Sep 13, 2024
Figure 1 for What You Say = What You Want? Teaching Humans to Articulate Requirements for LLMs
Figure 2 for What You Say = What You Want? Teaching Humans to Articulate Requirements for LLMs
Figure 3 for What You Say = What You Want? Teaching Humans to Articulate Requirements for LLMs
Figure 4 for What You Say = What You Want? Teaching Humans to Articulate Requirements for LLMs
Viaarxiv icon

SELF-GUIDE: Better Task-Specific Instruction Following via Self-Synthetic Finetuning

Add code
Jul 16, 2024
Figure 1 for SELF-GUIDE: Better Task-Specific Instruction Following via Self-Synthetic Finetuning
Figure 2 for SELF-GUIDE: Better Task-Specific Instruction Following via Self-Synthetic Finetuning
Figure 3 for SELF-GUIDE: Better Task-Specific Instruction Following via Self-Synthetic Finetuning
Figure 4 for SELF-GUIDE: Better Task-Specific Instruction Following via Self-Synthetic Finetuning
Viaarxiv icon

Synthetic Multimodal Question Generation

Add code
Jul 02, 2024
Figure 1 for Synthetic Multimodal Question Generation
Figure 2 for Synthetic Multimodal Question Generation
Figure 3 for Synthetic Multimodal Question Generation
Figure 4 for Synthetic Multimodal Question Generation
Viaarxiv icon

WebCanvas: Benchmarking Web Agents in Online Environments

Add code
Jun 18, 2024
Viaarxiv icon