Picture for Jesse Thomason

Jesse Thomason

University of Southern California

Language Models can Infer Action Semantics for Classical Planners from Environment Feedback

Add code
Jun 04, 2024
Viaarxiv icon

TwoStep: Multi-agent Task Planning using Classical Planners and Large Language Models

Add code
Mar 25, 2024
Figure 1 for TwoStep: Multi-agent Task Planning using Classical Planners and Large Language Models
Figure 2 for TwoStep: Multi-agent Task Planning using Classical Planners and Large Language Models
Figure 3 for TwoStep: Multi-agent Task Planning using Classical Planners and Large Language Models
Figure 4 for TwoStep: Multi-agent Task Planning using Classical Planners and Large Language Models
Viaarxiv icon

ViSaRL: Visual Reinforcement Learning Guided by Human Saliency

Add code
Mar 16, 2024
Figure 1 for ViSaRL: Visual Reinforcement Learning Guided by Human Saliency
Figure 2 for ViSaRL: Visual Reinforcement Learning Guided by Human Saliency
Figure 3 for ViSaRL: Visual Reinforcement Learning Guided by Human Saliency
Figure 4 for ViSaRL: Visual Reinforcement Learning Guided by Human Saliency
Viaarxiv icon

Selective "Selective Prediction": Reducing Unnecessary Abstention in Vision-Language Reasoning

Add code
Feb 23, 2024
Figure 1 for Selective "Selective Prediction": Reducing Unnecessary Abstention in Vision-Language Reasoning
Figure 2 for Selective "Selective Prediction": Reducing Unnecessary Abstention in Vision-Language Reasoning
Figure 3 for Selective "Selective Prediction": Reducing Unnecessary Abstention in Vision-Language Reasoning
Figure 4 for Selective "Selective Prediction": Reducing Unnecessary Abstention in Vision-Language Reasoning
Viaarxiv icon

WinoViz: Probing Visual Properties of Objects Under Different States

Add code
Feb 21, 2024
Figure 1 for WinoViz: Probing Visual Properties of Objects Under Different States
Figure 2 for WinoViz: Probing Visual Properties of Objects Under Different States
Figure 3 for WinoViz: Probing Visual Properties of Objects Under Different States
Figure 4 for WinoViz: Probing Visual Properties of Objects Under Different States
Viaarxiv icon

THE COLOSSEUM: A Benchmark for Evaluating Generalization for Robotic Manipulation

Add code
Feb 13, 2024
Viaarxiv icon

Does VLN Pretraining Work with Nonsensical or Irrelevant Instructions?

Add code
Dec 02, 2023
Figure 1 for Does VLN Pretraining Work with Nonsensical or Irrelevant Instructions?
Figure 2 for Does VLN Pretraining Work with Nonsensical or Irrelevant Instructions?
Figure 3 for Does VLN Pretraining Work with Nonsensical or Irrelevant Instructions?
Figure 4 for Does VLN Pretraining Work with Nonsensical or Irrelevant Instructions?
Viaarxiv icon

Efficient End-to-End Visual Document Understanding with Rationale Distillation

Add code
Nov 16, 2023
Figure 1 for Efficient End-to-End Visual Document Understanding with Rationale Distillation
Figure 2 for Efficient End-to-End Visual Document Understanding with Rationale Distillation
Figure 3 for Efficient End-to-End Visual Document Understanding with Rationale Distillation
Figure 4 for Efficient End-to-End Visual Document Understanding with Rationale Distillation
Viaarxiv icon

Do Localization Methods Actually Localize Memorized Data in LLMs?

Add code
Nov 15, 2023
Figure 1 for Do Localization Methods Actually Localize Memorized Data in LLMs?
Figure 2 for Do Localization Methods Actually Localize Memorized Data in LLMs?
Figure 3 for Do Localization Methods Actually Localize Memorized Data in LLMs?
Figure 4 for Do Localization Methods Actually Localize Memorized Data in LLMs?
Viaarxiv icon

Comparative Multi-View Language Grounding

Add code
Nov 14, 2023
Figure 1 for Comparative Multi-View Language Grounding
Figure 2 for Comparative Multi-View Language Grounding
Figure 3 for Comparative Multi-View Language Grounding
Figure 4 for Comparative Multi-View Language Grounding
Viaarxiv icon