Picture for William Yang Wang

William Yang Wang

Losing Visual Needles in Image Haystacks: Vision Language Models are Easily Distracted in Short and Long Contexts

Add code
Jun 24, 2024
Figure 1 for Losing Visual Needles in Image Haystacks: Vision Language Models are Easily Distracted in Short and Long Contexts
Figure 2 for Losing Visual Needles in Image Haystacks: Vision Language Models are Easily Distracted in Short and Long Contexts
Figure 3 for Losing Visual Needles in Image Haystacks: Vision Language Models are Easily Distracted in Short and Long Contexts
Figure 4 for Losing Visual Needles in Image Haystacks: Vision Language Models are Easily Distracted in Short and Long Contexts
Viaarxiv icon

DistiLRR: Transferring Code Repair for Low-Resource Programming Languages

Add code
Jun 21, 2024
Figure 1 for DistiLRR: Transferring Code Repair for Low-Resource Programming Languages
Figure 2 for DistiLRR: Transferring Code Repair for Low-Resource Programming Languages
Figure 3 for DistiLRR: Transferring Code Repair for Low-Resource Programming Languages
Figure 4 for DistiLRR: Transferring Code Repair for Low-Resource Programming Languages
Viaarxiv icon

Global Human-guided Counterfactual Explanations for Molecular Properties via Reinforcement Learning

Add code
Jun 19, 2024
Viaarxiv icon

BPO: Supercharging Online Preference Learning by Adhering to the Proximity of Behavior LLM

Add code
Jun 18, 2024
Figure 1 for BPO: Supercharging Online Preference Learning by Adhering to the Proximity of Behavior LLM
Figure 2 for BPO: Supercharging Online Preference Learning by Adhering to the Proximity of Behavior LLM
Figure 3 for BPO: Supercharging Online Preference Learning by Adhering to the Proximity of Behavior LLM
Figure 4 for BPO: Supercharging Online Preference Learning by Adhering to the Proximity of Behavior LLM
Viaarxiv icon

WildVision: Evaluating Vision-Language Models in the Wild with Human Preferences

Add code
Jun 16, 2024
Viaarxiv icon

TC-Bench: Benchmarking Temporal Compositionality in Text-to-Video and Image-to-Video Generation

Add code
Jun 12, 2024
Figure 1 for TC-Bench: Benchmarking Temporal Compositionality in Text-to-Video and Image-to-Video Generation
Figure 2 for TC-Bench: Benchmarking Temporal Compositionality in Text-to-Video and Image-to-Video Generation
Figure 3 for TC-Bench: Benchmarking Temporal Compositionality in Text-to-Video and Image-to-Video Generation
Figure 4 for TC-Bench: Benchmarking Temporal Compositionality in Text-to-Video and Image-to-Video Generation
Viaarxiv icon

MMWorld: Towards Multi-discipline Multi-faceted World Model Evaluation in Videos

Add code
Jun 12, 2024
Figure 1 for MMWorld: Towards Multi-discipline Multi-faceted World Model Evaluation in Videos
Figure 2 for MMWorld: Towards Multi-discipline Multi-faceted World Model Evaluation in Videos
Figure 3 for MMWorld: Towards Multi-discipline Multi-faceted World Model Evaluation in Videos
Figure 4 for MMWorld: Towards Multi-discipline Multi-faceted World Model Evaluation in Videos
Viaarxiv icon

Commonsense-T2I Challenge: Can Text-to-Image Generation Models Understand Commonsense?

Add code
Jun 11, 2024
Figure 1 for Commonsense-T2I Challenge: Can Text-to-Image Generation Models Understand Commonsense?
Figure 2 for Commonsense-T2I Challenge: Can Text-to-Image Generation Models Understand Commonsense?
Figure 3 for Commonsense-T2I Challenge: Can Text-to-Image Generation Models Understand Commonsense?
Figure 4 for Commonsense-T2I Challenge: Can Text-to-Image Generation Models Understand Commonsense?
Viaarxiv icon

Unveiling the Impact of Coding Data Instruction Fine-Tuning on Large Language Models Reasoning

Add code
May 30, 2024
Figure 1 for Unveiling the Impact of Coding Data Instruction Fine-Tuning on Large Language Models Reasoning
Figure 2 for Unveiling the Impact of Coding Data Instruction Fine-Tuning on Large Language Models Reasoning
Figure 3 for Unveiling the Impact of Coding Data Instruction Fine-Tuning on Large Language Models Reasoning
Figure 4 for Unveiling the Impact of Coding Data Instruction Fine-Tuning on Large Language Models Reasoning
Viaarxiv icon

T2V-Turbo: Breaking the Quality Bottleneck of Video Consistency Model with Mixed Reward Feedback

Add code
May 29, 2024
Figure 1 for T2V-Turbo: Breaking the Quality Bottleneck of Video Consistency Model with Mixed Reward Feedback
Figure 2 for T2V-Turbo: Breaking the Quality Bottleneck of Video Consistency Model with Mixed Reward Feedback
Figure 3 for T2V-Turbo: Breaking the Quality Bottleneck of Video Consistency Model with Mixed Reward Feedback
Figure 4 for T2V-Turbo: Breaking the Quality Bottleneck of Video Consistency Model with Mixed Reward Feedback
Viaarxiv icon