Picture for Lewei Lu

Lewei Lu

EVA: Efficient Reinforcement Learning for End-to-End Video Agent

Add code
Mar 24, 2026
Viaarxiv icon

ACPO: Counteracting Likelihood Displacement in Vision-Language Alignment with Asymmetric Constraints

Add code
Mar 23, 2026
Viaarxiv icon

FailureMem: A Failure-Aware Multimodal Framework for Autonomous Software Repair

Add code
Mar 18, 2026
Viaarxiv icon

GA-Drive: Geometry-Appearance Decoupled Modeling for Free-viewpoint Driving Scene Generatio

Add code
Feb 24, 2026
Viaarxiv icon

ICA: Information-Aware Credit Assignment for Visually Grounded Long-Horizon Information-Seeking Agents

Add code
Feb 11, 2026
Viaarxiv icon

SenseNova-MARS: Empowering Multimodal Agentic Reasoning and Search via Reinforcement Learning

Add code
Dec 30, 2025
Viaarxiv icon

Towards Fine-Grained Recognition with Large Visual Language Models: Benchmark and Optimization Strategies

Add code
Dec 11, 2025
Figure 1 for Towards Fine-Grained Recognition with Large Visual Language Models: Benchmark and Optimization Strategies
Figure 2 for Towards Fine-Grained Recognition with Large Visual Language Models: Benchmark and Optimization Strategies
Figure 3 for Towards Fine-Grained Recognition with Large Visual Language Models: Benchmark and Optimization Strategies
Figure 4 for Towards Fine-Grained Recognition with Large Visual Language Models: Benchmark and Optimization Strategies
Viaarxiv icon

Scaling Spatial Intelligence with Multimodal Foundation Models

Add code
Nov 17, 2025
Figure 1 for Scaling Spatial Intelligence with Multimodal Foundation Models
Figure 2 for Scaling Spatial Intelligence with Multimodal Foundation Models
Figure 3 for Scaling Spatial Intelligence with Multimodal Foundation Models
Figure 4 for Scaling Spatial Intelligence with Multimodal Foundation Models
Viaarxiv icon

Enhancing the Outcome Reward-based RL Training of MLLMs with Self-Consistency Sampling

Add code
Nov 13, 2025
Figure 1 for Enhancing the Outcome Reward-based RL Training of MLLMs with Self-Consistency Sampling
Figure 2 for Enhancing the Outcome Reward-based RL Training of MLLMs with Self-Consistency Sampling
Figure 3 for Enhancing the Outcome Reward-based RL Training of MLLMs with Self-Consistency Sampling
Figure 4 for Enhancing the Outcome Reward-based RL Training of MLLMs with Self-Consistency Sampling
Viaarxiv icon

Spatial Preference Rewarding for MLLMs Spatial Understanding

Add code
Oct 16, 2025
Viaarxiv icon