Picture for Zhe Chen

Zhe Chen

Pix2NPHM: Learning to Regress NPHM Reconstructions From a Single Image

Add code
Dec 19, 2025
Viaarxiv icon

Smile on the Face, Sadness in the Eyes: Bridging the Emotion Gap with a Multimodal Dataset of Eye and Facial Behaviors

Add code
Dec 18, 2025
Viaarxiv icon

MiroThinker: Pushing the Performance Boundaries of Open-Source Research Agents via Model, Context, and Interactive Scaling

Add code
Nov 18, 2025
Viaarxiv icon

Multi-Agent Deep Research: Training Multi-Agent Systems with M-GRPO

Add code
Nov 18, 2025
Viaarxiv icon

Collaborative Multi-Robot Non-Prehensile Manipulation via Flow-Matching Co-Generation

Add code
Nov 14, 2025
Viaarxiv icon

Gentle Manipulation Policy Learning via Demonstrations from VLM Planned Atomic Skills

Add code
Nov 08, 2025
Viaarxiv icon

WHU-STree: A Multi-modal Benchmark Dataset for Street Tree Inventory

Add code
Sep 16, 2025
Viaarxiv icon

InternVL3.5: Advancing Open-Source Multimodal Models in Versatility, Reasoning, and Efficiency

Add code
Aug 25, 2025
Figure 1 for InternVL3.5: Advancing Open-Source Multimodal Models in Versatility, Reasoning, and Efficiency
Figure 2 for InternVL3.5: Advancing Open-Source Multimodal Models in Versatility, Reasoning, and Efficiency
Figure 3 for InternVL3.5: Advancing Open-Source Multimodal Models in Versatility, Reasoning, and Efficiency
Figure 4 for InternVL3.5: Advancing Open-Source Multimodal Models in Versatility, Reasoning, and Efficiency
Viaarxiv icon

MedReseacher-R1: Expert-Level Medical Deep Researcher via A Knowledge-Informed Trajectory Synthesis Framework

Add code
Aug 20, 2025
Viaarxiv icon

REX-RAG: Reasoning Exploration with Policy Correction in Retrieval-Augmented Generation

Add code
Aug 12, 2025
Viaarxiv icon