Picture for Junwei Liu

Junwei Liu

PulseMind: A Multi-Modal Medical Model for Real-World Clinical Diagnosis

Add code
Jan 12, 2026
Viaarxiv icon

MedDialogRubrics: A Comprehensive Benchmark and Evaluation Framework for Multi-turn Medical Consultations in Large Language Models

Add code
Jan 07, 2026
Viaarxiv icon

Multi-Agent Deep Research: Training Multi-Agent Systems with M-GRPO

Add code
Nov 18, 2025
Figure 1 for Multi-Agent Deep Research: Training Multi-Agent Systems with M-GRPO
Figure 2 for Multi-Agent Deep Research: Training Multi-Agent Systems with M-GRPO
Figure 3 for Multi-Agent Deep Research: Training Multi-Agent Systems with M-GRPO
Figure 4 for Multi-Agent Deep Research: Training Multi-Agent Systems with M-GRPO
Viaarxiv icon

Can Agents Fix Agent Issues?

Add code
May 27, 2025
Viaarxiv icon

Effective Field Neural Network

Add code
Feb 24, 2025
Viaarxiv icon

Global Estimation of Building-Integrated Facade and Rooftop Photovoltaic Potential by Integrating 3D Building Footprint and Spatio-Temporal Datasets

Add code
Dec 02, 2024
Viaarxiv icon

Large Language Model-Based Agents for Software Engineering: A Survey

Add code
Sep 04, 2024
Figure 1 for Large Language Model-Based Agents for Software Engineering: A Survey
Figure 2 for Large Language Model-Based Agents for Software Engineering: A Survey
Figure 3 for Large Language Model-Based Agents for Software Engineering: A Survey
Figure 4 for Large Language Model-Based Agents for Software Engineering: A Survey
Viaarxiv icon

A Refer-and-Ground Multimodal Large Language Model for Biomedicine

Add code
Jun 26, 2024
Viaarxiv icon

ClassEval: A Manually-Crafted Benchmark for Evaluating LLMs on Class-level Code Generation

Add code
Aug 14, 2023
Figure 1 for ClassEval: A Manually-Crafted Benchmark for Evaluating LLMs on Class-level Code Generation
Figure 2 for ClassEval: A Manually-Crafted Benchmark for Evaluating LLMs on Class-level Code Generation
Figure 3 for ClassEval: A Manually-Crafted Benchmark for Evaluating LLMs on Class-level Code Generation
Figure 4 for ClassEval: A Manually-Crafted Benchmark for Evaluating LLMs on Class-level Code Generation
Viaarxiv icon

Evaluating Instruction-Tuned Large Language Models on Code Comprehension and Generation

Add code
Aug 02, 2023
Figure 1 for Evaluating Instruction-Tuned Large Language Models on Code Comprehension and Generation
Figure 2 for Evaluating Instruction-Tuned Large Language Models on Code Comprehension and Generation
Figure 3 for Evaluating Instruction-Tuned Large Language Models on Code Comprehension and Generation
Figure 4 for Evaluating Instruction-Tuned Large Language Models on Code Comprehension and Generation
Viaarxiv icon