Picture for Cheng Tang

Cheng Tang

MARS2 2025 Challenge on Multimodal Reasoning: Datasets, Methods, Results, Discussion, and Outlook

Add code
Sep 17, 2025
Figure 1 for MARS2 2025 Challenge on Multimodal Reasoning: Datasets, Methods, Results, Discussion, and Outlook
Figure 2 for MARS2 2025 Challenge on Multimodal Reasoning: Datasets, Methods, Results, Discussion, and Outlook
Figure 3 for MARS2 2025 Challenge on Multimodal Reasoning: Datasets, Methods, Results, Discussion, and Outlook
Figure 4 for MARS2 2025 Challenge on Multimodal Reasoning: Datasets, Methods, Results, Discussion, and Outlook
Viaarxiv icon

A Survey of Scientific Large Language Models: From Data Foundations to Agent Frontiers

Add code
Aug 28, 2025
Viaarxiv icon

RetinaLogos: Fine-Grained Synthesis of High-Resolution Retinal Images Through Captions

Add code
May 19, 2025
Viaarxiv icon

LECTOR: Summarizing E-book Reading Content for Personalized Student Support

Add code
May 12, 2025
Viaarxiv icon

Single-Agent vs. Multi-Agent LLM Strategies for Automated Student Reflection Assessment

Add code
Apr 08, 2025
Viaarxiv icon

Attention Mamba: Time Series Modeling with Adaptive Pooling Acceleration and Receptive Field Enhancements

Add code
Apr 02, 2025
Viaarxiv icon

GMAI-VL-R1: Harnessing Reinforcement Learning for Multimodal Medical Reasoning

Add code
Apr 02, 2025
Viaarxiv icon

An explainable transformer circuit for compositional generalization

Add code
Feb 19, 2025
Figure 1 for An explainable transformer circuit for compositional generalization
Figure 2 for An explainable transformer circuit for compositional generalization
Figure 3 for An explainable transformer circuit for compositional generalization
Figure 4 for An explainable transformer circuit for compositional generalization
Viaarxiv icon

LLaVA-SLT: Visual Language Tuning for Sign Language Translation

Add code
Dec 21, 2024
Figure 1 for LLaVA-SLT: Visual Language Tuning for Sign Language Translation
Figure 2 for LLaVA-SLT: Visual Language Tuning for Sign Language Translation
Figure 3 for LLaVA-SLT: Visual Language Tuning for Sign Language Translation
Figure 4 for LLaVA-SLT: Visual Language Tuning for Sign Language Translation
Viaarxiv icon

Knowledge Distillation in RNN-Attention Models for Early Prediction of Student Performance

Add code
Dec 19, 2024
Figure 1 for Knowledge Distillation in RNN-Attention Models for Early Prediction of Student Performance
Figure 2 for Knowledge Distillation in RNN-Attention Models for Early Prediction of Student Performance
Figure 3 for Knowledge Distillation in RNN-Attention Models for Early Prediction of Student Performance
Figure 4 for Knowledge Distillation in RNN-Attention Models for Early Prediction of Student Performance
Viaarxiv icon