Picture for Kanzhi Cheng

Kanzhi Cheng

Interactive Evolution: A Neural-Symbolic Self-Training Framework For Large Language Models

Add code
Jun 17, 2024
Figure 1 for Interactive Evolution: A Neural-Symbolic Self-Training Framework For Large Language Models
Figure 2 for Interactive Evolution: A Neural-Symbolic Self-Training Framework For Large Language Models
Figure 3 for Interactive Evolution: A Neural-Symbolic Self-Training Framework For Large Language Models
Figure 4 for Interactive Evolution: A Neural-Symbolic Self-Training Framework For Large Language Models
Viaarxiv icon

A Survey of Neural Code Intelligence: Paradigms, Advances and Beyond

Add code
Mar 21, 2024
Viaarxiv icon

SeeClick: Harnessing GUI Grounding for Advanced Visual GUI Agents

Add code
Jan 17, 2024
Viaarxiv icon

Food-500 Cap: A Fine-Grained Food Caption Benchmark for Evaluating Vision-Language Models

Add code
Aug 06, 2023
Figure 1 for Food-500 Cap: A Fine-Grained Food Caption Benchmark for Evaluating Vision-Language Models
Figure 2 for Food-500 Cap: A Fine-Grained Food Caption Benchmark for Evaluating Vision-Language Models
Figure 3 for Food-500 Cap: A Fine-Grained Food Caption Benchmark for Evaluating Vision-Language Models
Figure 4 for Food-500 Cap: A Fine-Grained Food Caption Benchmark for Evaluating Vision-Language Models
Viaarxiv icon

ADS-Cap: A Framework for Accurate and Diverse Stylized Captioning with Unpaired Stylistic Corpora

Add code
Aug 02, 2023
Viaarxiv icon

Beyond Generic: Enhancing Image Captioning with Real-World Knowledge using Vision-Language Pre-Training Model

Add code
Aug 02, 2023
Figure 1 for Beyond Generic: Enhancing Image Captioning with Real-World Knowledge using Vision-Language Pre-Training Model
Figure 2 for Beyond Generic: Enhancing Image Captioning with Real-World Knowledge using Vision-Language Pre-Training Model
Figure 3 for Beyond Generic: Enhancing Image Captioning with Real-World Knowledge using Vision-Language Pre-Training Model
Figure 4 for Beyond Generic: Enhancing Image Captioning with Real-World Knowledge using Vision-Language Pre-Training Model
Viaarxiv icon