Picture for Xinyi Yang

Xinyi Yang

VLM2Vec: Training Vision-Language Models for Massive Multimodal Embedding Tasks

Add code
Oct 07, 2024
Figure 1 for VLM2Vec: Training Vision-Language Models for Massive Multimodal Embedding Tasks
Figure 2 for VLM2Vec: Training Vision-Language Models for Massive Multimodal Embedding Tasks
Figure 3 for VLM2Vec: Training Vision-Language Models for Massive Multimodal Embedding Tasks
Figure 4 for VLM2Vec: Training Vision-Language Models for Massive Multimodal Embedding Tasks
Viaarxiv icon

ReGenesis: LLMs can Grow into Reasoning Generalists via Self-Improvement

Add code
Oct 03, 2024
Viaarxiv icon

Multi-UAV Pursuit-Evasion with Online Planning in Unknown Environments by Deep Reinforcement Learning

Add code
Sep 25, 2024
Viaarxiv icon

Rotatable Block-Controlled RIS: Bridging the Performance Gap to Element-Controlled Systems

Add code
Aug 22, 2024
Viaarxiv icon

CityLight: A Universal Model Towards Real-world City-scale Traffic Signal Control Coordination

Add code
Jun 04, 2024
Viaarxiv icon

Prefix Text as a Yarn: Eliciting Non-English Alignment in Foundation Language Model

Add code
Apr 25, 2024
Viaarxiv icon

FOFO: A Benchmark to Evaluate LLMs' Format-Following Capability

Add code
Feb 28, 2024
Viaarxiv icon

TaskFlex Solver for Multi-Agent Pursuit via Automatic Curriculum Learning

Add code
Dec 19, 2023
Viaarxiv icon

MASP: Scalable GNN-based Planning for Multi-Agent Navigation

Add code
Dec 05, 2023
Viaarxiv icon

Active Neural Topological Mapping for Multi-Agent Exploration

Add code
Nov 01, 2023
Viaarxiv icon