Picture for Junnan Li

Junnan Li

Accelerating Video Diffusion Models via Distribution Matching

Add code
Dec 08, 2024
Figure 1 for Accelerating Video Diffusion Models via Distribution Matching
Figure 2 for Accelerating Video Diffusion Models via Distribution Matching
Figure 3 for Accelerating Video Diffusion Models via Distribution Matching
Figure 4 for Accelerating Video Diffusion Models via Distribution Matching
Viaarxiv icon

VideoAutoArena: An Automated Arena for Evaluating Large Multimodal Models in Video Analysis through User Simulation

Add code
Nov 20, 2024
Figure 1 for VideoAutoArena: An Automated Arena for Evaluating Large Multimodal Models in Video Analysis through User Simulation
Figure 2 for VideoAutoArena: An Automated Arena for Evaluating Large Multimodal Models in Video Analysis through User Simulation
Figure 3 for VideoAutoArena: An Automated Arena for Evaluating Large Multimodal Models in Video Analysis through User Simulation
Figure 4 for VideoAutoArena: An Automated Arena for Evaluating Large Multimodal Models in Video Analysis through User Simulation
Viaarxiv icon

Aria: An Open Multimodal Native Mixture-of-Experts Model

Add code
Oct 08, 2024
Figure 1 for Aria: An Open Multimodal Native Mixture-of-Experts Model
Figure 2 for Aria: An Open Multimodal Native Mixture-of-Experts Model
Figure 3 for Aria: An Open Multimodal Native Mixture-of-Experts Model
Figure 4 for Aria: An Open Multimodal Native Mixture-of-Experts Model
Viaarxiv icon

Optimizing Interaction Space: Enlarging the Capture Volume for Multiple Portable Motion Capture Devices

Add code
Aug 30, 2024
Viaarxiv icon

Identification and validation of the dynamic model of a tendon-driven anthropomorphic finger

Add code
Aug 23, 2024
Figure 1 for Identification and validation of the dynamic model of a tendon-driven anthropomorphic finger
Figure 2 for Identification and validation of the dynamic model of a tendon-driven anthropomorphic finger
Figure 3 for Identification and validation of the dynamic model of a tendon-driven anthropomorphic finger
Viaarxiv icon

Object Augmentation Algorithm: Computing virtual object motion and object induced interaction wrench from optical markers

Add code
Aug 14, 2024
Figure 1 for Object Augmentation Algorithm: Computing virtual object motion and object induced interaction wrench from optical markers
Figure 2 for Object Augmentation Algorithm: Computing virtual object motion and object induced interaction wrench from optical markers
Figure 3 for Object Augmentation Algorithm: Computing virtual object motion and object induced interaction wrench from optical markers
Figure 4 for Object Augmentation Algorithm: Computing virtual object motion and object induced interaction wrench from optical markers
Viaarxiv icon

LongVideoBench: A Benchmark for Long-context Interleaved Video-Language Understanding

Add code
Jul 22, 2024
Figure 1 for LongVideoBench: A Benchmark for Long-context Interleaved Video-Language Understanding
Figure 2 for LongVideoBench: A Benchmark for Long-context Interleaved Video-Language Understanding
Figure 3 for LongVideoBench: A Benchmark for Long-context Interleaved Video-Language Understanding
Figure 4 for LongVideoBench: A Benchmark for Long-context Interleaved Video-Language Understanding
Viaarxiv icon

What Are We Measuring When We Evaluate Large Vision-Language Models? An Analysis of Latent Factors and Biases

Add code
Apr 03, 2024
Viaarxiv icon

X-InstructBLIP: A Framework for aligning X-Modal instruction-aware representations to LLMs and Emergent Cross-modal Reasoning

Add code
Nov 30, 2023
Figure 1 for X-InstructBLIP: A Framework for aligning X-Modal instruction-aware representations to LLMs and Emergent Cross-modal Reasoning
Figure 2 for X-InstructBLIP: A Framework for aligning X-Modal instruction-aware representations to LLMs and Emergent Cross-modal Reasoning
Figure 3 for X-InstructBLIP: A Framework for aligning X-Modal instruction-aware representations to LLMs and Emergent Cross-modal Reasoning
Figure 4 for X-InstructBLIP: A Framework for aligning X-Modal instruction-aware representations to LLMs and Emergent Cross-modal Reasoning
Viaarxiv icon

CodeTF: One-stop Transformer Library for State-of-the-art Code LLM

Add code
May 31, 2023
Figure 1 for CodeTF: One-stop Transformer Library for State-of-the-art Code LLM
Figure 2 for CodeTF: One-stop Transformer Library for State-of-the-art Code LLM
Figure 3 for CodeTF: One-stop Transformer Library for State-of-the-art Code LLM
Figure 4 for CodeTF: One-stop Transformer Library for State-of-the-art Code LLM
Viaarxiv icon