Picture for Thomas H. Li

Thomas H. Li

Learning Semantic Facial Descriptors for Accurate Face Animation

Add code
Jan 29, 2025
Viaarxiv icon

EPContrast: Effective Point-level Contrastive Learning for Large-scale Point Cloud Understanding

Add code
Oct 22, 2024
Figure 1 for EPContrast: Effective Point-level Contrastive Learning for Large-scale Point Cloud Understanding
Figure 2 for EPContrast: Effective Point-level Contrastive Learning for Large-scale Point Cloud Understanding
Figure 3 for EPContrast: Effective Point-level Contrastive Learning for Large-scale Point Cloud Understanding
Figure 4 for EPContrast: Effective Point-level Contrastive Learning for Large-scale Point Cloud Understanding
Viaarxiv icon

StreamFlow: Streamlined Multi-Frame Optical Flow Estimation for Video Sequences

Add code
Nov 28, 2023
Figure 1 for StreamFlow: Streamlined Multi-Frame Optical Flow Estimation for Video Sequences
Figure 2 for StreamFlow: Streamlined Multi-Frame Optical Flow Estimation for Video Sequences
Figure 3 for StreamFlow: Streamlined Multi-Frame Optical Flow Estimation for Video Sequences
Figure 4 for StreamFlow: Streamlined Multi-Frame Optical Flow Estimation for Video Sequences
Viaarxiv icon

Mug-STAN: Adapting Image-Language Pretrained Models for General Video Understanding

Add code
Nov 25, 2023
Viaarxiv icon

Efficient Test-Time Adaptation for Super-Resolution with Second-Order Degradation and Reconstruction

Add code
Oct 29, 2023
Viaarxiv icon

FGPrompt: Fine-grained Goal Prompting for Image-goal Navigation

Add code
Oct 11, 2023
Viaarxiv icon

One For All: Video Conversation is Feasible Without Video Instruction Tuning

Add code
Sep 27, 2023
Figure 1 for One For All: Video Conversation is Feasible Without Video Instruction Tuning
Figure 2 for One For All: Video Conversation is Feasible Without Video Instruction Tuning
Figure 3 for One For All: Video Conversation is Feasible Without Video Instruction Tuning
Figure 4 for One For All: Video Conversation is Feasible Without Video Instruction Tuning
Viaarxiv icon

$A^2$Nav: Action-Aware Zero-Shot Robot Navigation by Exploiting Vision-and-Language Ability of Foundation Models

Add code
Aug 15, 2023
Figure 1 for $A^2$Nav: Action-Aware Zero-Shot Robot Navigation by Exploiting Vision-and-Language Ability of Foundation Models
Figure 2 for $A^2$Nav: Action-Aware Zero-Shot Robot Navigation by Exploiting Vision-and-Language Ability of Foundation Models
Figure 3 for $A^2$Nav: Action-Aware Zero-Shot Robot Navigation by Exploiting Vision-and-Language Ability of Foundation Models
Figure 4 for $A^2$Nav: Action-Aware Zero-Shot Robot Navigation by Exploiting Vision-and-Language Ability of Foundation Models
Viaarxiv icon

Learning Vision-and-Language Navigation from YouTube Videos

Add code
Jul 22, 2023
Viaarxiv icon

Hard Sample Matters a Lot in Zero-Shot Quantization

Add code
Mar 24, 2023
Figure 1 for Hard Sample Matters a Lot in Zero-Shot Quantization
Figure 2 for Hard Sample Matters a Lot in Zero-Shot Quantization
Figure 3 for Hard Sample Matters a Lot in Zero-Shot Quantization
Figure 4 for Hard Sample Matters a Lot in Zero-Shot Quantization
Viaarxiv icon