Picture for Lin Wu

Lin Wu

In-context Prompt Learning for Test-time Vision Recognition with Frozen Vision-language Model

Add code
Mar 10, 2024
Figure 1 for In-context Prompt Learning for Test-time Vision Recognition with Frozen Vision-language Model
Figure 2 for In-context Prompt Learning for Test-time Vision Recognition with Frozen Vision-language Model
Figure 3 for In-context Prompt Learning for Test-time Vision Recognition with Frozen Vision-language Model
Figure 4 for In-context Prompt Learning for Test-time Vision Recognition with Frozen Vision-language Model
Viaarxiv icon

A Novel Approach to Industrial Defect Generation through Blended Latent Diffusion Model with Online Adaptation

Add code
Feb 29, 2024
Viaarxiv icon

DART: Depth-Enhanced Accurate and Real-Time Background Matting

Add code
Feb 24, 2024
Viaarxiv icon

Semantic-Aware Adversarial Training for Reliable Deep Hashing Retrieval

Add code
Oct 23, 2023
Viaarxiv icon

DSformer: A Double Sampling Transformer for Multivariate Time Series Long-term Prediction

Add code
Aug 07, 2023
Figure 1 for DSformer: A Double Sampling Transformer for Multivariate Time Series Long-term Prediction
Figure 2 for DSformer: A Double Sampling Transformer for Multivariate Time Series Long-term Prediction
Figure 3 for DSformer: A Double Sampling Transformer for Multivariate Time Series Long-term Prediction
Figure 4 for DSformer: A Double Sampling Transformer for Multivariate Time Series Long-term Prediction
Viaarxiv icon

Research on Multilingual News Clustering Based on Cross-Language Word Embeddings

Add code
May 30, 2023
Figure 1 for Research on Multilingual News Clustering Based on Cross-Language Word Embeddings
Figure 2 for Research on Multilingual News Clustering Based on Cross-Language Word Embeddings
Figure 3 for Research on Multilingual News Clustering Based on Cross-Language Word Embeddings
Figure 4 for Research on Multilingual News Clustering Based on Cross-Language Word Embeddings
Viaarxiv icon

LipFormer: Learning to Lipread Unseen Speakers based on Visual-Landmark Transformers

Add code
Feb 04, 2023
Figure 1 for LipFormer: Learning to Lipread Unseen Speakers based on Visual-Landmark Transformers
Figure 2 for LipFormer: Learning to Lipread Unseen Speakers based on Visual-Landmark Transformers
Figure 3 for LipFormer: Learning to Lipread Unseen Speakers based on Visual-Landmark Transformers
Figure 4 for LipFormer: Learning to Lipread Unseen Speakers based on Visual-Landmark Transformers
Viaarxiv icon

T-Person-GAN: Text-to-Person Image Generation with Identity-Consistency and Manifold Mix-Up

Add code
Aug 18, 2022
Figure 1 for T-Person-GAN: Text-to-Person Image Generation with Identity-Consistency and Manifold Mix-Up
Figure 2 for T-Person-GAN: Text-to-Person Image Generation with Identity-Consistency and Manifold Mix-Up
Figure 3 for T-Person-GAN: Text-to-Person Image Generation with Identity-Consistency and Manifold Mix-Up
Figure 4 for T-Person-GAN: Text-to-Person Image Generation with Identity-Consistency and Manifold Mix-Up
Viaarxiv icon

Cross-speaker Emotion Transfer Based on Speaker Condition Layer Normalization and Semi-Supervised Training in Text-To-Speech

Add code
Oct 11, 2021
Figure 1 for Cross-speaker Emotion Transfer Based on Speaker Condition Layer Normalization and Semi-Supervised Training in Text-To-Speech
Figure 2 for Cross-speaker Emotion Transfer Based on Speaker Condition Layer Normalization and Semi-Supervised Training in Text-To-Speech
Figure 3 for Cross-speaker Emotion Transfer Based on Speaker Condition Layer Normalization and Semi-Supervised Training in Text-To-Speech
Figure 4 for Cross-speaker Emotion Transfer Based on Speaker Condition Layer Normalization and Semi-Supervised Training in Text-To-Speech
Viaarxiv icon

Multi-modal Visual Place Recognition in Dynamics-Invariant Perception Space

Add code
May 17, 2021
Figure 1 for Multi-modal Visual Place Recognition in Dynamics-Invariant Perception Space
Figure 2 for Multi-modal Visual Place Recognition in Dynamics-Invariant Perception Space
Figure 3 for Multi-modal Visual Place Recognition in Dynamics-Invariant Perception Space
Figure 4 for Multi-modal Visual Place Recognition in Dynamics-Invariant Perception Space
Viaarxiv icon