Picture for Jun Yu

Jun Yu

Lehigh University

Enhancing Emotion Recognition in Conversation through Emotional Cross-Modal Fusion and Inter-class Contrastive Learning

Add code
May 28, 2024
Figure 1 for Enhancing Emotion Recognition in Conversation through Emotional Cross-Modal Fusion and Inter-class Contrastive Learning
Figure 2 for Enhancing Emotion Recognition in Conversation through Emotional Cross-Modal Fusion and Inter-class Contrastive Learning
Figure 3 for Enhancing Emotion Recognition in Conversation through Emotional Cross-Modal Fusion and Inter-class Contrastive Learning
Figure 4 for Enhancing Emotion Recognition in Conversation through Emotional Cross-Modal Fusion and Inter-class Contrastive Learning
Viaarxiv icon

RSET: Remapping-based Sorting Method for Emotion Transfer Speech Synthesis

Add code
May 27, 2024
Figure 1 for RSET: Remapping-based Sorting Method for Emotion Transfer Speech Synthesis
Figure 2 for RSET: Remapping-based Sorting Method for Emotion Transfer Speech Synthesis
Figure 3 for RSET: Remapping-based Sorting Method for Emotion Transfer Speech Synthesis
Figure 4 for RSET: Remapping-based Sorting Method for Emotion Transfer Speech Synthesis
Viaarxiv icon

Incremental Pseudo-Labeling for Black-Box Unsupervised Domain Adaptation

Add code
May 26, 2024
Viaarxiv icon

Imp: Highly Capable Large Multimodal Models for Mobile Devices

Add code
May 20, 2024
Viaarxiv icon

Unleashing the Power of Multi-Task Learning: A Comprehensive Survey Spanning Traditional, Deep, and Pretrained Foundation Model Eras

Add code
Apr 29, 2024
Figure 1 for Unleashing the Power of Multi-Task Learning: A Comprehensive Survey Spanning Traditional, Deep, and Pretrained Foundation Model Eras
Figure 2 for Unleashing the Power of Multi-Task Learning: A Comprehensive Survey Spanning Traditional, Deep, and Pretrained Foundation Model Eras
Figure 3 for Unleashing the Power of Multi-Task Learning: A Comprehensive Survey Spanning Traditional, Deep, and Pretrained Foundation Model Eras
Figure 4 for Unleashing the Power of Multi-Task Learning: A Comprehensive Survey Spanning Traditional, Deep, and Pretrained Foundation Model Eras
Viaarxiv icon

Frequency-Guided Multi-Level Human Action Anomaly Detection with Normalizing Flows

Add code
Apr 26, 2024
Figure 1 for Frequency-Guided Multi-Level Human Action Anomaly Detection with Normalizing Flows
Figure 2 for Frequency-Guided Multi-Level Human Action Anomaly Detection with Normalizing Flows
Figure 3 for Frequency-Guided Multi-Level Human Action Anomaly Detection with Normalizing Flows
Figure 4 for Frequency-Guided Multi-Level Human Action Anomaly Detection with Normalizing Flows
Viaarxiv icon

Leveraging Large Language Model to Generate a Novel Metaheuristic Algorithm with CRISPE Framework

Add code
Mar 25, 2024
Figure 1 for Leveraging Large Language Model to Generate a Novel Metaheuristic Algorithm with CRISPE Framework
Figure 2 for Leveraging Large Language Model to Generate a Novel Metaheuristic Algorithm with CRISPE Framework
Figure 3 for Leveraging Large Language Model to Generate a Novel Metaheuristic Algorithm with CRISPE Framework
Figure 4 for Leveraging Large Language Model to Generate a Novel Metaheuristic Algorithm with CRISPE Framework
Viaarxiv icon

Tackling Noisy Labels with Network Parameter Additive Decomposition

Add code
Mar 20, 2024
Figure 1 for Tackling Noisy Labels with Network Parameter Additive Decomposition
Figure 2 for Tackling Noisy Labels with Network Parameter Additive Decomposition
Figure 3 for Tackling Noisy Labels with Network Parameter Additive Decomposition
Figure 4 for Tackling Noisy Labels with Network Parameter Additive Decomposition
Viaarxiv icon

AUD-TGN: Advancing Action Unit Detection with Temporal Convolution and GPT-2 in Wild Audiovisual Contexts

Add code
Mar 20, 2024
Figure 1 for AUD-TGN: Advancing Action Unit Detection with Temporal Convolution and GPT-2 in Wild Audiovisual Contexts
Figure 2 for AUD-TGN: Advancing Action Unit Detection with Temporal Convolution and GPT-2 in Wild Audiovisual Contexts
Figure 3 for AUD-TGN: Advancing Action Unit Detection with Temporal Convolution and GPT-2 in Wild Audiovisual Contexts
Viaarxiv icon

Multimodal Fusion Method with Spatiotemporal Sequences and Relationship Learning for Valence-Arousal Estimation

Add code
Mar 20, 2024
Figure 1 for Multimodal Fusion Method with Spatiotemporal Sequences and Relationship Learning for Valence-Arousal Estimation
Figure 2 for Multimodal Fusion Method with Spatiotemporal Sequences and Relationship Learning for Valence-Arousal Estimation
Figure 3 for Multimodal Fusion Method with Spatiotemporal Sequences and Relationship Learning for Valence-Arousal Estimation
Viaarxiv icon