Picture for Yu Cheng

Yu Cheng

LLaMA-MoE v2: Exploring Sparsity of LLaMA from Perspective of Mixture-of-Experts with Post-Training

Add code
Nov 24, 2024
Viaarxiv icon

MoE-I$^2$: Compressing Mixture of Experts Models through Inter-Expert Pruning and Intra-Expert Low-Rank Decomposition

Add code
Nov 01, 2024
Figure 1 for MoE-I$^2$: Compressing Mixture of Experts Models through Inter-Expert Pruning and Intra-Expert Low-Rank Decomposition
Figure 2 for MoE-I$^2$: Compressing Mixture of Experts Models through Inter-Expert Pruning and Intra-Expert Low-Rank Decomposition
Figure 3 for MoE-I$^2$: Compressing Mixture of Experts Models through Inter-Expert Pruning and Intra-Expert Low-Rank Decomposition
Figure 4 for MoE-I$^2$: Compressing Mixture of Experts Models through Inter-Expert Pruning and Intra-Expert Low-Rank Decomposition
Viaarxiv icon

Continuous Speech Tokenizer in Text To Speech

Add code
Oct 22, 2024
Viaarxiv icon

RoRA-VLM: Robust Retrieval-Augmented Vision Language Models

Add code
Oct 11, 2024
Figure 1 for RoRA-VLM: Robust Retrieval-Augmented Vision Language Models
Figure 2 for RoRA-VLM: Robust Retrieval-Augmented Vision Language Models
Figure 3 for RoRA-VLM: Robust Retrieval-Augmented Vision Language Models
Figure 4 for RoRA-VLM: Robust Retrieval-Augmented Vision Language Models
Viaarxiv icon

What makes your model a low-empathy or warmth person: Exploring the Origins of Personality in LLMs

Add code
Oct 07, 2024
Figure 1 for What makes your model a low-empathy or warmth person: Exploring the Origins of Personality in LLMs
Figure 2 for What makes your model a low-empathy or warmth person: Exploring the Origins of Personality in LLMs
Figure 3 for What makes your model a low-empathy or warmth person: Exploring the Origins of Personality in LLMs
Figure 4 for What makes your model a low-empathy or warmth person: Exploring the Origins of Personality in LLMs
Viaarxiv icon

Towards World Simulator: Crafting Physical Commonsense-Based Benchmark for Video Generation

Add code
Oct 07, 2024
Viaarxiv icon

Grounded-VideoLLM: Sharpening Fine-grained Temporal Grounding in Video Large Language Models

Add code
Oct 04, 2024
Figure 1 for Grounded-VideoLLM: Sharpening Fine-grained Temporal Grounding in Video Large Language Models
Figure 2 for Grounded-VideoLLM: Sharpening Fine-grained Temporal Grounding in Video Large Language Models
Figure 3 for Grounded-VideoLLM: Sharpening Fine-grained Temporal Grounding in Video Large Language Models
Figure 4 for Grounded-VideoLLM: Sharpening Fine-grained Temporal Grounding in Video Large Language Models
Viaarxiv icon

CLIP-MoE: Towards Building Mixture of Experts for CLIP with Diversified Multiplet Upcycling

Add code
Sep 28, 2024
Viaarxiv icon

SURf: Teaching Large Vision-Language Models to Selectively Utilize Retrieved Information

Add code
Sep 21, 2024
Figure 1 for SURf: Teaching Large Vision-Language Models to Selectively Utilize Retrieved Information
Figure 2 for SURf: Teaching Large Vision-Language Models to Selectively Utilize Retrieved Information
Figure 3 for SURf: Teaching Large Vision-Language Models to Selectively Utilize Retrieved Information
Figure 4 for SURf: Teaching Large Vision-Language Models to Selectively Utilize Retrieved Information
Viaarxiv icon

Look, Compare, Decide: Alleviating Hallucination in Large Vision-Language Models via Multi-View Multi-Path Reasoning

Add code
Aug 30, 2024
Figure 1 for Look, Compare, Decide: Alleviating Hallucination in Large Vision-Language Models via Multi-View Multi-Path Reasoning
Figure 2 for Look, Compare, Decide: Alleviating Hallucination in Large Vision-Language Models via Multi-View Multi-Path Reasoning
Figure 3 for Look, Compare, Decide: Alleviating Hallucination in Large Vision-Language Models via Multi-View Multi-Path Reasoning
Figure 4 for Look, Compare, Decide: Alleviating Hallucination in Large Vision-Language Models via Multi-View Multi-Path Reasoning
Viaarxiv icon