Picture for Hai Yu

Hai Yu

Skip-Layer Attention: Bridging Abstract and Detailed Dependencies in Transformers

Add code
Jun 17, 2024
Figure 1 for Skip-Layer Attention: Bridging Abstract and Detailed Dependencies in Transformers
Figure 2 for Skip-Layer Attention: Bridging Abstract and Detailed Dependencies in Transformers
Figure 3 for Skip-Layer Attention: Bridging Abstract and Detailed Dependencies in Transformers
Figure 4 for Skip-Layer Attention: Bridging Abstract and Detailed Dependencies in Transformers
Viaarxiv icon

Loss Masking Is Not Needed in Decoder-only Transformer for Discrete-token Based ASR

Add code
Nov 08, 2023
Figure 1 for Loss Masking Is Not Needed in Decoder-only Transformer for Discrete-token Based ASR
Figure 2 for Loss Masking Is Not Needed in Decoder-only Transformer for Discrete-token Based ASR
Figure 3 for Loss Masking Is Not Needed in Decoder-only Transformer for Discrete-token Based ASR
Figure 4 for Loss Masking Is Not Needed in Decoder-only Transformer for Discrete-token Based ASR
Viaarxiv icon

Improving Long Document Topic Segmentation Models With Enhanced Coherence Modeling

Add code
Oct 23, 2023
Figure 1 for Improving Long Document Topic Segmentation Models With Enhanced Coherence Modeling
Figure 2 for Improving Long Document Topic Segmentation Models With Enhanced Coherence Modeling
Figure 3 for Improving Long Document Topic Segmentation Models With Enhanced Coherence Modeling
Figure 4 for Improving Long Document Topic Segmentation Models With Enhanced Coherence Modeling
Viaarxiv icon

Ditto: A Simple and Efficient Approach to Improve Sentence Embeddings

Add code
May 18, 2023
Figure 1 for Ditto: A Simple and Efficient Approach to Improve Sentence Embeddings
Figure 2 for Ditto: A Simple and Efficient Approach to Improve Sentence Embeddings
Figure 3 for Ditto: A Simple and Efficient Approach to Improve Sentence Embeddings
Figure 4 for Ditto: A Simple and Efficient Approach to Improve Sentence Embeddings
Viaarxiv icon

MUG: A General Meeting Understanding and Generation Benchmark

Add code
Mar 27, 2023
Figure 1 for MUG: A General Meeting Understanding and Generation Benchmark
Figure 2 for MUG: A General Meeting Understanding and Generation Benchmark
Figure 3 for MUG: A General Meeting Understanding and Generation Benchmark
Figure 4 for MUG: A General Meeting Understanding and Generation Benchmark
Viaarxiv icon

Overview of the ICASSP 2023 General Meeting Understanding and Generation Challenge

Add code
Mar 24, 2023
Figure 1 for Overview of the ICASSP 2023 General Meeting Understanding and Generation Challenge
Figure 2 for Overview of the ICASSP 2023 General Meeting Understanding and Generation Challenge
Viaarxiv icon

Persia: An Open, Hybrid System Scaling Deep Learning-based Recommenders up to 100 Trillion Parameters

Add code
Nov 23, 2021
Figure 1 for Persia: An Open, Hybrid System Scaling Deep Learning-based Recommenders up to 100 Trillion Parameters
Figure 2 for Persia: An Open, Hybrid System Scaling Deep Learning-based Recommenders up to 100 Trillion Parameters
Figure 3 for Persia: An Open, Hybrid System Scaling Deep Learning-based Recommenders up to 100 Trillion Parameters
Figure 4 for Persia: An Open, Hybrid System Scaling Deep Learning-based Recommenders up to 100 Trillion Parameters
Viaarxiv icon