Picture for Shaohan Huang

Shaohan Huang

Retentive Network: A Successor to Transformer for Large Language Models

Add code
Aug 09, 2023
Viaarxiv icon

Scaling Sentence Embeddings with Large Language Models

Add code
Jul 31, 2023
Viaarxiv icon

LongNet: Scaling Transformers to 1,000,000,000 Tokens

Add code
Jul 19, 2023
Viaarxiv icon

Kosmos-2: Grounding Multimodal Large Language Models to the World

Add code
Jul 13, 2023
Viaarxiv icon

Learning Music Sequence Representation from Text Supervision

Add code
May 31, 2023
Viaarxiv icon

Dual-Alignment Pre-training for Cross-lingual Sentence Embedding

Add code
May 16, 2023
Viaarxiv icon

Pre-training Language Model as a Multi-perspective Course Learner

Add code
May 06, 2023
Viaarxiv icon

UPRISE: Universal Prompt Retrieval for Improving Zero-Shot Evaluation

Add code
Mar 22, 2023
Figure 1 for UPRISE: Universal Prompt Retrieval for Improving Zero-Shot Evaluation
Figure 2 for UPRISE: Universal Prompt Retrieval for Improving Zero-Shot Evaluation
Figure 3 for UPRISE: Universal Prompt Retrieval for Improving Zero-Shot Evaluation
Figure 4 for UPRISE: Universal Prompt Retrieval for Improving Zero-Shot Evaluation
Viaarxiv icon

Language Is Not All You Need: Aligning Perception with Language Models

Add code
Mar 01, 2023
Figure 1 for Language Is Not All You Need: Aligning Perception with Language Models
Figure 2 for Language Is Not All You Need: Aligning Perception with Language Models
Figure 3 for Language Is Not All You Need: Aligning Perception with Language Models
Figure 4 for Language Is Not All You Need: Aligning Perception with Language Models
Viaarxiv icon

A Length-Extrapolatable Transformer

Add code
Dec 20, 2022
Viaarxiv icon