Alert button
Picture for Xiang Kong

Xiang Kong

Alert button

MM1: Methods, Analysis & Insights from Multimodal LLM Pre-training

Add code
Bookmark button
Alert button
Mar 22, 2024
Brandon McKinzie, Zhe Gan, Jean-Philippe Fauconnier, Sam Dodge, Bowen Zhang, Philipp Dufter, Dhruti Shah, Xianzhi Du, Futang Peng, Floris Weers, Anton Belyi, Haotian Zhang, Karanjeet Singh, Doug Kang, Ankur Jain, Hongyu Hè, Max Schwarzer, Tom Gunter, Xiang Kong, Aonan Zhang, Jianyu Wang, Chong Wang, Nan Du, Tao Lei, Sam Wiseman, Guoli Yin, Mark Lee, Zirui Wang, Ruoming Pang, Peter Grasch, Alexander Toshev, Yinfei Yang

Figure 1 for MM1: Methods, Analysis & Insights from Multimodal LLM Pre-training
Figure 2 for MM1: Methods, Analysis & Insights from Multimodal LLM Pre-training
Figure 3 for MM1: Methods, Analysis & Insights from Multimodal LLM Pre-training
Figure 4 for MM1: Methods, Analysis & Insights from Multimodal LLM Pre-training
Viaarxiv icon

Direct Large Language Model Alignment Through Self-Rewarding Contrastive Prompt Distillation

Add code
Bookmark button
Alert button
Feb 19, 2024
Aiwei Liu, Haoping Bai, Zhiyun Lu, Xiang Kong, Simon Wang, Jiulong Shan, Meng Cao, Lijie Wen

Figure 1 for Direct Large Language Model Alignment Through Self-Rewarding Contrastive Prompt Distillation
Figure 2 for Direct Large Language Model Alignment Through Self-Rewarding Contrastive Prompt Distillation
Figure 3 for Direct Large Language Model Alignment Through Self-Rewarding Contrastive Prompt Distillation
Figure 4 for Direct Large Language Model Alignment Through Self-Rewarding Contrastive Prompt Distillation
Viaarxiv icon

Mega: Moving Average Equipped Gated Attention

Add code
Bookmark button
Alert button
Sep 26, 2022
Xuezhe Ma, Chunting Zhou, Xiang Kong, Junxian He, Liangke Gui, Graham Neubig, Jonathan May, Luke Zettlemoyer

Figure 1 for Mega: Moving Average Equipped Gated Attention
Figure 2 for Mega: Moving Average Equipped Gated Attention
Figure 3 for Mega: Moving Average Equipped Gated Attention
Figure 4 for Mega: Moving Average Equipped Gated Attention
Viaarxiv icon

Multilingual Neural Machine Translation with Deep Encoder and Multiple Shallow Decoders

Add code
Bookmark button
Alert button
Jun 05, 2022
Xiang Kong, Adithya Renduchintala, James Cross, Yuqing Tang, Jiatao Gu, Xian Li

Figure 1 for Multilingual Neural Machine Translation with Deep Encoder and Multiple Shallow Decoders
Figure 2 for Multilingual Neural Machine Translation with Deep Encoder and Multiple Shallow Decoders
Figure 3 for Multilingual Neural Machine Translation with Deep Encoder and Multiple Shallow Decoders
Figure 4 for Multilingual Neural Machine Translation with Deep Encoder and Multiple Shallow Decoders
Viaarxiv icon

BLT: Bidirectional Layout Transformer for Controllable Layout Generation

Add code
Bookmark button
Alert button
Dec 09, 2021
Xiang Kong, Lu Jiang, Huiwen Chang, Han Zhang, Yuan Hao, Haifeng Gong, Irfan Essa

Figure 1 for BLT: Bidirectional Layout Transformer for Controllable Layout Generation
Figure 2 for BLT: Bidirectional Layout Transformer for Controllable Layout Generation
Figure 3 for BLT: Bidirectional Layout Transformer for Controllable Layout Generation
Figure 4 for BLT: Bidirectional Layout Transformer for Controllable Layout Generation
Viaarxiv icon

Luna: Linear Unified Nested Attention

Add code
Bookmark button
Alert button
Jun 03, 2021
Xuezhe Ma, Xiang Kong, Sinong Wang, Chunting Zhou, Jonathan May, Hao Ma, Luke Zettlemoyer

Figure 1 for Luna: Linear Unified Nested Attention
Figure 2 for Luna: Linear Unified Nested Attention
Figure 3 for Luna: Linear Unified Nested Attention
Figure 4 for Luna: Linear Unified Nested Attention
Viaarxiv icon

Fully Non-autoregressive Neural Machine Translation: Tricks of the Trade

Add code
Bookmark button
Alert button
Dec 31, 2020
Jiatao Gu, Xiang Kong

Figure 1 for Fully Non-autoregressive Neural Machine Translation: Tricks of the Trade
Figure 2 for Fully Non-autoregressive Neural Machine Translation: Tricks of the Trade
Figure 3 for Fully Non-autoregressive Neural Machine Translation: Tricks of the Trade
Figure 4 for Fully Non-autoregressive Neural Machine Translation: Tricks of the Trade
Viaarxiv icon

Incorporating a Local Translation Mechanism into Non-autoregressive Translation

Add code
Bookmark button
Alert button
Nov 12, 2020
Xiang Kong, Zhisong Zhang, Eduard Hovy

Figure 1 for Incorporating a Local Translation Mechanism into Non-autoregressive Translation
Figure 2 for Incorporating a Local Translation Mechanism into Non-autoregressive Translation
Figure 3 for Incorporating a Local Translation Mechanism into Non-autoregressive Translation
Figure 4 for Incorporating a Local Translation Mechanism into Non-autoregressive Translation
Viaarxiv icon

Deep Transformers with Latent Depth

Add code
Bookmark button
Alert button
Oct 16, 2020
Xian Li, Asa Cooper Stickland, Yuqing Tang, Xiang Kong

Figure 1 for Deep Transformers with Latent Depth
Figure 2 for Deep Transformers with Latent Depth
Figure 3 for Deep Transformers with Latent Depth
Figure 4 for Deep Transformers with Latent Depth
Viaarxiv icon

SCDE: Sentence Cloze Dataset with High Quality Distractors From Examinations

Add code
Bookmark button
Alert button
Apr 27, 2020
Xiang Kong, Varun Gangal, Eduard Hovy

Figure 1 for SCDE: Sentence Cloze Dataset with High Quality Distractors From Examinations
Figure 2 for SCDE: Sentence Cloze Dataset with High Quality Distractors From Examinations
Figure 3 for SCDE: Sentence Cloze Dataset with High Quality Distractors From Examinations
Figure 4 for SCDE: Sentence Cloze Dataset with High Quality Distractors From Examinations
Viaarxiv icon