Alert button
Picture for Nan Ding

Nan Ding

Alert button

CausalLM is not optimal for in-context learning

Add code
Bookmark button
Alert button
Sep 03, 2023
Nan Ding, Tomer Levinboim, Jialin Wu, Sebastian Goodman, Radu Soricut

Figure 1 for CausalLM is not optimal for in-context learning
Figure 2 for CausalLM is not optimal for in-context learning
Figure 3 for CausalLM is not optimal for in-context learning
Figure 4 for CausalLM is not optimal for in-context learning
Viaarxiv icon

Improving Robust Generalization by Direct PAC-Bayesian Bound Minimization

Add code
Bookmark button
Alert button
Nov 22, 2022
Zifan Wang, Nan Ding, Tomer Levinboim, Xi Chen, Radu Soricut

Figure 1 for Improving Robust Generalization by Direct PAC-Bayesian Bound Minimization
Figure 2 for Improving Robust Generalization by Direct PAC-Bayesian Bound Minimization
Figure 3 for Improving Robust Generalization by Direct PAC-Bayesian Bound Minimization
Figure 4 for Improving Robust Generalization by Direct PAC-Bayesian Bound Minimization
Viaarxiv icon

PaLI: A Jointly-Scaled Multilingual Language-Image Model

Add code
Bookmark button
Alert button
Sep 16, 2022
Xi Chen, Xiao Wang, Soravit Changpinyo, AJ Piergiovanni, Piotr Padlewski, Daniel Salz, Sebastian Goodman, Adam Grycner, Basil Mustafa, Lucas Beyer, Alexander Kolesnikov, Joan Puigcerver, Nan Ding, Keran Rong, Hassan Akbari, Gaurav Mishra, Linting Xue, Ashish Thapliyal, James Bradbury, Weicheng Kuo, Mojtaba Seyedhosseini, Chao Jia, Burcu Karagol Ayan, Carlos Riquelme, Andreas Steiner, Anelia Angelova, Xiaohua Zhai, Neil Houlsby, Radu Soricut

Figure 1 for PaLI: A Jointly-Scaled Multilingual Language-Image Model
Figure 2 for PaLI: A Jointly-Scaled Multilingual Language-Image Model
Figure 3 for PaLI: A Jointly-Scaled Multilingual Language-Image Model
Figure 4 for PaLI: A Jointly-Scaled Multilingual Language-Image Model
Viaarxiv icon

All You May Need for VQA are Image Captions

Add code
Bookmark button
Alert button
May 04, 2022
Soravit Changpinyo, Doron Kukliansky, Idan Szpektor, Xi Chen, Nan Ding, Radu Soricut

Figure 1 for All You May Need for VQA are Image Captions
Figure 2 for All You May Need for VQA are Image Captions
Figure 3 for All You May Need for VQA are Image Captions
Figure 4 for All You May Need for VQA are Image Captions
Viaarxiv icon

PACTran: PAC-Bayesian Metrics for Estimating the Transferability of Pretrained Models to Classification Tasks

Add code
Bookmark button
Alert button
Mar 10, 2022
Nan Ding, Xi Chen, Tomer Levinboim, Beer Changpinyo, Radu Soricut

Figure 1 for PACTran: PAC-Bayesian Metrics for Estimating the Transferability of Pretrained Models to Classification Tasks
Figure 2 for PACTran: PAC-Bayesian Metrics for Estimating the Transferability of Pretrained Models to Classification Tasks
Figure 3 for PACTran: PAC-Bayesian Metrics for Estimating the Transferability of Pretrained Models to Classification Tasks
Figure 4 for PACTran: PAC-Bayesian Metrics for Estimating the Transferability of Pretrained Models to Classification Tasks
Viaarxiv icon

Bridging the Gap Between Practice and PAC-Bayes Theory in Few-Shot Meta-Learning

Add code
Bookmark button
Alert button
May 28, 2021
Nan Ding, Xi Chen, Tomer Levinboim, Sebastian Goodman, Radu Soricut

Figure 1 for Bridging the Gap Between Practice and PAC-Bayes Theory in Few-Shot Meta-Learning
Figure 2 for Bridging the Gap Between Practice and PAC-Bayes Theory in Few-Shot Meta-Learning
Figure 3 for Bridging the Gap Between Practice and PAC-Bayes Theory in Few-Shot Meta-Learning
Figure 4 for Bridging the Gap Between Practice and PAC-Bayes Theory in Few-Shot Meta-Learning
Viaarxiv icon

Do Transformer Modifications Transfer Across Implementations and Applications?

Add code
Bookmark button
Alert button
Feb 23, 2021
Sharan Narang, Hyung Won Chung, Yi Tay, William Fedus, Thibault Fevry, Michael Matena, Karishma Malkan, Noah Fiedel, Noam Shazeer, Zhenzhong Lan, Yanqi Zhou, Wei Li, Nan Ding, Jake Marcus, Adam Roberts, Colin Raffel

Figure 1 for Do Transformer Modifications Transfer Across Implementations and Applications?
Figure 2 for Do Transformer Modifications Transfer Across Implementations and Applications?
Figure 3 for Do Transformer Modifications Transfer Across Implementations and Applications?
Viaarxiv icon

Conceptual 12M: Pushing Web-Scale Image-Text Pre-Training To Recognize Long-Tail Visual Concepts

Add code
Bookmark button
Alert button
Feb 17, 2021
Soravit Changpinyo, Piyush Sharma, Nan Ding, Radu Soricut

Figure 1 for Conceptual 12M: Pushing Web-Scale Image-Text Pre-Training To Recognize Long-Tail Visual Concepts
Figure 2 for Conceptual 12M: Pushing Web-Scale Image-Text Pre-Training To Recognize Long-Tail Visual Concepts
Figure 3 for Conceptual 12M: Pushing Web-Scale Image-Text Pre-Training To Recognize Long-Tail Visual Concepts
Figure 4 for Conceptual 12M: Pushing Web-Scale Image-Text Pre-Training To Recognize Long-Tail Visual Concepts
Viaarxiv icon

Improving Text Generation Evaluation with Batch Centering and Tempered Word Mover Distance

Add code
Bookmark button
Alert button
Oct 13, 2020
Xi Chen, Nan Ding, Tomer Levinboim, Radu Soricut

Figure 1 for Improving Text Generation Evaluation with Batch Centering and Tempered Word Mover Distance
Figure 2 for Improving Text Generation Evaluation with Batch Centering and Tempered Word Mover Distance
Figure 3 for Improving Text Generation Evaluation with Batch Centering and Tempered Word Mover Distance
Figure 4 for Improving Text Generation Evaluation with Batch Centering and Tempered Word Mover Distance
Viaarxiv icon