Alert button
Picture for Rulin Shao

Rulin Shao

Alert button

Language models scale reliably with over-training and on downstream tasks

Add code
Bookmark button
Alert button
Mar 13, 2024
Samir Yitzhak Gadre, Georgios Smyrnis, Vaishaal Shankar, Suchin Gururangan, Mitchell Wortsman, Rulin Shao, Jean Mercat, Alex Fang, Jeffrey Li, Sedrick Keh, Rui Xin, Marianna Nezhurina, Igor Vasiljevic, Jenia Jitsev, Alexandros G. Dimakis, Gabriel Ilharco, Shuran Song, Thomas Kollar, Yair Carmon, Achal Dave, Reinhard Heckel, Niklas Muennighoff, Ludwig Schmidt

Figure 1 for Language models scale reliably with over-training and on downstream tasks
Figure 2 for Language models scale reliably with over-training and on downstream tasks
Figure 3 for Language models scale reliably with over-training and on downstream tasks
Figure 4 for Language models scale reliably with over-training and on downstream tasks
Viaarxiv icon

Vision-Flan: Scaling Human-Labeled Tasks in Visual Instruction Tuning

Add code
Bookmark button
Alert button
Feb 18, 2024
Zhiyang Xu, Chao Feng, Rulin Shao, Trevor Ashby, Ying Shen, Di Jin, Yu Cheng, Qifan Wang, Lifu Huang

Viaarxiv icon

LightSeq: Sequence Level Parallelism for Distributed Training of Long Context Transformers

Add code
Bookmark button
Alert button
Oct 05, 2023
Dacheng Li, Rulin Shao, Anze Xie, Eric P. Xing, Joseph E. Gonzalez, Ion Stoica, Xuezhe Ma, Hao Zhang

Figure 1 for LightSeq: Sequence Level Parallelism for Distributed Training of Long Context Transformers
Figure 2 for LightSeq: Sequence Level Parallelism for Distributed Training of Long Context Transformers
Figure 3 for LightSeq: Sequence Level Parallelism for Distributed Training of Long Context Transformers
Figure 4 for LightSeq: Sequence Level Parallelism for Distributed Training of Long Context Transformers
Viaarxiv icon

VisIT-Bench: A Benchmark for Vision-Language Instruction Following Inspired by Real-World Use

Add code
Bookmark button
Alert button
Aug 12, 2023
Yonatan Bitton, Hritik Bansal, Jack Hessel, Rulin Shao, Wanrong Zhu, Anas Awadalla, Josh Gardner, Rohan Taori, Ludwig Schimdt

Figure 1 for VisIT-Bench: A Benchmark for Vision-Language Instruction Following Inspired by Real-World Use
Figure 2 for VisIT-Bench: A Benchmark for Vision-Language Instruction Following Inspired by Real-World Use
Figure 3 for VisIT-Bench: A Benchmark for Vision-Language Instruction Following Inspired by Real-World Use
Figure 4 for VisIT-Bench: A Benchmark for Vision-Language Instruction Following Inspired by Real-World Use
Viaarxiv icon

Cross-modal Attention Congruence Regularization for Vision-Language Relation Alignment

Add code
Bookmark button
Alert button
Dec 20, 2022
Rohan Pandey, Rulin Shao, Paul Pu Liang, Ruslan Salakhutdinov, Louis-Philippe Morency

Figure 1 for Cross-modal Attention Congruence Regularization for Vision-Language Relation Alignment
Figure 2 for Cross-modal Attention Congruence Regularization for Vision-Language Relation Alignment
Figure 3 for Cross-modal Attention Congruence Regularization for Vision-Language Relation Alignment
Figure 4 for Cross-modal Attention Congruence Regularization for Vision-Language Relation Alignment
Viaarxiv icon

MPCFormer: fast, performant and private Transformer inference with MPC

Add code
Bookmark button
Alert button
Nov 02, 2022
Dacheng Li, Rulin Shao, Hongyi Wang, Han Guo, Eric P. Xing, Hao Zhang

Figure 1 for MPCFormer: fast, performant and private Transformer inference with MPC
Figure 2 for MPCFormer: fast, performant and private Transformer inference with MPC
Figure 3 for MPCFormer: fast, performant and private Transformer inference with MPC
Figure 4 for MPCFormer: fast, performant and private Transformer inference with MPC
Viaarxiv icon

How and When Adversarial Robustness Transfers in Knowledge Distillation?

Add code
Bookmark button
Alert button
Oct 22, 2021
Rulin Shao, Jinfeng Yi, Pin-Yu Chen, Cho-Jui Hsieh

Figure 1 for How and When Adversarial Robustness Transfers in Knowledge Distillation?
Figure 2 for How and When Adversarial Robustness Transfers in Knowledge Distillation?
Figure 3 for How and When Adversarial Robustness Transfers in Knowledge Distillation?
Figure 4 for How and When Adversarial Robustness Transfers in Knowledge Distillation?
Viaarxiv icon

On the Adversarial Robustness of Visual Transformers

Add code
Bookmark button
Alert button
Mar 29, 2021
Rulin Shao, Zhouxing Shi, Jinfeng Yi, Pin-Yu Chen, Cho-Jui Hsieh

Figure 1 for On the Adversarial Robustness of Visual Transformers
Figure 2 for On the Adversarial Robustness of Visual Transformers
Figure 3 for On the Adversarial Robustness of Visual Transformers
Figure 4 for On the Adversarial Robustness of Visual Transformers
Viaarxiv icon

Robust Text CAPTCHAs Using Adversarial Examples

Add code
Bookmark button
Alert button
Jan 07, 2021
Rulin Shao, Zhouxing Shi, Jinfeng Yi, Pin-Yu Chen, Cho-Jui Hsieh

Figure 1 for Robust Text CAPTCHAs Using Adversarial Examples
Figure 2 for Robust Text CAPTCHAs Using Adversarial Examples
Figure 3 for Robust Text CAPTCHAs Using Adversarial Examples
Figure 4 for Robust Text CAPTCHAs Using Adversarial Examples
Viaarxiv icon

Stochastic Channel-Based Federated Learning for Medical Data Privacy Preserving

Add code
Bookmark button
Alert button
Nov 15, 2019
Rulin Shao, Hongyu He, Hui Liu, Dianbo Liu

Figure 1 for Stochastic Channel-Based Federated Learning for Medical Data Privacy Preserving
Figure 2 for Stochastic Channel-Based Federated Learning for Medical Data Privacy Preserving
Viaarxiv icon