Get our free extension to see links to code for papers anywhere online!

Chrome logo  Add to Chrome

Firefox logo Add to Firefox

SE-MoE: A Scalable and Efficient Mixture-of-Experts Distributed Training and Inference System


May 20, 2022
Liang Shen , Zhihua Wu , WeiBao Gong , Hongxiang Hao , Yangfan Bai , HuaChao Wu , Xinxuan Wu , Haoyi Xiong , Dianhai Yu , Yanjun Ma


   Access Paper or Ask Questions

  • Share via Twitter
  • Share via Facebook
  • Share via LinkedIn
  • Share via Whatsapp
  • Share via Messenger
  • Share via Email

ERNIE 3.0 Titan: Exploring Larger-scale Knowledge Enhanced Pre-training for Language Understanding and Generation


Dec 23, 2021
Shuohuan Wang , Yu Sun , Yang Xiang , Zhihua Wu , Siyu Ding , Weibao Gong , Shikun Feng , Junyuan Shang , Yanbin Zhao , Chao Pang , Jiaxiang Liu , Xuyi Chen , Yuxiang Lu , Weixin Liu , Xi Wang , Yangfan Bai , Qiuliang Chen , Li Zhao , Shiyong Li , Peng Sun , Dianhai Yu , Yanjun Ma , Hao Tian , Hua Wu , Tian Wu , Wei Zeng , Ge Li , Wen Gao , Haifeng Wang

* arXiv admin note: text overlap with arXiv:2107.02137 

   Access Paper or Ask Questions

  • Share via Twitter
  • Share via Facebook
  • Share via LinkedIn
  • Share via Whatsapp
  • Share via Messenger
  • Share via Email