Alert button
Picture for Xuxi Chen

Xuxi Chen

Alert button

Take the Bull by the Horns: Hard Sample-Reweighted Continual Training Improves LLM Generalization

Add code
Bookmark button
Alert button
Mar 01, 2024
Xuxi Chen, Zhendong Wang, Daouda Sow, Junjie Yang, Tianlong Chen, Yingbin Liang, Mingyuan Zhou, Zhangyang Wang

Viaarxiv icon

Rethinking PGD Attack: Is Sign Function Necessary?

Add code
Bookmark button
Alert button
Dec 03, 2023
Junjie Yang, Tianlong Chen, Xuxi Chen, Zhangyang Wang, Yingbin Liang

Figure 1 for Rethinking PGD Attack: Is Sign Function Necessary?
Figure 2 for Rethinking PGD Attack: Is Sign Function Necessary?
Figure 3 for Rethinking PGD Attack: Is Sign Function Necessary?
Figure 4 for Rethinking PGD Attack: Is Sign Function Necessary?
Viaarxiv icon

Orca 2: Teaching Small Language Models How to Reason

Add code
Bookmark button
Alert button
Nov 21, 2023
Arindam Mitra, Luciano Del Corro, Shweti Mahajan, Andres Codas, Clarisse Simoes, Sahaj Agarwal, Xuxi Chen, Anastasia Razdaibiedina, Erik Jones, Kriti Aggarwal, Hamid Palangi, Guoqing Zheng, Corby Rosset, Hamed Khanpour, Ahmed Awadallah

Viaarxiv icon

Data Distillation Can Be Like Vodka: Distilling More Times For Better Quality

Add code
Bookmark button
Alert button
Oct 10, 2023
Xuxi Chen, Yu Yang, Zhangyang Wang, Baharan Mirzasoleiman

Figure 1 for Data Distillation Can Be Like Vodka: Distilling More Times For Better Quality
Figure 2 for Data Distillation Can Be Like Vodka: Distilling More Times For Better Quality
Figure 3 for Data Distillation Can Be Like Vodka: Distilling More Times For Better Quality
Figure 4 for Data Distillation Can Be Like Vodka: Distilling More Times For Better Quality
Viaarxiv icon

Sparsity May Cry: Let Us Fail (Current) Sparse Neural Networks Together!

Add code
Bookmark button
Alert button
Mar 03, 2023
Shiwei Liu, Tianlong Chen, Zhenyu Zhang, Xuxi Chen, Tianjin Huang, Ajay Jaiswal, Zhangyang Wang

Figure 1 for Sparsity May Cry: Let Us Fail (Current) Sparse Neural Networks Together!
Figure 2 for Sparsity May Cry: Let Us Fail (Current) Sparse Neural Networks Together!
Figure 3 for Sparsity May Cry: Let Us Fail (Current) Sparse Neural Networks Together!
Figure 4 for Sparsity May Cry: Let Us Fail (Current) Sparse Neural Networks Together!
Viaarxiv icon

M-L2O: Towards Generalizable Learning-to-Optimize by Test-Time Fast Self-Adaptation

Add code
Bookmark button
Alert button
Feb 28, 2023
Junjie Yang, Xuxi Chen, Tianlong Chen, Zhangyang Wang, Yingbin Liang

Figure 1 for M-L2O: Towards Generalizable Learning-to-Optimize by Test-Time Fast Self-Adaptation
Figure 2 for M-L2O: Towards Generalizable Learning-to-Optimize by Test-Time Fast Self-Adaptation
Figure 3 for M-L2O: Towards Generalizable Learning-to-Optimize by Test-Time Fast Self-Adaptation
Figure 4 for M-L2O: Towards Generalizable Learning-to-Optimize by Test-Time Fast Self-Adaptation
Viaarxiv icon

Is Attention All NeRF Needs?

Add code
Bookmark button
Alert button
Jul 27, 2022
Mukund Varma T, Peihao Wang, Xuxi Chen, Tianlong Chen, Subhashini Venugopalan, Zhangyang Wang

Figure 1 for Is Attention All NeRF Needs?
Figure 2 for Is Attention All NeRF Needs?
Figure 3 for Is Attention All NeRF Needs?
Figure 4 for Is Attention All NeRF Needs?
Viaarxiv icon

More ConvNets in the 2020s: Scaling up Kernels Beyond 51x51 using Sparsity

Add code
Bookmark button
Alert button
Jul 07, 2022
Shiwei Liu, Tianlong Chen, Xiaohan Chen, Xuxi Chen, Qiao Xiao, Boqian Wu, Mykola Pechenizkiy, Decebal Mocanu, Zhangyang Wang

Figure 1 for More ConvNets in the 2020s: Scaling up Kernels Beyond 51x51 using Sparsity
Figure 2 for More ConvNets in the 2020s: Scaling up Kernels Beyond 51x51 using Sparsity
Figure 3 for More ConvNets in the 2020s: Scaling up Kernels Beyond 51x51 using Sparsity
Figure 4 for More ConvNets in the 2020s: Scaling up Kernels Beyond 51x51 using Sparsity
Viaarxiv icon