Alert button
Picture for Yanghao Li

Yanghao Li

Alert button

Idempotence and Perceptual Image Compression

Jan 17, 2024
Tongda Xu, Ziran Zhu, Dailan He, Yanghao Li, Lina Guo, Yuanyuan Wang, Zhe Wang, Hongwei Qin, Yan Wang, Jingjing Liu, Ya-Qin Zhang

Viaarxiv icon

Bandwidth-efficient Inference for Neural Image Compression

Sep 07, 2023
Shanzhi Yin, Tongda Xu, Yongsheng Liang, Yuanyuan Wang, Yanghao Li, Yan Wang, Jingjing Liu

Figure 1 for Bandwidth-efficient Inference for Neural Image Compression
Figure 2 for Bandwidth-efficient Inference for Neural Image Compression
Figure 3 for Bandwidth-efficient Inference for Neural Image Compression
Figure 4 for Bandwidth-efficient Inference for Neural Image Compression
Viaarxiv icon

Conditional Perceptual Quality Preserving Image Compression

Aug 16, 2023
Tongda Xu, Qian Zhang, Yanghao Li, Dailan He, Zhe Wang, Yuanyuan Wang, Hongwei Qin, Yan Wang, Jingjing Liu, Ya-Qin Zhang

Figure 1 for Conditional Perceptual Quality Preserving Image Compression
Figure 2 for Conditional Perceptual Quality Preserving Image Compression
Figure 3 for Conditional Perceptual Quality Preserving Image Compression
Figure 4 for Conditional Perceptual Quality Preserving Image Compression
Viaarxiv icon

R-MAE: Regions Meet Masked Autoencoders

Jun 08, 2023
Duy-Kien Nguyen, Vaibhav Aggarwal, Yanghao Li, Martin R. Oswald, Alexander Kirillov, Cees G. M. Snoek, Xinlei Chen

Figure 1 for R-MAE: Regions Meet Masked Autoencoders
Figure 2 for R-MAE: Regions Meet Masked Autoencoders
Figure 3 for R-MAE: Regions Meet Masked Autoencoders
Figure 4 for R-MAE: Regions Meet Masked Autoencoders
Viaarxiv icon

Hiera: A Hierarchical Vision Transformer without the Bells-and-Whistles

Jun 01, 2023
Chaitanya Ryali, Yuan-Ting Hu, Daniel Bolya, Chen Wei, Haoqi Fan, Po-Yao Huang, Vaibhav Aggarwal, Arkabandhu Chowdhury, Omid Poursaeed, Judy Hoffman, Jitendra Malik, Yanghao Li, Christoph Feichtenhofer

Figure 1 for Hiera: A Hierarchical Vision Transformer without the Bells-and-Whistles
Figure 2 for Hiera: A Hierarchical Vision Transformer without the Bells-and-Whistles
Figure 3 for Hiera: A Hierarchical Vision Transformer without the Bells-and-Whistles
Figure 4 for Hiera: A Hierarchical Vision Transformer without the Bells-and-Whistles
Viaarxiv icon

Diffusion Models as Masked Autoencoders

Apr 06, 2023
Chen Wei, Karttikeya Mangalam, Po-Yao Huang, Yanghao Li, Haoqi Fan, Hu Xu, Huiyu Wang, Cihang Xie, Alan Yuille, Christoph Feichtenhofer

Figure 1 for Diffusion Models as Masked Autoencoders
Figure 2 for Diffusion Models as Masked Autoencoders
Figure 3 for Diffusion Models as Masked Autoencoders
Figure 4 for Diffusion Models as Masked Autoencoders
Viaarxiv icon

Efficient Semantic Segmentation by Altering Resolutions for Compressed Videos

Mar 13, 2023
Yubin Hu, Yuze He, Yanghao Li, Jisheng Li, Yuxing Han, Jiangtao Wen, Yong-Jin Liu

Figure 1 for Efficient Semantic Segmentation by Altering Resolutions for Compressed Videos
Figure 2 for Efficient Semantic Segmentation by Altering Resolutions for Compressed Videos
Figure 3 for Efficient Semantic Segmentation by Altering Resolutions for Compressed Videos
Figure 4 for Efficient Semantic Segmentation by Altering Resolutions for Compressed Videos
Viaarxiv icon

Reversible Vision Transformers

Feb 09, 2023
Karttikeya Mangalam, Haoqi Fan, Yanghao Li, Chao-Yuan Wu, Bo Xiong, Christoph Feichtenhofer, Jitendra Malik

Figure 1 for Reversible Vision Transformers
Figure 2 for Reversible Vision Transformers
Figure 3 for Reversible Vision Transformers
Figure 4 for Reversible Vision Transformers
Viaarxiv icon

MAViL: Masked Audio-Video Learners

Dec 15, 2022
Po-Yao Huang, Vasu Sharma, Hu Xu, Chaitanya Ryali, Haoqi Fan, Yanghao Li, Shang-Wen Li, Gargi Ghosh, Jitendra Malik, Christoph Feichtenhofer

Figure 1 for MAViL: Masked Audio-Video Learners
Figure 2 for MAViL: Masked Audio-Video Learners
Figure 3 for MAViL: Masked Audio-Video Learners
Figure 4 for MAViL: Masked Audio-Video Learners
Viaarxiv icon

Scaling Language-Image Pre-training via Masking

Dec 01, 2022
Yanghao Li, Haoqi Fan, Ronghang Hu, Christoph Feichtenhofer, Kaiming He

Figure 1 for Scaling Language-Image Pre-training via Masking
Figure 2 for Scaling Language-Image Pre-training via Masking
Figure 3 for Scaling Language-Image Pre-training via Masking
Figure 4 for Scaling Language-Image Pre-training via Masking
Viaarxiv icon