Alert button
Picture for Ping Luo

Ping Luo

Alert button

LVLM-eHub: A Comprehensive Evaluation Benchmark for Large Vision-Language Models

Add code
Bookmark button
Alert button
Jun 15, 2023
Peng Xu, Wenqi Shao, Kaipeng Zhang, Peng Gao, Shuo Liu, Meng Lei, Fanqing Meng, Siyuan Huang, Yu Qiao, Ping Luo

Figure 1 for LVLM-eHub: A Comprehensive Evaluation Benchmark for Large Vision-Language Models
Figure 2 for LVLM-eHub: A Comprehensive Evaluation Benchmark for Large Vision-Language Models
Figure 3 for LVLM-eHub: A Comprehensive Evaluation Benchmark for Large Vision-Language Models
Figure 4 for LVLM-eHub: A Comprehensive Evaluation Benchmark for Large Vision-Language Models
Viaarxiv icon

Scene as Occupancy

Add code
Bookmark button
Alert button
Jun 06, 2023
Wenwen Tong, Chonghao Sima, Tai Wang, Silei Wu, Hanming Deng, Li Chen, Yi Gu, Lewei Lu, Ping Luo, Dahua Lin, Hongyang Li

Figure 1 for Scene as Occupancy
Figure 2 for Scene as Occupancy
Figure 3 for Scene as Occupancy
Figure 4 for Scene as Occupancy
Viaarxiv icon

RAPHAEL: Text-to-Image Generation via Large Mixture of Diffusion Paths

Add code
Bookmark button
Alert button
May 29, 2023
Zeyue Xue, Guanglu Song, Qiushan Guo, Boxiao Liu, Zhuofan Zong, Yu Liu, Ping Luo

Figure 1 for RAPHAEL: Text-to-Image Generation via Large Mixture of Diffusion Paths
Figure 2 for RAPHAEL: Text-to-Image Generation via Large Mixture of Diffusion Paths
Figure 3 for RAPHAEL: Text-to-Image Generation via Large Mixture of Diffusion Paths
Figure 4 for RAPHAEL: Text-to-Image Generation via Large Mixture of Diffusion Paths
Viaarxiv icon

DiffRate : Differentiable Compression Rate for Efficient Vision Transformers

Add code
Bookmark button
Alert button
May 29, 2023
Mengzhao Chen, Wenqi Shao, Peng Xu, Mingbao Lin, Kaipeng Zhang, Fei Chao, Rongrong Ji, Yu Qiao, Ping Luo

Figure 1 for DiffRate : Differentiable Compression Rate for Efficient Vision Transformers
Figure 2 for DiffRate : Differentiable Compression Rate for Efficient Vision Transformers
Figure 3 for DiffRate : Differentiable Compression Rate for Efficient Vision Transformers
Figure 4 for DiffRate : Differentiable Compression Rate for Efficient Vision Transformers
Viaarxiv icon

VisionLLM: Large Language Model is also an Open-Ended Decoder for Vision-Centric Tasks

Add code
Bookmark button
Alert button
May 25, 2023
Wenhai Wang, Zhe Chen, Xiaokang Chen, Jiannan Wu, Xizhou Zhu, Gang Zeng, Ping Luo, Tong Lu, Jie Zhou, Yu Qiao, Jifeng Dai

Figure 1 for VisionLLM: Large Language Model is also an Open-Ended Decoder for Vision-Centric Tasks
Figure 2 for VisionLLM: Large Language Model is also an Open-Ended Decoder for Vision-Centric Tasks
Figure 3 for VisionLLM: Large Language Model is also an Open-Ended Decoder for Vision-Centric Tasks
Figure 4 for VisionLLM: Large Language Model is also an Open-Ended Decoder for Vision-Centric Tasks
Viaarxiv icon

SyNDock: N Rigid Protein Docking via Learnable Group Synchronization

Add code
Bookmark button
Alert button
May 25, 2023
Yuanfeng Ji, Yatao Bian, Guoji Fu, Peilin Zhao, Ping Luo

Figure 1 for SyNDock: N Rigid Protein Docking via Learnable Group Synchronization
Figure 2 for SyNDock: N Rigid Protein Docking via Learnable Group Synchronization
Figure 3 for SyNDock: N Rigid Protein Docking via Learnable Group Synchronization
Figure 4 for SyNDock: N Rigid Protein Docking via Learnable Group Synchronization
Viaarxiv icon

EmbodiedGPT: Vision-Language Pre-Training via Embodied Chain of Thought

Add code
Bookmark button
Alert button
May 24, 2023
Yao Mu, Qinglong Zhang, Mengkang Hu, Wenhai Wang, Mingyu Ding, Jun Jin, Bin Wang, Jifeng Dai, Yu Qiao, Ping Luo

Figure 1 for EmbodiedGPT: Vision-Language Pre-Training via Embodied Chain of Thought
Figure 2 for EmbodiedGPT: Vision-Language Pre-Training via Embodied Chain of Thought
Figure 3 for EmbodiedGPT: Vision-Language Pre-Training via Embodied Chain of Thought
Figure 4 for EmbodiedGPT: Vision-Language Pre-Training via Embodied Chain of Thought
Viaarxiv icon

VDT: An Empirical Study on Video Diffusion with Transformers

Add code
Bookmark button
Alert button
May 22, 2023
Haoyu Lu, Guoxing Yang, Nanyi Fei, Yuqi Huo, Zhiwu Lu, Ping Luo, Mingyu Ding

Figure 1 for VDT: An Empirical Study on Video Diffusion with Transformers
Figure 2 for VDT: An Empirical Study on Video Diffusion with Transformers
Figure 3 for VDT: An Empirical Study on Video Diffusion with Transformers
Figure 4 for VDT: An Empirical Study on Video Diffusion with Transformers
Viaarxiv icon