Alert button
Picture for Peng Gao

Peng Gao

Alert button

ImageBind-LLM: Multi-modality Instruction Tuning

Add code
Bookmark button
Alert button
Sep 11, 2023
Jiaming Han, Renrui Zhang, Wenqi Shao, Peng Gao, Peng Xu, Han Xiao, Kaipeng Zhang, Chris Liu, Song Wen, Ziyu Guo, Xudong Lu, Shuai Ren, Yafei Wen, Xiaoxin Chen, Xiangyu Yue, Hongsheng Li, Yu Qiao

Figure 1 for ImageBind-LLM: Multi-modality Instruction Tuning
Figure 2 for ImageBind-LLM: Multi-modality Instruction Tuning
Figure 3 for ImageBind-LLM: Multi-modality Instruction Tuning
Figure 4 for ImageBind-LLM: Multi-modality Instruction Tuning
Viaarxiv icon

Point-Bind & Point-LLM: Aligning Point Cloud with Multi-modality for 3D Understanding, Generation, and Instruction Following

Add code
Bookmark button
Alert button
Sep 01, 2023
Ziyu Guo, Renrui Zhang, Xiangyang Zhu, Yiwen Tang, Xianzheng Ma, Jiaming Han, Kexin Chen, Peng Gao, Xianzhi Li, Hongsheng Li, Pheng-Ann Heng

Figure 1 for Point-Bind & Point-LLM: Aligning Point Cloud with Multi-modality for 3D Understanding, Generation, and Instruction Following
Figure 2 for Point-Bind & Point-LLM: Aligning Point Cloud with Multi-modality for 3D Understanding, Generation, and Instruction Following
Figure 3 for Point-Bind & Point-LLM: Aligning Point Cloud with Multi-modality for 3D Understanding, Generation, and Instruction Following
Figure 4 for Point-Bind & Point-LLM: Aligning Point Cloud with Multi-modality for 3D Understanding, Generation, and Instruction Following
Viaarxiv icon

OmniQuant: Omnidirectionally Calibrated Quantization for Large Language Models

Add code
Bookmark button
Alert button
Aug 25, 2023
Wenqi Shao, Mengzhao Chen, Zhaoyang Zhang, Peng Xu, Lirui Zhao, Zhiqian Li, Kaipeng Zhang, Peng Gao, Yu Qiao, Ping Luo

Figure 1 for OmniQuant: Omnidirectionally Calibrated Quantization for Large Language Models
Figure 2 for OmniQuant: Omnidirectionally Calibrated Quantization for Large Language Models
Figure 3 for OmniQuant: Omnidirectionally Calibrated Quantization for Large Language Models
Figure 4 for OmniQuant: Omnidirectionally Calibrated Quantization for Large Language Models
Viaarxiv icon

Less is More: Towards Efficient Few-shot 3D Semantic Segmentation via Training-free Networks

Add code
Bookmark button
Alert button
Aug 24, 2023
Xiangyang Zhu, Renrui Zhang, Bowei He, Ziyu Guo, Jiaming Liu, Hao Dong, Peng Gao

Viaarxiv icon

Tiny LVLM-eHub: Early Multimodal Experiments with Bard

Add code
Bookmark button
Alert button
Aug 07, 2023
Wenqi Shao, Yutao Hu, Peng Gao, Meng Lei, Kaipeng Zhang, Fanqing Meng, Peng Xu, Siyuan Huang, Hongsheng Li, Yu Qiao, Ping Luo

Figure 1 for Tiny LVLM-eHub: Early Multimodal Experiments with Bard
Figure 2 for Tiny LVLM-eHub: Early Multimodal Experiments with Bard
Figure 3 for Tiny LVLM-eHub: Early Multimodal Experiments with Bard
Figure 4 for Tiny LVLM-eHub: Early Multimodal Experiments with Bard
Viaarxiv icon

Enhanced Neural Beamformer with Spatial Information for Target Speech Extraction

Add code
Bookmark button
Alert button
Jun 28, 2023
Aoqi Guo, Junnan Wu, Peng Gao, Wenbo Zhu, Qinwen Guo, Dazhi Gao, Yujun Wang

Figure 1 for Enhanced Neural Beamformer with Spatial Information for Target Speech Extraction
Figure 2 for Enhanced Neural Beamformer with Spatial Information for Target Speech Extraction
Figure 3 for Enhanced Neural Beamformer with Spatial Information for Target Speech Extraction
Figure 4 for Enhanced Neural Beamformer with Spatial Information for Target Speech Extraction
Viaarxiv icon

LVLM-eHub: A Comprehensive Evaluation Benchmark for Large Vision-Language Models

Add code
Bookmark button
Alert button
Jun 15, 2023
Peng Xu, Wenqi Shao, Kaipeng Zhang, Peng Gao, Shuo Liu, Meng Lei, Fanqing Meng, Siyuan Huang, Yu Qiao, Ping Luo

Figure 1 for LVLM-eHub: A Comprehensive Evaluation Benchmark for Large Vision-Language Models
Figure 2 for LVLM-eHub: A Comprehensive Evaluation Benchmark for Large Vision-Language Models
Figure 3 for LVLM-eHub: A Comprehensive Evaluation Benchmark for Large Vision-Language Models
Figure 4 for LVLM-eHub: A Comprehensive Evaluation Benchmark for Large Vision-Language Models
Viaarxiv icon

Referred by Multi-Modality: A Unified Temporal Transformer for Video Object Segmentation

Add code
Bookmark button
Alert button
May 25, 2023
Shilin Yan, Renrui Zhang, Ziyu Guo, Wenchao Chen, Wei Zhang, Hongyang Li, Yu Qiao, Zhongjiang He, Peng Gao

Figure 1 for Referred by Multi-Modality: A Unified Temporal Transformer for Video Object Segmentation
Figure 2 for Referred by Multi-Modality: A Unified Temporal Transformer for Video Object Segmentation
Figure 3 for Referred by Multi-Modality: A Unified Temporal Transformer for Video Object Segmentation
Figure 4 for Referred by Multi-Modality: A Unified Temporal Transformer for Video Object Segmentation
Viaarxiv icon

Instruct2Act: Mapping Multi-modality Instructions to Robotic Actions with Large Language Model

Add code
Bookmark button
Alert button
May 24, 2023
Siyuan Huang, Zhengkai Jiang, Hao Dong, Yu Qiao, Peng Gao, Hongsheng Li

Figure 1 for Instruct2Act: Mapping Multi-modality Instructions to Robotic Actions with Large Language Model
Figure 2 for Instruct2Act: Mapping Multi-modality Instructions to Robotic Actions with Large Language Model
Figure 3 for Instruct2Act: Mapping Multi-modality Instructions to Robotic Actions with Large Language Model
Figure 4 for Instruct2Act: Mapping Multi-modality Instructions to Robotic Actions with Large Language Model
Viaarxiv icon