Alert button
Picture for Lei Ji

Lei Ji

Alert button

An Efficient COarse-to-fiNE Alignment Framework @ Ego4D Natural Language Queries Challenge 2022

Add code
Bookmark button
Alert button
Nov 16, 2022
Zhijian Hou, Wanjun Zhong, Lei Ji, Difei Gao, Kun Yan, Wing-Kwong Chan, Chong-Wah Ngo, Zheng Shou, Nan Duan

Figure 1 for An Efficient COarse-to-fiNE Alignment Framework @ Ego4D Natural Language Queries Challenge 2022
Figure 2 for An Efficient COarse-to-fiNE Alignment Framework @ Ego4D Natural Language Queries Challenge 2022
Figure 3 for An Efficient COarse-to-fiNE Alignment Framework @ Ego4D Natural Language Queries Challenge 2022
Figure 4 for An Efficient COarse-to-fiNE Alignment Framework @ Ego4D Natural Language Queries Challenge 2022
Viaarxiv icon

HORIZON: A High-Resolution Panorama Synthesis Framework

Add code
Bookmark button
Alert button
Oct 10, 2022
Kun Yan, Lei Ji, Chenfei Wu, Jian Liang, Ming Zhou, Nan Duan, Shuai Ma

Figure 1 for HORIZON: A High-Resolution Panorama Synthesis Framework
Figure 2 for HORIZON: A High-Resolution Panorama Synthesis Framework
Figure 3 for HORIZON: A High-Resolution Panorama Synthesis Framework
Figure 4 for HORIZON: A High-Resolution Panorama Synthesis Framework
Viaarxiv icon

CONE: An Efficient COarse-to-fiNE Alignment Framework for Long Video Temporal Grounding

Add code
Bookmark button
Alert button
Sep 22, 2022
Zhijian Hou, Wanjun Zhong, Lei Ji, Difei Gao, Kun Yan, Wing-Kwong Chan, Chong-Wah Ngo, Zheng Shou, Nan Duan

Figure 1 for CONE: An Efficient COarse-to-fiNE Alignment Framework for Long Video Temporal Grounding
Figure 2 for CONE: An Efficient COarse-to-fiNE Alignment Framework for Long Video Temporal Grounding
Figure 3 for CONE: An Efficient COarse-to-fiNE Alignment Framework for Long Video Temporal Grounding
Figure 4 for CONE: An Efficient COarse-to-fiNE Alignment Framework for Long Video Temporal Grounding
Viaarxiv icon

ScaleVLAD: Improving Multimodal Sentiment Analysis via Multi-Scale Fusion of Locally Descriptors

Add code
Bookmark button
Alert button
Dec 02, 2021
Huaishao Luo, Lei Ji, Yanyong Huang, Bin Wang, Shenggong Ji, Tianrui Li

Figure 1 for ScaleVLAD: Improving Multimodal Sentiment Analysis via Multi-Scale Fusion of Locally Descriptors
Figure 2 for ScaleVLAD: Improving Multimodal Sentiment Analysis via Multi-Scale Fusion of Locally Descriptors
Figure 3 for ScaleVLAD: Improving Multimodal Sentiment Analysis via Multi-Scale Fusion of Locally Descriptors
Figure 4 for ScaleVLAD: Improving Multimodal Sentiment Analysis via Multi-Scale Fusion of Locally Descriptors
Viaarxiv icon

NÜWA: Visual Synthesis Pre-training for Neural visUal World creAtion

Add code
Bookmark button
Alert button
Nov 24, 2021
Chenfei Wu, Jian Liang, Lei Ji, Fan Yang, Yuejian Fang, Daxin Jiang, Nan Duan

Figure 1 for NÜWA: Visual Synthesis Pre-training for Neural visUal World creAtion
Figure 2 for NÜWA: Visual Synthesis Pre-training for Neural visUal World creAtion
Figure 3 for NÜWA: Visual Synthesis Pre-training for Neural visUal World creAtion
Figure 4 for NÜWA: Visual Synthesis Pre-training for Neural visUal World creAtion
Viaarxiv icon

Hybrid Reasoning Network for Video-based Commonsense Captioning

Add code
Bookmark button
Alert button
Aug 05, 2021
Weijiang Yu, Jian Liang, Lei Ji, Lu Li, Yuejian Fang, Nong Xiao, Nan Duan

Viaarxiv icon

GEM: A General Evaluation Benchmark for Multimodal Tasks

Add code
Bookmark button
Alert button
Jun 18, 2021
Lin Su, Nan Duan, Edward Cui, Lei Ji, Chenfei Wu, Huaishao Luo, Yongfei Liu, Ming Zhong, Taroon Bharti, Arun Sacheti

Figure 1 for GEM: A General Evaluation Benchmark for Multimodal Tasks
Figure 2 for GEM: A General Evaluation Benchmark for Multimodal Tasks
Figure 3 for GEM: A General Evaluation Benchmark for Multimodal Tasks
Figure 4 for GEM: A General Evaluation Benchmark for Multimodal Tasks
Viaarxiv icon

CLIP4Clip: An Empirical Study of CLIP for End to End Video Clip Retrieval

Add code
Bookmark button
Alert button
May 08, 2021
Huaishao Luo, Lei Ji, Ming Zhong, Yang Chen, Wen Lei, Nan Duan, Tianrui Li

Figure 1 for CLIP4Clip: An Empirical Study of CLIP for End to End Video Clip Retrieval
Figure 2 for CLIP4Clip: An Empirical Study of CLIP for End to End Video Clip Retrieval
Figure 3 for CLIP4Clip: An Empirical Study of CLIP for End to End Video Clip Retrieval
Figure 4 for CLIP4Clip: An Empirical Study of CLIP for End to End Video Clip Retrieval
Viaarxiv icon

GODIVA: Generating Open-DomaIn Videos from nAtural Descriptions

Add code
Bookmark button
Alert button
Apr 30, 2021
Chenfei Wu, Lun Huang, Qianxi Zhang, Binyang Li, Lei Ji, Fan Yang, Guillermo Sapiro, Nan Duan

Figure 1 for GODIVA: Generating Open-DomaIn Videos from nAtural Descriptions
Figure 2 for GODIVA: Generating Open-DomaIn Videos from nAtural Descriptions
Figure 3 for GODIVA: Generating Open-DomaIn Videos from nAtural Descriptions
Figure 4 for GODIVA: Generating Open-DomaIn Videos from nAtural Descriptions
Viaarxiv icon