Alert button
Picture for Junfeng Tian

Junfeng Tian

Alert button

RethinkingTMSC: An Empirical Study for Target-Oriented Multimodal Sentiment Classification

Add code
Bookmark button
Alert button
Oct 14, 2023
Junjie Ye, Jie Zhou, Junfeng Tian, Rui Wang, Qi Zhang, Tao Gui, Xuanjing Huang

Figure 1 for RethinkingTMSC: An Empirical Study for Target-Oriented Multimodal Sentiment Classification
Figure 2 for RethinkingTMSC: An Empirical Study for Target-Oriented Multimodal Sentiment Classification
Figure 3 for RethinkingTMSC: An Empirical Study for Target-Oriented Multimodal Sentiment Classification
Figure 4 for RethinkingTMSC: An Empirical Study for Target-Oriented Multimodal Sentiment Classification
Viaarxiv icon

UReader: Universal OCR-free Visually-situated Language Understanding with Multimodal Large Language Model

Add code
Bookmark button
Alert button
Oct 08, 2023
Jiabo Ye, Anwen Hu, Haiyang Xu, Qinghao Ye, Ming Yan, Guohai Xu, Chenliang Li, Junfeng Tian, Qi Qian, Ji Zhang, Qin Jin, Liang He, Xin Alex Lin, Fei Huang

Figure 1 for UReader: Universal OCR-free Visually-situated Language Understanding with Multimodal Large Language Model
Figure 2 for UReader: Universal OCR-free Visually-situated Language Understanding with Multimodal Large Language Model
Figure 3 for UReader: Universal OCR-free Visually-situated Language Understanding with Multimodal Large Language Model
Figure 4 for UReader: Universal OCR-free Visually-situated Language Understanding with Multimodal Large Language Model
Viaarxiv icon

mPLUG-DocOwl: Modularized Multimodal Large Language Model for Document Understanding

Add code
Bookmark button
Alert button
Jul 04, 2023
Jiabo Ye, Anwen Hu, Haiyang Xu, Qinghao Ye, Ming Yan, Yuhao Dan, Chenlin Zhao, Guohai Xu, Chenliang Li, Junfeng Tian, Qian Qi, Ji Zhang, Fei Huang

Figure 1 for mPLUG-DocOwl: Modularized Multimodal Large Language Model for Document Understanding
Figure 2 for mPLUG-DocOwl: Modularized Multimodal Large Language Model for Document Understanding
Figure 3 for mPLUG-DocOwl: Modularized Multimodal Large Language Model for Document Understanding
Figure 4 for mPLUG-DocOwl: Modularized Multimodal Large Language Model for Document Understanding
Viaarxiv icon

ChatPLUG: Open-Domain Generative Dialogue System with Internet-Augmented Instruction Tuning for Digital Human

Add code
Bookmark button
Alert button
Apr 28, 2023
Junfeng Tian, Hehong Chen, Guohai Xu, Ming Yan, Xing Gao, Jianhai Zhang, Chenliang Li, Jiayi Liu, Wenshen Xu, Haiyang Xu, Qi Qian, Wei Wang, Qinghao Ye, Jiejing Zhang, Ji Zhang, Fei Huang, Jingren Zhou

Figure 1 for ChatPLUG: Open-Domain Generative Dialogue System with Internet-Augmented Instruction Tuning for Digital Human
Figure 2 for ChatPLUG: Open-Domain Generative Dialogue System with Internet-Augmented Instruction Tuning for Digital Human
Figure 3 for ChatPLUG: Open-Domain Generative Dialogue System with Internet-Augmented Instruction Tuning for Digital Human
Figure 4 for ChatPLUG: Open-Domain Generative Dialogue System with Internet-Augmented Instruction Tuning for Digital Human
Viaarxiv icon

mPLUG-Owl: Modularization Empowers Large Language Models with Multimodality

Add code
Bookmark button
Alert button
Apr 27, 2023
Qinghao Ye, Haiyang Xu, Guohai Xu, Jiabo Ye, Ming Yan, Yiyang Zhou, Junyang Wang, Anwen Hu, Pengcheng Shi, Yaya Shi, Chenliang Li, Yuanhong Xu, Hehong Chen, Junfeng Tian, Qian Qi, Ji Zhang, Fei Huang

Figure 1 for mPLUG-Owl: Modularization Empowers Large Language Models with Multimodality
Figure 2 for mPLUG-Owl: Modularization Empowers Large Language Models with Multimodality
Figure 3 for mPLUG-Owl: Modularization Empowers Large Language Models with Multimodality
Figure 4 for mPLUG-Owl: Modularization Empowers Large Language Models with Multimodality
Viaarxiv icon

mPLUG: Effective and Efficient Vision-Language Learning by Cross-modal Skip-connections

Add code
Bookmark button
Alert button
May 25, 2022
Chenliang Li, Haiyang Xu, Junfeng Tian, Wei Wang, Ming Yan, Bin Bi, Jiabo Ye, Hehong Chen, Guohai Xu, Zheng Cao, Ji Zhang, Songfang Huang, Fei Huang, Jingren Zhou, Luo Si

Figure 1 for mPLUG: Effective and Efficient Vision-Language Learning by Cross-modal Skip-connections
Figure 2 for mPLUG: Effective and Efficient Vision-Language Learning by Cross-modal Skip-connections
Figure 3 for mPLUG: Effective and Efficient Vision-Language Learning by Cross-modal Skip-connections
Figure 4 for mPLUG: Effective and Efficient Vision-Language Learning by Cross-modal Skip-connections
Viaarxiv icon

WikiDiverse: A Multimodal Entity Linking Dataset with Diversified Contextual Topics and Entity Types

Add code
Bookmark button
Alert button
Apr 13, 2022
Xuwu Wang, Junfeng Tian, Min Gui, Zhixu Li, Rui Wang, Ming Yan, Lihan Chen, Yanghua Xiao

Figure 1 for WikiDiverse: A Multimodal Entity Linking Dataset with Diversified Contextual Topics and Entity Types
Figure 2 for WikiDiverse: A Multimodal Entity Linking Dataset with Diversified Contextual Topics and Entity Types
Figure 3 for WikiDiverse: A Multimodal Entity Linking Dataset with Diversified Contextual Topics and Entity Types
Figure 4 for WikiDiverse: A Multimodal Entity Linking Dataset with Diversified Contextual Topics and Entity Types
Viaarxiv icon

Shifting More Attention to Visual Backbone: Query-modulated Refinement Networks for End-to-End Visual Grounding

Add code
Bookmark button
Alert button
Mar 29, 2022
Jiabo Ye, Junfeng Tian, Ming Yan, Xiaoshan Yang, Xuwu Wang, Ji Zhang, Liang He, Xin Lin

Figure 1 for Shifting More Attention to Visual Backbone: Query-modulated Refinement Networks for End-to-End Visual Grounding
Figure 2 for Shifting More Attention to Visual Backbone: Query-modulated Refinement Networks for End-to-End Visual Grounding
Figure 3 for Shifting More Attention to Visual Backbone: Query-modulated Refinement Networks for End-to-End Visual Grounding
Figure 4 for Shifting More Attention to Visual Backbone: Query-modulated Refinement Networks for End-to-End Visual Grounding
Viaarxiv icon