Picture for Guoxin Wang

Guoxin Wang

Multi-Dimension-Embedding-Aware Modality Fusion Transformer for Psychiatric Disorder Clasification

Add code
Oct 04, 2023
Viaarxiv icon

Kosmos-2.5: A Multimodal Literate Model

Add code
Sep 20, 2023
Figure 1 for Kosmos-2.5: A Multimodal Literate Model
Figure 2 for Kosmos-2.5: A Multimodal Literate Model
Figure 3 for Kosmos-2.5: A Multimodal Literate Model
Figure 4 for Kosmos-2.5: A Multimodal Literate Model
Viaarxiv icon

Unifying Vision, Text, and Layout for Universal Document Processing

Add code
Dec 20, 2022
Viaarxiv icon

Understanding Long Documents with Different Position-Aware Attentions

Add code
Aug 17, 2022
Figure 1 for Understanding Long Documents with Different Position-Aware Attentions
Figure 2 for Understanding Long Documents with Different Position-Aware Attentions
Figure 3 for Understanding Long Documents with Different Position-Aware Attentions
Figure 4 for Understanding Long Documents with Different Position-Aware Attentions
Viaarxiv icon

BoningKnife: Joint Entity Mention Detection and Typing for Nested NER via prior Boundary Knowledge

Add code
Jul 20, 2021
Figure 1 for BoningKnife: Joint Entity Mention Detection and Typing for Nested NER via prior Boundary Knowledge
Figure 2 for BoningKnife: Joint Entity Mention Detection and Typing for Nested NER via prior Boundary Knowledge
Figure 3 for BoningKnife: Joint Entity Mention Detection and Typing for Nested NER via prior Boundary Knowledge
Figure 4 for BoningKnife: Joint Entity Mention Detection and Typing for Nested NER via prior Boundary Knowledge
Viaarxiv icon

Understanding Chinese Video and Language via Contrastive Multimodal Pre-Training

Add code
Apr 19, 2021
Figure 1 for Understanding Chinese Video and Language via Contrastive Multimodal Pre-Training
Figure 2 for Understanding Chinese Video and Language via Contrastive Multimodal Pre-Training
Figure 3 for Understanding Chinese Video and Language via Contrastive Multimodal Pre-Training
Figure 4 for Understanding Chinese Video and Language via Contrastive Multimodal Pre-Training
Viaarxiv icon

LayoutXLM: Multimodal Pre-training for Multilingual Visually-rich Document Understanding

Add code
Apr 18, 2021
Figure 1 for LayoutXLM: Multimodal Pre-training for Multilingual Visually-rich Document Understanding
Figure 2 for LayoutXLM: Multimodal Pre-training for Multilingual Visually-rich Document Understanding
Figure 3 for LayoutXLM: Multimodal Pre-training for Multilingual Visually-rich Document Understanding
Figure 4 for LayoutXLM: Multimodal Pre-training for Multilingual Visually-rich Document Understanding
Viaarxiv icon

LayoutLMv2: Multi-modal Pre-training for Visually-Rich Document Understanding

Add code
Dec 29, 2020
Figure 1 for LayoutLMv2: Multi-modal Pre-training for Visually-Rich Document Understanding
Figure 2 for LayoutLMv2: Multi-modal Pre-training for Visually-Rich Document Understanding
Figure 3 for LayoutLMv2: Multi-modal Pre-training for Visually-Rich Document Understanding
Figure 4 for LayoutLMv2: Multi-modal Pre-training for Visually-Rich Document Understanding
Viaarxiv icon

A Pre-training Strategy for Recommendation

Add code
Oct 23, 2020
Figure 1 for A Pre-training Strategy for Recommendation
Figure 2 for A Pre-training Strategy for Recommendation
Figure 3 for A Pre-training Strategy for Recommendation
Figure 4 for A Pre-training Strategy for Recommendation
Viaarxiv icon

Enhanced Meta-Learning for Cross-lingual Named Entity Recognition with Minimal Resources

Add code
Nov 14, 2019
Figure 1 for Enhanced Meta-Learning for Cross-lingual Named Entity Recognition with Minimal Resources
Figure 2 for Enhanced Meta-Learning for Cross-lingual Named Entity Recognition with Minimal Resources
Figure 3 for Enhanced Meta-Learning for Cross-lingual Named Entity Recognition with Minimal Resources
Figure 4 for Enhanced Meta-Learning for Cross-lingual Named Entity Recognition with Minimal Resources
Viaarxiv icon