Alert button
Picture for Caren Han

Caren Han

Alert button

M3-VRD: Multimodal Multi-task Multi-teacher Visually-Rich Form Document Understanding

Add code
Bookmark button
Alert button
Feb 28, 2024
Yihao Ding, Lorenzo Vaiani, Caren Han, Jean Lee, Paolo Garza, Josiah Poon, Luca Cagliero

Viaarxiv icon

SceneGATE: Scene-Graph based co-Attention networks for TExt visual question answering

Add code
Bookmark button
Alert button
Dec 16, 2022
Siwen Luo, Feiqi Cao, Felipe Nunez, Zean Wen, Josiah Poon, Caren Han

Figure 1 for SceneGATE: Scene-Graph based co-Attention networks for TExt visual question answering
Figure 2 for SceneGATE: Scene-Graph based co-Attention networks for TExt visual question answering
Figure 3 for SceneGATE: Scene-Graph based co-Attention networks for TExt visual question answering
Figure 4 for SceneGATE: Scene-Graph based co-Attention networks for TExt visual question answering
Viaarxiv icon

An Analysis of Deep Reinforcement Learning Agents for Text-based Games

Add code
Bookmark button
Alert button
Sep 12, 2022
Chen Chen, Yue Dai, Josiah Poon, Caren Han

Figure 1 for An Analysis of Deep Reinforcement Learning Agents for Text-based Games
Figure 2 for An Analysis of Deep Reinforcement Learning Agents for Text-based Games
Figure 3 for An Analysis of Deep Reinforcement Learning Agents for Text-based Games
Figure 4 for An Analysis of Deep Reinforcement Learning Agents for Text-based Games
Viaarxiv icon

RoViST:Learning Robust Metrics for Visual Storytelling

Add code
Bookmark button
Alert button
May 08, 2022
Eileen Wang, Caren Han, Josiah Poon

Figure 1 for RoViST:Learning Robust Metrics for Visual Storytelling
Figure 2 for RoViST:Learning Robust Metrics for Visual Storytelling
Figure 3 for RoViST:Learning Robust Metrics for Visual Storytelling
Figure 4 for RoViST:Learning Robust Metrics for Visual Storytelling
Viaarxiv icon

Local Interpretations for Explainable Natural Language Processing: A Survey

Add code
Bookmark button
Alert button
Mar 20, 2021
Siwen Luo, Hamish Ivison, Caren Han, Josiah Poon

Figure 1 for Local Interpretations for Explainable Natural Language Processing: A Survey
Figure 2 for Local Interpretations for Explainable Natural Language Processing: A Survey
Figure 3 for Local Interpretations for Explainable Natural Language Processing: A Survey
Figure 4 for Local Interpretations for Explainable Natural Language Processing: A Survey
Viaarxiv icon

VICTR: Visual Information Captured Text Representation for Text-to-Image Multimodal Tasks

Add code
Bookmark button
Alert button
Oct 07, 2020
Caren Han, Siqu Long, Siwen Luo, Kunze Wang, Josiah Poon

Figure 1 for VICTR: Visual Information Captured Text Representation for Text-to-Image Multimodal Tasks
Figure 2 for VICTR: Visual Information Captured Text Representation for Text-to-Image Multimodal Tasks
Figure 3 for VICTR: Visual Information Captured Text Representation for Text-to-Image Multimodal Tasks
Figure 4 for VICTR: Visual Information Captured Text Representation for Text-to-Image Multimodal Tasks
Viaarxiv icon