Alert button
Picture for Joya Chen

Joya Chen

Alert button

Bootstrapping SparseFormers from Vision Foundation Models

Add code
Bookmark button
Alert button
Dec 04, 2023
Ziteng Gao, Zhan Tong, Kevin Qinghong Lin, Joya Chen, Mike Zheng Shou

Viaarxiv icon

Ego-Exo4D: Understanding Skilled Human Activity from First- and Third-Person Perspectives

Add code
Bookmark button
Alert button
Nov 30, 2023
Kristen Grauman, Andrew Westbury, Lorenzo Torresani, Kris Kitani, Jitendra Malik, Triantafyllos Afouras, Kumar Ashutosh, Vijay Baiyya, Siddhant Bansal, Bikram Boote, Eugene Byrne, Zach Chavis, Joya Chen, Feng Cheng, Fu-Jen Chu, Sean Crane, Avijit Dasgupta, Jing Dong, Maria Escobar, Cristhian Forigua, Abrham Gebreselasie, Sanjay Haresh, Jing Huang, Md Mohaiminul Islam, Suyog Jain, Rawal Khirodkar, Devansh Kukreja, Kevin J Liang, Jia-Wei Liu, Sagnik Majumder, Yongsen Mao, Miguel Martin, Effrosyni Mavroudi, Tushar Nagarajan, Francesco Ragusa, Santhosh Kumar Ramakrishnan, Luigi Seminara, Arjun Somayazulu, Yale Song, Shan Su, Zihui Xue, Edward Zhang, Jinxu Zhang, Angela Castillo, Changan Chen, Xinzhu Fu, Ryosuke Furuta, Cristina Gonzalez, Prince Gupta, Jiabo Hu, Yifei Huang, Yiming Huang, Weslie Khoo, Anush Kumar, Robert Kuo, Sach Lakhavani, Miao Liu, Mi Luo, Zhengyi Luo, Brighid Meredith, Austin Miller, Oluwatumininu Oguntola, Xiaqing Pan, Penny Peng, Shraman Pramanick, Merey Ramazanova, Fiona Ryan, Wei Shan, Kiran Somasundaram, Chenan Song, Audrey Southerland, Masatoshi Tateno, Huiyu Wang, Yuchen Wang, Takuma Yagi, Mingfei Yan, Xitong Yang, Zecheng Yu, Shengxin Cindy Zha, Chen Zhao, Ziwei Zhao, Zhifan Zhu, Jeff Zhuo, Pablo Arbelaez, Gedas Bertasius, David Crandall, Dima Damen, Jakob Engel, Giovanni Maria Farinella, Antonino Furnari, Bernard Ghanem, Judy Hoffman, C. V. Jawahar, Richard Newcombe, Hyun Soo Park, James M. Rehg, Yoichi Sato, Manolis Savva, Jianbo Shi, Mike Zheng Shou, Michael Wray

Figure 1 for Ego-Exo4D: Understanding Skilled Human Activity from First- and Third-Person Perspectives
Figure 2 for Ego-Exo4D: Understanding Skilled Human Activity from First- and Third-Person Perspectives
Figure 3 for Ego-Exo4D: Understanding Skilled Human Activity from First- and Third-Person Perspectives
Figure 4 for Ego-Exo4D: Understanding Skilled Human Activity from First- and Third-Person Perspectives
Viaarxiv icon

UniVTG: Towards Unified Video-Language Temporal Grounding

Add code
Bookmark button
Alert button
Aug 18, 2023
Kevin Qinghong Lin, Pengchuan Zhang, Joya Chen, Shraman Pramanick, Difei Gao, Alex Jinpeng Wang, Rui Yan, Mike Zheng Shou

Figure 1 for UniVTG: Towards Unified Video-Language Temporal Grounding
Figure 2 for UniVTG: Towards Unified Video-Language Temporal Grounding
Figure 3 for UniVTG: Towards Unified Video-Language Temporal Grounding
Figure 4 for UniVTG: Towards Unified Video-Language Temporal Grounding
Viaarxiv icon

AssistGPT: A General Multi-modal Assistant that can Plan, Execute, Inspect, and Learn

Add code
Bookmark button
Alert button
Jun 28, 2023
Difei Gao, Lei Ji, Luowei Zhou, Kevin Qinghong Lin, Joya Chen, Zihan Fan, Mike Zheng Shou

Figure 1 for AssistGPT: A General Multi-modal Assistant that can Plan, Execute, Inspect, and Learn
Figure 2 for AssistGPT: A General Multi-modal Assistant that can Plan, Execute, Inspect, and Learn
Figure 3 for AssistGPT: A General Multi-modal Assistant that can Plan, Execute, Inspect, and Learn
Figure 4 for AssistGPT: A General Multi-modal Assistant that can Plan, Execute, Inspect, and Learn
Viaarxiv icon

Affordance Grounding from Demonstration Video to Target Image

Add code
Bookmark button
Alert button
Mar 26, 2023
Joya Chen, Difei Gao, Kevin Qinghong Lin, Mike Zheng Shou

Figure 1 for Affordance Grounding from Demonstration Video to Target Image
Figure 2 for Affordance Grounding from Demonstration Video to Target Image
Figure 3 for Affordance Grounding from Demonstration Video to Target Image
Figure 4 for Affordance Grounding from Demonstration Video to Target Image
Viaarxiv icon

AssistQ: Affordance-centric Question-driven Task Completion for Egocentric Assistant

Add code
Bookmark button
Alert button
Mar 08, 2022
Benita Wong, Joya Chen, You Wu, Stan Weixian Lei, Dongxing Mao, Difei Gao, Mike Zheng Shou

Figure 1 for AssistQ: Affordance-centric Question-driven Task Completion for Egocentric Assistant
Figure 2 for AssistQ: Affordance-centric Question-driven Task Completion for Egocentric Assistant
Figure 3 for AssistQ: Affordance-centric Question-driven Task Completion for Egocentric Assistant
Figure 4 for AssistQ: Affordance-centric Question-driven Task Completion for Egocentric Assistant
Viaarxiv icon

DropIT: Dropping Intermediate Tensors for Memory-Efficient DNN Training

Add code
Bookmark button
Alert button
Feb 28, 2022
Joya Chen, Kai Xu, Yifei Cheng, Angela Yao

Figure 1 for DropIT: Dropping Intermediate Tensors for Memory-Efficient DNN Training
Figure 2 for DropIT: Dropping Intermediate Tensors for Memory-Efficient DNN Training
Figure 3 for DropIT: Dropping Intermediate Tensors for Memory-Efficient DNN Training
Figure 4 for DropIT: Dropping Intermediate Tensors for Memory-Efficient DNN Training
Viaarxiv icon

Foreground-Background Imbalance Problem in Deep Object Detectors: A Review

Add code
Bookmark button
Alert button
Jun 16, 2020
Joya Chen, Qi Wu, Dong Liu, Tong Xu

Figure 1 for Foreground-Background Imbalance Problem in Deep Object Detectors: A Review
Figure 2 for Foreground-Background Imbalance Problem in Deep Object Detectors: A Review
Viaarxiv icon