Alert button
Picture for Yu Shi

Yu Shi

Alert button

Florence: A New Foundation Model for Computer Vision

Add code
Bookmark button
Alert button
Nov 22, 2021
Lu Yuan, Dongdong Chen, Yi-Ling Chen, Noel Codella, Xiyang Dai, Jianfeng Gao, Houdong Hu, Xuedong Huang, Boxin Li, Chunyuan Li, Ce Liu, Mengchen Liu, Zicheng Liu, Yumao Lu, Yu Shi, Lijuan Wang, Jianfeng Wang, Bin Xiao, Zhen Xiao, Jianwei Yang, Michael Zeng, Luowei Zhou, Pengchuan Zhang

Figure 1 for Florence: A New Foundation Model for Computer Vision
Figure 2 for Florence: A New Foundation Model for Computer Vision
Figure 3 for Florence: A New Foundation Model for Computer Vision
Figure 4 for Florence: A New Foundation Model for Computer Vision
Viaarxiv icon

Optimizing Alignment of Speech and Language Latent Spaces for End-to-End Speech Recognition and Understanding

Add code
Bookmark button
Alert button
Oct 23, 2021
Wei Wang, Shuo Ren, Yao Qian, Shujie Liu, Yu Shi, Yanmin Qian, Michael Zeng

Figure 1 for Optimizing Alignment of Speech and Language Latent Spaces for End-to-End Speech Recognition and Understanding
Figure 2 for Optimizing Alignment of Speech and Language Latent Spaces for End-to-End Speech Recognition and Understanding
Figure 3 for Optimizing Alignment of Speech and Language Latent Spaces for End-to-End Speech Recognition and Understanding
Figure 4 for Optimizing Alignment of Speech and Language Latent Spaces for End-to-End Speech Recognition and Understanding
Viaarxiv icon

Temporal Pyramid Transformer with Multimodal Interaction for Video Question Answering

Add code
Bookmark button
Alert button
Sep 10, 2021
Min Peng, Chongyang Wang, Yuan Gao, Yu Shi, Xiang-Dong Zhou

Figure 1 for Temporal Pyramid Transformer with Multimodal Interaction for Video Question Answering
Figure 2 for Temporal Pyramid Transformer with Multimodal Interaction for Video Question Answering
Figure 3 for Temporal Pyramid Transformer with Multimodal Interaction for Video Question Answering
Figure 4 for Temporal Pyramid Transformer with Multimodal Interaction for Video Question Answering
Viaarxiv icon

A Joint and Domain-Adaptive Approach to Spoken Language Understanding

Add code
Bookmark button
Alert button
Jul 25, 2021
Linhao Zhang, Yu Shi, Linjun Shou, Ming Gong, Houfeng Wang, Michael Zeng

Figure 1 for A Joint and Domain-Adaptive Approach to Spoken Language Understanding
Figure 2 for A Joint and Domain-Adaptive Approach to Spoken Language Understanding
Figure 3 for A Joint and Domain-Adaptive Approach to Spoken Language Understanding
Figure 4 for A Joint and Domain-Adaptive Approach to Spoken Language Understanding
Viaarxiv icon

Transformer-F: A Transformer network with effective methods for learning universal sentence representation

Add code
Bookmark button
Alert button
Jul 02, 2021
Yu Shi

Figure 1 for Transformer-F: A Transformer network with effective methods for learning universal sentence representation
Figure 2 for Transformer-F: A Transformer network with effective methods for learning universal sentence representation
Figure 3 for Transformer-F: A Transformer network with effective methods for learning universal sentence representation
Figure 4 for Transformer-F: A Transformer network with effective methods for learning universal sentence representation
Viaarxiv icon

Generating Human Readable Transcript for Automatic Speech Recognition with Pre-trained Language Model

Add code
Bookmark button
Alert button
Feb 22, 2021
Junwei Liao, Yu Shi, Ming Gong, Linjun Shou, Sefik Eskimez, Liyang Lu, Hong Qu, Michael Zeng

Figure 1 for Generating Human Readable Transcript for Automatic Speech Recognition with Pre-trained Language Model
Figure 2 for Generating Human Readable Transcript for Automatic Speech Recognition with Pre-trained Language Model
Figure 3 for Generating Human Readable Transcript for Automatic Speech Recognition with Pre-trained Language Model
Viaarxiv icon

Improving Zero-shot Neural Machine Translation on Language-specific Encoders-Decoders

Add code
Bookmark button
Alert button
Feb 12, 2021
Junwei Liao, Yu Shi, Ming Gong, Linjun Shou, Hong Qu, Michael Zeng

Figure 1 for Improving Zero-shot Neural Machine Translation on Language-specific Encoders-Decoders
Figure 2 for Improving Zero-shot Neural Machine Translation on Language-specific Encoders-Decoders
Figure 3 for Improving Zero-shot Neural Machine Translation on Language-specific Encoders-Decoders
Figure 4 for Improving Zero-shot Neural Machine Translation on Language-specific Encoders-Decoders
Viaarxiv icon

Speech-language Pre-training for End-to-end Spoken Language Understanding

Add code
Bookmark button
Alert button
Feb 11, 2021
Yao Qian, Ximo Bian, Yu Shi, Naoyuki Kanda, Leo Shen, Zhen Xiao, Michael Zeng

Figure 1 for Speech-language Pre-training for End-to-end Spoken Language Understanding
Figure 2 for Speech-language Pre-training for End-to-end Spoken Language Understanding
Figure 3 for Speech-language Pre-training for End-to-end Spoken Language Understanding
Figure 4 for Speech-language Pre-training for End-to-end Spoken Language Understanding
Viaarxiv icon

Mixed-Lingual Pre-training for Cross-lingual Summarization

Add code
Bookmark button
Alert button
Oct 18, 2020
Ruochen Xu, Chenguang Zhu, Yu Shi, Michael Zeng, Xuedong Huang

Figure 1 for Mixed-Lingual Pre-training for Cross-lingual Summarization
Figure 2 for Mixed-Lingual Pre-training for Cross-lingual Summarization
Figure 3 for Mixed-Lingual Pre-training for Cross-lingual Summarization
Figure 4 for Mixed-Lingual Pre-training for Cross-lingual Summarization
Viaarxiv icon

MaP: A Matrix-based Prediction Approach to Improve Span Extraction in Machine Reading Comprehension

Add code
Bookmark button
Alert button
Sep 29, 2020
Huaishao Luo, Yu Shi, Ming Gong, Linjun Shou, Tianrui Li

Figure 1 for MaP: A Matrix-based Prediction Approach to Improve Span Extraction in Machine Reading Comprehension
Figure 2 for MaP: A Matrix-based Prediction Approach to Improve Span Extraction in Machine Reading Comprehension
Figure 3 for MaP: A Matrix-based Prediction Approach to Improve Span Extraction in Machine Reading Comprehension
Figure 4 for MaP: A Matrix-based Prediction Approach to Improve Span Extraction in Machine Reading Comprehension
Viaarxiv icon