Alert button
Picture for Yifeng Liu

Yifeng Liu

Alert button

UniRel: Unified Representation and Interaction for Joint Relational Triple Extraction

Nov 16, 2022
Wei Tang, Benfeng Xu, Yuyue Zhao, Zhendong Mao, Yifeng Liu, Yong Liao, Haiyong Xie

Figure 1 for UniRel: Unified Representation and Interaction for Joint Relational Triple Extraction
Figure 2 for UniRel: Unified Representation and Interaction for Joint Relational Triple Extraction
Figure 3 for UniRel: Unified Representation and Interaction for Joint Relational Triple Extraction
Figure 4 for UniRel: Unified Representation and Interaction for Joint Relational Triple Extraction

Relational triple extraction is challenging for its difficulty in capturing rich correlations between entities and relations. Existing works suffer from 1) heterogeneous representations of entities and relations, and 2) heterogeneous modeling of entity-entity interactions and entity-relation interactions. Therefore, the rich correlations are not fully exploited by existing works. In this paper, we propose UniRel to address these challenges. Specifically, we unify the representations of entities and relations by jointly encoding them within a concatenated natural language sequence, and unify the modeling of interactions with a proposed Interaction Map, which is built upon the off-the-shelf self-attention mechanism within any Transformer block. With comprehensive experiments on two popular relational triple extraction datasets, we demonstrate that UniRel is more effective and computationally efficient. The source code is available at https://github.com/wtangdev/UniRel.

* Accepted at EMNLP 2022. Camera-ready version 
Viaarxiv icon

Progressive Multi-stage Interactive Training in Mobile Network for Fine-grained Recognition

Dec 08, 2021
Zhenxin Wu, Qingliang Chen, Yifeng Liu, Yinqi Zhang, Chengkai Zhu, Yang Yu

Figure 1 for Progressive Multi-stage Interactive Training in Mobile Network for Fine-grained Recognition
Figure 2 for Progressive Multi-stage Interactive Training in Mobile Network for Fine-grained Recognition
Figure 3 for Progressive Multi-stage Interactive Training in Mobile Network for Fine-grained Recognition
Figure 4 for Progressive Multi-stage Interactive Training in Mobile Network for Fine-grained Recognition

Fine-grained Visual Classification (FGVC) aims to identify objects from subcategories. It is a very challenging task because of the subtle inter-class differences. Existing research applies large-scale convolutional neural networks or visual transformers as the feature extractor, which is extremely computationally expensive. In fact, real-world scenarios of fine-grained recognition often require a more lightweight mobile network that can be utilized offline. However, the fundamental mobile network feature extraction capability is weaker than large-scale models. In this paper, based on the lightweight MobilenetV2, we propose a Progressive Multi-Stage Interactive training method with a Recursive Mosaic Generator (RMG-PMSI). First, we propose a Recursive Mosaic Generator (RMG) that generates images with different granularities in different phases. Then, the features of different stages pass through a Multi-Stage Interaction (MSI) module, which strengthens and complements the corresponding features of different stages. Finally, using the progressive training (P), the features extracted by the model in different stages can be fully utilized and fused with each other. Experiments on three prestigious fine-grained benchmarks show that RMG-PMSI can significantly improve the performance with good robustness and transferability.

Viaarxiv icon

Dilated Fully Convolutional Neural Network for Depth Estimation from a Single Image

Mar 12, 2021
Binghan Li, Yindong Hua, Yifeng Liu, Mi Lu

Figure 1 for Dilated Fully Convolutional Neural Network for Depth Estimation from a Single Image
Figure 2 for Dilated Fully Convolutional Neural Network for Depth Estimation from a Single Image
Figure 3 for Dilated Fully Convolutional Neural Network for Depth Estimation from a Single Image
Figure 4 for Dilated Fully Convolutional Neural Network for Depth Estimation from a Single Image

Depth prediction plays a key role in understanding a 3D scene. Several techniques have been developed throughout the years, among which Convolutional Neural Network has recently achieved state-of-the-art performance on estimating depth from a single image. However, traditional CNNs suffer from the lower resolution and information loss caused by the pooling layers. And oversized parameters generated from fully connected layers often lead to a exploded memory usage problem. In this paper, we present an advanced Dilated Fully Convolutional Neural Network to address the deficiencies. Taking advantages of the exponential expansion of the receptive field in dilated convolutions, our model can minimize the loss of resolution. It also reduces the amount of parameters significantly by replacing the fully connected layers with the fully convolutional layers. We show experimentally on NYU Depth V2 datasets that the depth prediction obtained from our model is considerably closer to ground truth than that from traditional CNNs techniques.

Viaarxiv icon