Alert button
Picture for Qianying Liu

Qianying Liu

Alert button

Multi-Scale Cross Contrastive Learning for Semi-Supervised Medical Image Segmentation

Jun 25, 2023
Qianying Liu, Xiao Gu, Paul Henderson, Fani Deligianni

Figure 1 for Multi-Scale Cross Contrastive Learning for Semi-Supervised Medical Image Segmentation
Figure 2 for Multi-Scale Cross Contrastive Learning for Semi-Supervised Medical Image Segmentation
Figure 3 for Multi-Scale Cross Contrastive Learning for Semi-Supervised Medical Image Segmentation
Figure 4 for Multi-Scale Cross Contrastive Learning for Semi-Supervised Medical Image Segmentation

Semi-supervised learning has demonstrated great potential in medical image segmentation by utilizing knowledge from unlabeled data. However, most existing approaches do not explicitly capture high-level semantic relations between distant regions, which limits their performance. In this paper, we focus on representation learning for semi-supervised learning, by developing a novel Multi-Scale Cross Supervised Contrastive Learning (MCSC) framework, to segment structures in medical images. We jointly train CNN and Transformer models, regularising their features to be semantically consistent across different scales. Our approach contrasts multi-scale features based on ground-truth and cross-predicted labels, in order to extract robust feature representations that reflect intra- and inter-slice relationships across the whole dataset. To tackle class imbalance, we take into account the prevalence of each class to guide contrastive learning and ensure that features adequately capture infrequent classes. Extensive experiments on two multi-structure medical segmentation datasets demonstrate the effectiveness of MCSC. It not only outperforms state-of-the-art semi-supervised methods by more than 3.0% in Dice, but also greatly reduces the performance gap with fully supervised methods.

Viaarxiv icon

Exploring the Impact of Layer Normalization for Zero-shot Neural Machine Translation

May 16, 2023
Zhuoyuan Mao, Raj Dabre, Qianying Liu, Haiyue Song, Chenhui Chu, Sadao Kurohashi

Figure 1 for Exploring the Impact of Layer Normalization for Zero-shot Neural Machine Translation
Figure 2 for Exploring the Impact of Layer Normalization for Zero-shot Neural Machine Translation
Figure 3 for Exploring the Impact of Layer Normalization for Zero-shot Neural Machine Translation
Figure 4 for Exploring the Impact of Layer Normalization for Zero-shot Neural Machine Translation

This paper studies the impact of layer normalization (LayerNorm) on zero-shot translation (ZST). Recent efforts for ZST often utilize the Transformer architecture as the backbone, with LayerNorm at the input of layers (PreNorm) set as the default. However, Xu et al. (2019) has revealed that PreNorm carries the risk of overfitting the training data. Based on this, we hypothesize that PreNorm may overfit supervised directions and thus have low generalizability for ZST. Through experiments on OPUS, IWSLT, and Europarl datasets for 54 ZST directions, we demonstrate that the original Transformer setting of LayerNorm after residual connections (PostNorm) consistently outperforms PreNorm by up to 12.3 BLEU points. We then study the performance disparities by analyzing the differences in off-target rates and structural variations between PreNorm and PostNorm. This study highlights the need for careful consideration of the LayerNorm setting for ZST.

* Accepted to ACL 2023 main conference 
Viaarxiv icon

Comprehensive Solution Program Centric Pretraining for Table-and-Text Hybrid Numerical Reasoning

May 12, 2023
Qianying Liu, Dongsheng Yang, Wenjie Zhong, Fei Cheng, Sadao Kurohashi

Figure 1 for Comprehensive Solution Program Centric Pretraining for Table-and-Text Hybrid Numerical Reasoning
Figure 2 for Comprehensive Solution Program Centric Pretraining for Table-and-Text Hybrid Numerical Reasoning
Figure 3 for Comprehensive Solution Program Centric Pretraining for Table-and-Text Hybrid Numerical Reasoning
Figure 4 for Comprehensive Solution Program Centric Pretraining for Table-and-Text Hybrid Numerical Reasoning

Numerical reasoning over table-and-text hybrid passages, such as financial reports, poses significant challenges and has numerous potential applications. Noise and irrelevant variables in the model input have been a hindrance to its performance. Additionally, coarse-grained supervision of the whole solution program has impeded the model's ability to learn the underlying numerical reasoning process. In this paper, we propose three pretraining tasks that operate at both the whole program and sub-program level: Variable Integrity Ranking, which guides the model to focus on useful variables; Variable Operator Prediction, which decomposes the supervision into fine-grained single operator prediction; and Variable Keyphrase Masking, which encourages the model to identify key evidence that sub-programs are derived from. Experimental results demonstrate the effectiveness of our proposed methods, surpassing transformer-based model baselines.

* 11 pages 
Viaarxiv icon

GPT-RE: In-context Learning for Relation Extraction using Large Language Models

May 03, 2023
Zhen Wan, Fei Cheng, Zhuoyuan Mao, Qianying Liu, Haiyue Song, Jiwei Li, Sadao Kurohashi

Figure 1 for GPT-RE: In-context Learning for Relation Extraction using Large Language Models
Figure 2 for GPT-RE: In-context Learning for Relation Extraction using Large Language Models
Figure 3 for GPT-RE: In-context Learning for Relation Extraction using Large Language Models
Figure 4 for GPT-RE: In-context Learning for Relation Extraction using Large Language Models

In spite of the potential for ground-breaking achievements offered by large language models (LLMs) (e.g., GPT-3), they still lag significantly behind fully-supervised baselines (e.g., fine-tuned BERT) in relation extraction (RE). This is due to the two major shortcomings of LLMs in RE: (1) low relevance regarding entity and relation in retrieved demonstrations for in-context learning; and (2) the strong inclination to wrongly classify NULL examples into other pre-defined labels. In this paper, we propose GPT-RE to bridge the gap between LLMs and fully-supervised baselines. GPT-RE successfully addresses the aforementioned issues by (1) incorporating task-specific entity representations in demonstration retrieval; and (2) enriching the demonstrations with gold label-induced reasoning logic. We evaluate GPT-RE on four widely-used RE datasets, and observe that GPT-RE achieves improvements over not only existing GPT-3 baselines, but also fully-supervised baselines. Specifically, GPT-RE achieves SOTA performances on the Semeval and SciERC datasets, and competitive performances on the TACRED and ACE05 datasets.

Viaarxiv icon

Optimizing Facial Expressions of an Android Robot Effectively: a Bayesian Optimization Approach

Jan 13, 2023
Dongsheng Yang, Wataru Sato, Qianying Liu, Takashi Minato, Shushi Namba, Shin'ya Nishida

Figure 1 for Optimizing Facial Expressions of an Android Robot Effectively: a Bayesian Optimization Approach
Figure 2 for Optimizing Facial Expressions of an Android Robot Effectively: a Bayesian Optimization Approach
Figure 3 for Optimizing Facial Expressions of an Android Robot Effectively: a Bayesian Optimization Approach
Figure 4 for Optimizing Facial Expressions of an Android Robot Effectively: a Bayesian Optimization Approach

Expressing various facial emotions is an important social ability for efficient communication between humans. A key challenge in human-robot interaction research is providing androids with the ability to make various human-like facial expressions for efficient communication with humans. The android Nikola, we have developed, is equipped with many actuators for facial muscle control. While this enables Nikola to simulate various human expressions, it also complicates identification of the optimal parameters for producing desired expressions. Here, we propose a novel method that automatically optimizes the facial expressions of our android. We use a machine vision algorithm to evaluate the magnitudes of seven basic emotions, and employ the Bayesian Optimization algorithm to identify the parameters that produce the most convincing facial expressions. Evaluations by naive human participants demonstrate that our method improves the rated strength of the android's facial expressions of anger, disgust, sadness, and surprise compared with the previous method that relied on Ekman's theory and parameter adjustments by a human expert.

* 8 pages, 8 figures, accepted by Humanoids2022 
Viaarxiv icon

Textual Enhanced Contrastive Learning for Solving Math Word Problems

Nov 29, 2022
Yibin Shen, Qianying Liu, Zhuoyuan Mao, Fei Cheng, Sadao Kurohashi

Figure 1 for Textual Enhanced Contrastive Learning for Solving Math Word Problems
Figure 2 for Textual Enhanced Contrastive Learning for Solving Math Word Problems
Figure 3 for Textual Enhanced Contrastive Learning for Solving Math Word Problems
Figure 4 for Textual Enhanced Contrastive Learning for Solving Math Word Problems

Solving math word problems is the task that analyses the relation of quantities and requires an accurate understanding of contextual natural language information. Recent studies show that current models rely on shallow heuristics to predict solutions and could be easily misled by small textual perturbations. To address this problem, we propose a Textual Enhanced Contrastive Learning framework, which enforces the models to distinguish semantically similar examples while holding different mathematical logic. We adopt a self-supervised manner strategy to enrich examples with subtle textual variance by textual reordering or problem re-construction. We then retrieve the hardest to differentiate samples from both equation and textual perspectives and guide the model to learn their representations. Experimental results show that our method achieves state-of-the-art on both widely used benchmark datasets and also exquisitely designed challenge datasets in English and Chinese. \footnote{Our code and data is available at \url{https://github.com/yiyunya/Textual_CL_MWP}

* Findings of EMNLP 2022 
Viaarxiv icon

Optimizing Vision Transformers for Medical Image Segmentation and Few-Shot Domain Adaptation

Oct 14, 2022
Qianying Liu, Chaitanya Kaul, Christos Anagnostopoulos, Roderick Murray-Smith, Fani Deligianni

Figure 1 for Optimizing Vision Transformers for Medical Image Segmentation and Few-Shot Domain Adaptation
Figure 2 for Optimizing Vision Transformers for Medical Image Segmentation and Few-Shot Domain Adaptation
Figure 3 for Optimizing Vision Transformers for Medical Image Segmentation and Few-Shot Domain Adaptation
Figure 4 for Optimizing Vision Transformers for Medical Image Segmentation and Few-Shot Domain Adaptation

The adaptation of transformers to computer vision is not straightforward because the modelling of image contextual information results in quadratic computational complexity with relation to the input features. Most of existing methods require extensive pre-training on massive datasets such as ImageNet and therefore their application to fields such as healthcare is less effective. CNNs are the dominant architecture in computer vision tasks because convolutional filters can effectively model local dependencies and reduce drastically the parameters required. However, convolutional filters cannot handle more complex interactions, which are beyond a small neighbour of pixels. Furthermore, their weights are fixed after training and thus they do not take into consideration changes in the visual input. Inspired by recent work on hybrid visual transformers with convolutions and hierarchical transformers, we propose Convolutional Swin-Unet (CS-Unet) transformer blocks and optimise their settings with relation to patch embedding, projection, the feed-forward network, up sampling and skip connections. CS-Unet can be trained from scratch and inherits the superiority of convolutions in each feature process phase. It helps to encode precise spatial information and produce hierarchical representations that contribute to object concepts at various scales. Experiments show that CS-Unet without pre-training surpasses other state-of-the-art counterparts by large margins on two medical CT and MRI datasets with fewer parameters. In addition, two domain-adaptation experiments on optic disc and polyp image segmentation further prove that our method is highly generalizable and effectively bridges the domain gap between images from different sources.

Viaarxiv icon

ComSearch: Equation Searching with Combinatorial Strategy for Solving Math Word Problems with Weak Supervision

Oct 13, 2022
Qianying Liu, Wenyu Guan, Jianhao Shen, Fei Cheng, Sadao Kurohashi

Figure 1 for ComSearch: Equation Searching with Combinatorial Strategy for Solving Math Word Problems with Weak Supervision
Figure 2 for ComSearch: Equation Searching with Combinatorial Strategy for Solving Math Word Problems with Weak Supervision
Figure 3 for ComSearch: Equation Searching with Combinatorial Strategy for Solving Math Word Problems with Weak Supervision
Figure 4 for ComSearch: Equation Searching with Combinatorial Strategy for Solving Math Word Problems with Weak Supervision

Previous studies have introduced a weakly-supervised paradigm for solving math word problems requiring only the answer value annotation. While these methods search for correct value equation candidates as pseudo labels, they search among a narrow sub-space of the enormous equation space. To address this problem, we propose a novel search algorithm with combinatorial strategy \textbf{ComSearch}, which can compress the search space by excluding mathematically equivalent equations. The compression allows the searching algorithm to enumerate all possible equations and obtain high-quality data. We investigate the noise in the pseudo labels that hold wrong mathematical logic, which we refer to as the \textit{false-matching} problem, and propose a ranking model to denoise the pseudo labels. Our approach holds a flexible framework to utilize two existing supervised math word problem solvers to train pseudo labels, and both achieve state-of-the-art performance in the weak supervision task.

* 13 pages 
Viaarxiv icon

Seeking Diverse Reasoning Logic: Controlled Equation Expression Generation for Solving Math Word Problems

Sep 21, 2022
Yibin Shen, Qianying Liu, Zhuoyuan Mao, Zhen Wan, Fei Cheng, Sadao Kurohashi

Figure 1 for Seeking Diverse Reasoning Logic: Controlled Equation Expression Generation for Solving Math Word Problems
Figure 2 for Seeking Diverse Reasoning Logic: Controlled Equation Expression Generation for Solving Math Word Problems
Figure 3 for Seeking Diverse Reasoning Logic: Controlled Equation Expression Generation for Solving Math Word Problems
Figure 4 for Seeking Diverse Reasoning Logic: Controlled Equation Expression Generation for Solving Math Word Problems

To solve Math Word Problems, human students leverage diverse reasoning logic that reaches different possible equation solutions. However, the mainstream sequence-to-sequence approach of automatic solvers aims to decode a fixed solution equation supervised by human annotation. In this paper, we propose a controlled equation generation solver by leveraging a set of control codes to guide the model to consider certain reasoning logic and decode the corresponding equations expressions transformed from the human reference. The empirical results suggest that our method universally improves the performance on single-unknown (Math23K) and multiple-unknown (DRAW1K, HMWP) benchmarks, with substantial improvements up to 13.2% accuracy on the challenging multiple-unknown datasets.

* AACL 2022 short paper 
Viaarxiv icon

Relation Extraction with Weighted Contrastive Pre-training on Distant Supervision

May 18, 2022
Zhen Wan, Fei Cheng, Qianying Liu, Zhuoyuan Mao, Haiyue Song, Sadao Kurohashi

Figure 1 for Relation Extraction with Weighted Contrastive Pre-training on Distant Supervision
Figure 2 for Relation Extraction with Weighted Contrastive Pre-training on Distant Supervision
Figure 3 for Relation Extraction with Weighted Contrastive Pre-training on Distant Supervision
Figure 4 for Relation Extraction with Weighted Contrastive Pre-training on Distant Supervision

Contrastive pre-training on distant supervision has shown remarkable effectiveness for improving supervised relation extraction tasks. However, the existing methods ignore the intrinsic noise of distant supervision during the pre-training stage. In this paper, we propose a weighted contrastive learning method by leveraging the supervised data to estimate the reliability of pre-training instances and explicitly reduce the effect of noise. Experimental results on three supervised datasets demonstrate the advantages of our proposed weighted contrastive learning approach, compared to two state-of-the-art non-weighted baselines.

* Under review 
Viaarxiv icon