Alert button
Picture for Ke Liu

Ke Liu

Alert button

Unified Matrix Factorization with Dynamic Multi-view Clustering

Aug 13, 2023
Shangde Gao, Ke Liu, Yichao Fu

Figure 1 for Unified Matrix Factorization with Dynamic Multi-view Clustering
Figure 2 for Unified Matrix Factorization with Dynamic Multi-view Clustering
Figure 3 for Unified Matrix Factorization with Dynamic Multi-view Clustering
Figure 4 for Unified Matrix Factorization with Dynamic Multi-view Clustering

Matrix factorization (MF) is a classical collaborative filtering algorithm for recommender systems. It decomposes the user-item interaction matrix into a product of low-dimensional user representation matrix and item representation matrix. In typical recommendation scenarios, the user-item interaction paradigm is usually a two-stage process and requires static clustering analysis of the obtained user and item representations. The above process, however, is time and computationally intensive, making it difficult to apply in real-time to e-commerce or Internet of Things environments with billions of users and trillions of items. To address this, we propose a unified matrix factorization method based on dynamic multi-view clustering (MFDMC) that employs an end-to-end training paradigm. Specifically, in each view, a user/item representation is regarded as a weighted projection of all clusters. The representation of each cluster is learnable, enabling the dynamic discarding of bad clusters. Furthermore, we employ multi-view clustering to represent multiple roles of users/items, effectively utilizing the representation space and improving the interpretability of the user/item representations for downstream tasks. Extensive experiments show that our proposed MFDMC achieves state-of-the-art performance on real-world recommendation datasets. Additionally, comprehensive visualization and ablation studies interpretably confirm that our method provides meaningful representations for downstream tasks of users/items.

Viaarxiv icon

Contrastive Knowledge Amalgamation for Unsupervised Image Classification

Jul 27, 2023
Shangde Gao, Yichao Fu, Ke Liu, Yuqiang Han

Figure 1 for Contrastive Knowledge Amalgamation for Unsupervised Image Classification
Figure 2 for Contrastive Knowledge Amalgamation for Unsupervised Image Classification
Figure 3 for Contrastive Knowledge Amalgamation for Unsupervised Image Classification
Figure 4 for Contrastive Knowledge Amalgamation for Unsupervised Image Classification

Knowledge amalgamation (KA) aims to learn a compact student model to handle the joint objective from multiple teacher models that are are specialized for their own tasks respectively. Current methods focus on coarsely aligning teachers and students in the common representation space, making it difficult for the student to learn the proper decision boundaries from a set of heterogeneous teachers. Besides, the KL divergence in previous works only minimizes the probability distribution difference between teachers and the student, ignoring the intrinsic characteristics of teachers. Therefore, we propose a novel Contrastive Knowledge Amalgamation (CKA) framework, which introduces contrastive losses and an alignment loss to achieve intra-class cohesion and inter-class separation.Contrastive losses intra- and inter- models are designed to widen the distance between representations of different classes. The alignment loss is introduced to minimize the sample-level distribution differences of teacher-student models in the common representation space.Furthermore, the student learns heterogeneous unsupervised classification tasks through soft targets efficiently and flexibly in the task-level amalgamation. Extensive experiments on benchmarks demonstrate the generalization capability of CKA in the amalgamation of specific task as well as multiple tasks. Comprehensive ablation studies provide a further insight into our CKA.

* arXiv admin note: text overlap with arXiv:2112.07327 by other authors 
Viaarxiv icon

$E(2)$-Equivariant Vision Transformer

Jul 04, 2023
Renjun Xu, Kaifan Yang, Ke Liu, Fengxiang He

Figure 1 for $E(2)$-Equivariant Vision Transformer
Figure 2 for $E(2)$-Equivariant Vision Transformer
Figure 3 for $E(2)$-Equivariant Vision Transformer
Figure 4 for $E(2)$-Equivariant Vision Transformer

Vision Transformer (ViT) has achieved remarkable performance in computer vision. However, positional encoding in ViT makes it substantially difficult to learn the intrinsic equivariance in data. Initial attempts have been made on designing equivariant ViT but are proved defective in some cases in this paper. To address this issue, we design a Group Equivariant Vision Transformer (GE-ViT) via a novel, effective positional encoding operator. We prove that GE-ViT meets all the theoretical requirements of an equivariant neural network. Comprehensive experiments are conducted on standard benchmark datasets, demonstrating that GE-ViT significantly outperforms non-equivariant self-attention networks. The code is available at https://github.com/ZJUCDSYangKaifan/GEVit.

* Accept to UAI2023 
Viaarxiv icon

S2SNet: A Pretrained Neural Network for Superconductivity Discovery

Jun 28, 2023
Ke Liu, Kaifan Yang, Jiahong Zhang, Renjun Xu

Superconductivity allows electrical current to flow without any energy loss, and thus making solids superconducting is a grand goal of physics, material science, and electrical engineering. More than 16 Nobel Laureates have been awarded for their contribution to superconductivity research. Superconductors are valuable for sustainable development goals (SDGs), such as climate change mitigation, affordable and clean energy, industry, innovation and infrastructure, and so on. However, a unified physics theory explaining all superconductivity mechanism is still unknown. It is believed that superconductivity is microscopically due to not only molecular compositions but also the geometric crystal structure. Hence a new dataset, S2S, containing both crystal structures and superconducting critical temperature, is built upon SuperCon and Material Project. Based on this new dataset, we propose a novel model, S2SNet, which utilizes the attention mechanism for superconductivity prediction. To overcome the shortage of data, S2SNet is pre-trained on the whole Material Project dataset with Masked-Language Modeling (MLM). S2SNet makes a new state-of-the-art, with out-of-sample accuracy of 92% and Area Under Curve (AUC) of 0.92. To the best of our knowledge, S2SNet is the first work to predict superconductivity with only information of crystal structures. This work is beneficial to superconductivity discovery and further SDGs. Code and datasets are available in https://github.com/zjuKeLiu/S2SNet

* Accepted to IJCAI 2022 
Viaarxiv icon

Supervised Deep Hashing for High-dimensional and Heterogeneous Case-based Reasoning

Jun 29, 2022
Qi Zhang, Liang Hu, Chongyang Shi, Ke Liu, Longbing Cao

Figure 1 for Supervised Deep Hashing for High-dimensional and Heterogeneous Case-based Reasoning
Figure 2 for Supervised Deep Hashing for High-dimensional and Heterogeneous Case-based Reasoning
Figure 3 for Supervised Deep Hashing for High-dimensional and Heterogeneous Case-based Reasoning
Figure 4 for Supervised Deep Hashing for High-dimensional and Heterogeneous Case-based Reasoning

Case-based Reasoning (CBR) on high-dimensional and heterogeneous data is a trending yet challenging and computationally expensive task in the real world. A promising approach is to obtain low-dimensional hash codes representing cases and perform a similarity retrieval of cases in Hamming space. However, previous methods based on data-independent hashing rely on random projections or manual construction, inapplicable to address specific data issues (e.g., high-dimensionality and heterogeneity) due to their insensitivity to data characteristics. To address these issues, this work introduces a novel deep hashing network to learn similarity-preserving compact hash codes for efficient case retrieval and proposes a deep-hashing-enabled CBR model HeCBR. Specifically, we introduce position embedding to represent heterogeneous features and utilize a multilinear interaction layer to obtain case embeddings, which effectively filtrates zero-valued features to tackle high-dimensionality and sparsity and captures inter-feature couplings. Then, we feed the case embeddings into fully-connected layers, and subsequently a hash layer generates hash codes with a quantization regularizer to control the quantization loss during relaxation. To cater to incremental learning of CBR, we further propose an adaptive learning strategy to update the hash function. Extensive experiments on public datasets show that HeCBR greatly reduces storage and significantly accelerates case retrieval. HeCBR achieves desirable performance compared with the state-of-the-art CBR methods and performs significantly better than hashing-based CBR methods in classification.

* 27 pages 
Viaarxiv icon

Learning Universal User Representations via Self-Supervised Lifelong Behaviors Modeling

Oct 25, 2021
Bei Yang, Ke Liu, Xiaoxiao Xu, Renjun Xu, Hong Liu, Huan Xu

Figure 1 for Learning Universal User Representations via Self-Supervised Lifelong Behaviors Modeling
Figure 2 for Learning Universal User Representations via Self-Supervised Lifelong Behaviors Modeling
Figure 3 for Learning Universal User Representations via Self-Supervised Lifelong Behaviors Modeling
Figure 4 for Learning Universal User Representations via Self-Supervised Lifelong Behaviors Modeling

Universal user representation is an important research topic in industry, and is widely used in diverse downstream user analysis tasks, such as user profiling and user preference prediction. With the rapid development of Internet service platforms, extremely long user behavior sequences have been accumulated. However, existing researches have little ability to model universal user representation based on lifelong sequences of user behavior since registration. In this study, we propose a novel framework called Lifelong User Representation Model (LURM) to tackle this challenge. Specifically, LURM consists of two cascaded sub-models: (i) Bag of Interests (BoI) encodes user behaviors in any time period into a sparse vector with super-high dimension (e.g.,105); (ii) Self-supervised Multi-anchor EncoderNetwork (SMEN) maps sequences of BoI features to multiple low-dimensional user representations by contrastive learning. SMEN achieves almost lossless dimensionality reduction, benefiting from a novel multi-anchor module which can learn different aspects of user preferences. Experiments on several benchmark datasets show that our approach outperforms state-of-the-art unsupervised representation methods in downstream tasks

* during peer review 
Viaarxiv icon

Focusing on Persons: Colorizing Old Images Learning from Modern Historical Movies

Aug 14, 2021
Xin Jin, Zhonglan Li, Ke Liu, Dongqing Zou, Xiaodong Li, Xingfan Zhu, Ziyin Zhou, Qilong Sun, Qingyu Liu

Figure 1 for Focusing on Persons: Colorizing Old Images Learning from Modern Historical Movies
Figure 2 for Focusing on Persons: Colorizing Old Images Learning from Modern Historical Movies
Figure 3 for Focusing on Persons: Colorizing Old Images Learning from Modern Historical Movies
Figure 4 for Focusing on Persons: Colorizing Old Images Learning from Modern Historical Movies

In industry, there exist plenty of scenarios where old gray photos need to be automatically colored, such as video sites and archives. In this paper, we present the HistoryNet focusing on historical person's diverse high fidelity clothing colorization based on fine grained semantic understanding and prior. Colorization of historical persons is realistic and practical, however, existing methods do not perform well in the regards. In this paper, a HistoryNet including three parts, namely, classification, fine grained semantic parsing and colorization, is proposed. Classification sub-module supplies classifying of images according to the eras, nationalities and garment types; Parsing sub-network supplies the semantic for person contours, clothing and background in the image to achieve more accurate colorization of clothes and persons and prevent color overflow. In the training process, we integrate classification and semantic parsing features into the coloring generation network to improve colorization. Through the design of classification and parsing subnetwork, the accuracy of image colorization can be improved and the boundary of each part of image can be more clearly. Moreover, we also propose a novel Modern Historical Movies Dataset (MHMD) containing 1,353,166 images and 42 labels of eras, nationalities, and garment types for automatic colorization from 147 historical movies or TV series made in modern time. Various quantitative and qualitative comparisons demonstrate that our method outperforms the state-of-the-art colorization methods, especially on military uniforms, which has correct colors according to the historical literatures.

* ACM Multimedia 2021 Industrial Track 
Viaarxiv icon

ChemiRise: a data-driven retrosynthesis engine

Aug 09, 2021
Xiangyan Sun, Ke Liu, Yuquan Lin, Lingjie Wu, Haoming Xing, Minghong Gao, Ji Liu, Suocheng Tan, Zekun Ni, Qi Han, Junqiu Wu, Jie Fan

Figure 1 for ChemiRise: a data-driven retrosynthesis engine
Figure 2 for ChemiRise: a data-driven retrosynthesis engine
Figure 3 for ChemiRise: a data-driven retrosynthesis engine
Figure 4 for ChemiRise: a data-driven retrosynthesis engine

We have developed an end-to-end, retrosynthesis system, named ChemiRise, that can propose complete retrosynthesis routes for organic compounds rapidly and reliably. The system was trained on a processed patent database of over 3 million organic reactions. Experimental reactions were atom-mapped, clustered, and extracted into reaction templates. We then trained a graph convolutional neural network-based one-step reaction proposer using template embeddings and developed a guiding algorithm on the directed acyclic graph (DAG) of chemical compounds to find the best candidate to explore. The atom-mapping algorithm and the one-step reaction proposer were benchmarked against previous studies and showed better results. The final product was demonstrated by retrosynthesis routes reviewed and rated by human experts, showing satisfying functionality and a potential productivity boost in real-life use cases.

Viaarxiv icon