Alert button
Picture for Mojtaba Nayyeri

Mojtaba Nayyeri

Alert button

Geometric Relational Embeddings: A Survey

Apr 24, 2023
Bo Xiong, Mojtaba Nayyeri, Ming Jin, Yunjie He, Michael Cochez, Shirui Pan, Steffen Staab

Figure 1 for Geometric Relational Embeddings: A Survey
Figure 2 for Geometric Relational Embeddings: A Survey
Figure 3 for Geometric Relational Embeddings: A Survey
Figure 4 for Geometric Relational Embeddings: A Survey

Geometric relational embeddings map relational data as geometric objects that combine vector information suitable for machine learning and structured/relational information for structured/relational reasoning, typically in low dimensions. Their preservation of relational structures and their appealing properties and interpretability have led to their uptake for tasks such as knowledge graph completion, ontology and hierarchy reasoning, logical query answering, and hierarchical multi-label classification. We survey methods that underly geometric relational embeddings and categorize them based on (i) the embedding geometries that are used to represent the data; and (ii) the relational reasoning tasks that they aim to improve. We identify the desired properties (i.e., inductive biases) of each kind of embedding and discuss some potential future work.

* Work in progress 
Viaarxiv icon

Modeling Relational Patterns for Logical Query Answering over Knowledge Graphs

Mar 21, 2023
Yunjie He, Mojtaba Nayyeri, Bo Xiong, Evgeny Kharlamov, Steffen Staab

Figure 1 for Modeling Relational Patterns for Logical Query Answering over Knowledge Graphs
Figure 2 for Modeling Relational Patterns for Logical Query Answering over Knowledge Graphs
Figure 3 for Modeling Relational Patterns for Logical Query Answering over Knowledge Graphs
Figure 4 for Modeling Relational Patterns for Logical Query Answering over Knowledge Graphs

Answering first-order logical (FOL) queries over knowledge graphs (KG) remains a challenging task mainly due to KG incompleteness. Query embedding approaches this problem by computing the low-dimensional vector representations of entities, relations, and logical queries. KGs exhibit relational patterns such as symmetry and composition and modeling the patterns can further enhance the performance of query embedding models. However, the role of such patterns in answering FOL queries by query embedding models has not been yet studied in the literature. In this paper, we fill in this research gap and empower FOL queries reasoning with pattern inference by introducing an inductive bias that allows for learning relation patterns. To this end, we develop a novel query embedding method, RoConE, that defines query regions as geometric cones and algebraic query operators by rotations in complex space. RoConE combines the advantages of Cone as a well-specified geometric representation for query embedding, and also the rotation operator as a powerful algebraic operation for pattern inference. Our experimental results on several benchmark datasets confirm the advantage of relational patterns for enhancing logical query answering task.

Viaarxiv icon

Link Prediction with Attention Applied on Multiple Knowledge Graph Embedding Models

Feb 13, 2023
Cosimo Gregucci, Mojtaba Nayyeri, Daniel Hernández, Steffen Staab

Figure 1 for Link Prediction with Attention Applied on Multiple Knowledge Graph Embedding Models
Figure 2 for Link Prediction with Attention Applied on Multiple Knowledge Graph Embedding Models
Figure 3 for Link Prediction with Attention Applied on Multiple Knowledge Graph Embedding Models
Figure 4 for Link Prediction with Attention Applied on Multiple Knowledge Graph Embedding Models

Predicting missing links between entities in a knowledge graph is a fundamental task to deal with the incompleteness of data on the Web. Knowledge graph embeddings map nodes into a vector space to predict new links, scoring them according to geometric criteria. Relations in the graph may follow patterns that can be learned, e.g., some relations might be symmetric and others might be hierarchical. However, the learning capability of different embedding models varies for each pattern and, so far, no single model can learn all patterns equally well. In this paper, we combine the query representations from several models in a unified one to incorporate patterns that are independently captured by each model. Our combination uses attention to select the most suitable model to answer each query. The models are also mapped onto a non-Euclidean manifold, the Poincar\'e ball, to capture structural patterns, such as hierarchies, besides relational patterns, such as symmetry. We prove that our combination provides a higher expressiveness and inference power than each model on its own. As a result, the combined model can learn relational and structural patterns. We conduct extensive experimental analysis with various link prediction benchmarks showing that the combined model outperforms individual models, including state-of-the-art approaches.

* ACM Web Conference 2023 
Viaarxiv icon

Integrating Knowledge Graph embedding and pretrained Language Models in Hypercomplex Spaces

Aug 05, 2022
Mojtaba Nayyeri, Zihao Wang, Mst. Mahfuja Akter, Mirza Mohtashim Alam, Md Rashad Al Hasan Rony, Jens Lehmann, Steffen Staab

Figure 1 for Integrating Knowledge Graph embedding and pretrained Language Models in Hypercomplex Spaces
Figure 2 for Integrating Knowledge Graph embedding and pretrained Language Models in Hypercomplex Spaces
Figure 3 for Integrating Knowledge Graph embedding and pretrained Language Models in Hypercomplex Spaces
Figure 4 for Integrating Knowledge Graph embedding and pretrained Language Models in Hypercomplex Spaces

Knowledge Graphs, such as Wikidata, comprise structural and textual knowledge in order to represent knowledge. For each of the two modalities dedicated approaches for graph embedding and language models learn patterns that allow for predicting novel structural knowledge. Few approaches have integrated learning and inference with both modalities and these existing ones could only partially exploit the interaction of structural and textual knowledge. In our approach, we build on existing strong representations of single modalities and we use hypercomplex algebra to represent both, (i), single-modality embedding as well as, (ii), the interaction between different modalities and their complementary means of knowledge representation. More specifically, we suggest Dihedron and Quaternion representations of 4D hypercomplex numbers to integrate four modalities namely structural knowledge graph embedding, word-level representations (e.g.\ Word2vec, Fasttext), sentence-level representations (Sentence transformer), and document-level representations (sentence transformer, Doc2vec). Our unified vector representation scores the plausibility of labelled edges via Hamilton and Dihedron products, thus modeling pairwise interactions between different modalities. Extensive experimental evaluation on standard benchmark datasets shows the superiority of our two new models using abundant textual information besides sparse structural knowledge to enhance performance in link prediction tasks.

* add acknowledgement 
Viaarxiv icon

Ultrahyperbolic Knowledge Graph Embeddings

Jun 01, 2022
Bo Xiong, Shichao Zhu, Mojtaba Nayyeri, Chengjin Xu, Shirui Pan, Chuan Zhou, Steffen Staab

Figure 1 for Ultrahyperbolic Knowledge Graph Embeddings
Figure 2 for Ultrahyperbolic Knowledge Graph Embeddings
Figure 3 for Ultrahyperbolic Knowledge Graph Embeddings
Figure 4 for Ultrahyperbolic Knowledge Graph Embeddings

Recent knowledge graph (KG) embeddings have been advanced by hyperbolic geometry due to its superior capability for representing hierarchies. The topological structures of real-world KGs, however, are rather heterogeneous, i.e., a KG is composed of multiple distinct hierarchies and non-hierarchical graph structures. Therefore, a homogeneous (either Euclidean or hyperbolic) geometry is not sufficient for fairly representing such heterogeneous structures. To capture the topological heterogeneity of KGs, we present an ultrahyperbolic KG embedding (UltraE) in an ultrahyperbolic (or pseudo-Riemannian) manifold that seamlessly interleaves hyperbolic and spherical manifolds. In particular, we model each relation as a pseudo-orthogonal transformation that preserves the pseudo-Riemannian bilinear form. The pseudo-orthogonal transformation is decomposed into various operators (i.e., circular rotations, reflections and hyperbolic rotations), allowing for simultaneously modeling heterogeneous structures as well as complex relational patterns. Experimental results on three standard KGs show that UltraE outperforms previous Euclidean- and hyperbolic-based approaches.

* Accepted to SIGKDD'22 
Viaarxiv icon

Geometric Algebra based Embeddings for Static and Temporal Knowledge Graph Completion

Feb 25, 2022
Chengjin Xu, Mojtaba Nayyeri, Yung-Yu Chen, Jens Lehmann

Figure 1 for Geometric Algebra based Embeddings for Static and Temporal Knowledge Graph Completion
Figure 2 for Geometric Algebra based Embeddings for Static and Temporal Knowledge Graph Completion
Figure 3 for Geometric Algebra based Embeddings for Static and Temporal Knowledge Graph Completion
Figure 4 for Geometric Algebra based Embeddings for Static and Temporal Knowledge Graph Completion

Recent years, Knowledge Graph Embeddings (KGEs) have shown promising performance on link prediction tasks by mapping the entities and relations from a Knowledge Graph (KG) into a geometric space and thus have gained increasing attentions. In addition, many recent Knowledge Graphs involve evolving data, e.g., the fact (\textit{Obama}, \textit{PresidentOf}, \textit{USA}) is valid only from 2009 to 2017. This introduces important challenges for knowledge representation learning since such temporal KGs change over time. In this work, we strive to move beyond the complex or hypercomplex space for KGE and propose a novel geometric algebra based embedding approach, GeomE, which uses multivector representations and the geometric product to model entities and relations. GeomE subsumes several state-of-the-art KGE models and is able to model diverse relations patterns. On top of this, we extend GeomE to TGeomE for temporal KGE, which performs 4th-order tensor factorization of a temporal KG and devises a new linear temporal regularization for time representation learning. Moreover, we study the effect of time granularity on the performance of TGeomE models. Experimental results show that our proposed models achieve the state-of-the-art performances on link prediction over four commonly-used static KG datasets and four well-established temporal KG datasets across various metrics.

* There are some theorem mistakes in the Appendix section need to be fixed. And we are still trying to solve them. We submitted the Arxiv version for providing the supplementary analysis, but now we hope to withdraw the current version to avoid misleading the readers from Arxiv 
Viaarxiv icon

Geometric Algebra based Embeddings for Staticand Temporal Knowledge Graph Completion

Feb 18, 2022
Chengjin Xu, Mojtaba Nayyeri, Yung-Yu Chen, Jens Lehmann

Figure 1 for Geometric Algebra based Embeddings for Staticand Temporal Knowledge Graph Completion
Figure 2 for Geometric Algebra based Embeddings for Staticand Temporal Knowledge Graph Completion
Figure 3 for Geometric Algebra based Embeddings for Staticand Temporal Knowledge Graph Completion
Figure 4 for Geometric Algebra based Embeddings for Staticand Temporal Knowledge Graph Completion

Recent years, Knowledge Graph Embeddings (KGEs) have shown promising performance on link prediction tasks by mapping the entities and relations from a Knowledge Graph (KG) into a geometric space and thus have gained increasing attentions. In addition, many recent Knowledge Graphs involve evolving data, e.g., the fact (\textit{Obama}, \textit{PresidentOf}, \textit{USA}) is valid only from 2009 to 2017. This introduces important challenges for knowledge representation learning since such temporal KGs change over time. In this work, we strive to move beyond the complex or hypercomplex space for KGE and propose a novel geometric algebra based embedding approach, GeomE, which uses multivector representations and the geometric product to model entities and relations. GeomE subsumes several state-of-the-art KGE models and is able to model diverse relations patterns. On top of this, we extend GeomE to TGeomE for temporal KGE, which performs 4th-order tensor factorization of a temporal KG and devises a new linear temporal regularization for time representation learning. Moreover, we study the effect of time granularity on the performance of TGeomE models. Experimental results show that our proposed models achieve the state-of-the-art performances on link prediction over four commonly-used static KG datasets and four well-established temporal KG datasets across various metrics.

* Accepted by IEEE Transactions on Knowledge and Data Engineering 
Viaarxiv icon

Box Embeddings for the Description Logic EL++

Jan 24, 2022
Bo Xiong, Nico Potyka, Trung-Kien Tran, Mojtaba Nayyeri, Steffen Staab

Figure 1 for Box Embeddings for the Description Logic EL++
Figure 2 for Box Embeddings for the Description Logic EL++
Figure 3 for Box Embeddings for the Description Logic EL++
Figure 4 for Box Embeddings for the Description Logic EL++

Recently, various methods for representation learning on Knowledge Bases (KBs) have been developed. However, these approaches either only focus on learning the embeddings of the data-level knowledge (ABox) or exhibit inherent limitations when dealing with the concept-level knowledge (TBox), e.g., not properly modelling the structure of the logical knowledge. We present BoxEL, a geometric KB embedding approach that allows for better capturing logical structure expressed in the theories of Description Logic EL++. BoxEL models concepts in a KB as axis-parallel boxes exhibiting the advantage of intersectional closure, entities as points inside boxes, and relations between concepts/entities as affine transformations. We show theoretical guarantees (soundness) of BoxEL for preserving logical structure. Namely, the trained model of BoxEL embedding with loss 0 is a (logical) model of the KB. Experimental results on subsumption reasoning and a real-world application--protein-protein prediction show that BoxEL outperforms traditional knowledge graph embedding methods as well as state-of-the-art EL++ embedding approaches.

* 11 pages, 3 figures 
Viaarxiv icon

Trans4E: Link Prediction on Scholarly Knowledge Graphs

Jul 03, 2021
Mojtaba Nayyeri, Gokce Muge Cil, Sahar Vahdati, Francesco Osborne, Mahfuzur Rahman, Simone Angioni, Angelo Salatino, Diego Reforgiato Recupero, Nadezhda Vassilyeva, Enrico Motta, Jens Lehmann

Figure 1 for Trans4E: Link Prediction on Scholarly Knowledge Graphs
Figure 2 for Trans4E: Link Prediction on Scholarly Knowledge Graphs
Figure 3 for Trans4E: Link Prediction on Scholarly Knowledge Graphs
Figure 4 for Trans4E: Link Prediction on Scholarly Knowledge Graphs

The incompleteness of Knowledge Graphs (KGs) is a crucial issue affecting the quality of AI-based services. In the scholarly domain, KGs describing research publications typically lack important information, hindering our ability to analyse and predict research dynamics. In recent years, link prediction approaches based on Knowledge Graph Embedding models became the first aid for this issue. In this work, we present Trans4E, a novel embedding model that is particularly fit for KGs which include N to M relations with N$\gg$M. This is typical for KGs that categorize a large number of entities (e.g., research articles, patents, persons) according to a relatively small set of categories. Trans4E was applied on two large-scale knowledge graphs, the Academia/Industry DynAmics (AIDA) and Microsoft Academic Graph (MAG), for completing the information about Fields of Study (e.g., 'neural networks', 'machine learning', 'artificial intelligence'), and affiliation types (e.g., 'education', 'company', 'government'), improving the scope and accuracy of the resulting data. We evaluated our approach against alternative solutions on AIDA, MAG, and four other benchmarks (FB15k, FB15k-237, WN18, and WN18RR). Trans4E outperforms the other models when using low embedding dimensions and obtains competitive results in high dimensions.

Viaarxiv icon

Multiple Run Ensemble Learning withLow-Dimensional Knowledge Graph Embeddings

Apr 11, 2021
Chengjin Xu, Mojtaba Nayyeri, Sahar Vahdati, Jens Lehmann

Figure 1 for Multiple Run Ensemble Learning withLow-Dimensional Knowledge Graph Embeddings
Figure 2 for Multiple Run Ensemble Learning withLow-Dimensional Knowledge Graph Embeddings
Figure 3 for Multiple Run Ensemble Learning withLow-Dimensional Knowledge Graph Embeddings
Figure 4 for Multiple Run Ensemble Learning withLow-Dimensional Knowledge Graph Embeddings

Among the top approaches of recent years, link prediction using knowledge graph embedding (KGE) models has gained significant attention for knowledge graph completion. Various embedding models have been proposed so far, among which, some recent KGE models obtain state-of-the-art performance on link prediction tasks by using embeddings with a high dimension (e.g. 1000) which accelerate the costs of training and evaluation considering the large scale of KGs. In this paper, we propose a simple but effective performance boosting strategy for KGE models by using multiple low dimensions in different repetition rounds of the same model. For example, instead of training a model one time with a large embedding size of 1200, we repeat the training of the model 6 times in parallel with an embedding size of 200 and then combine the 6 separate models for testing while the overall numbers of adjustable parameters are same (6*200=1200) and the total memory footprint remains the same. We show that our approach enables different models to better cope with their expressiveness issues on modeling various graph patterns such as symmetric, 1-n, n-1 and n-n. In order to justify our findings, we conduct experiments on various KGE models. Experimental results on standard benchmark datasets, namely FB15K, FB15K-237 and WN18RR, show that multiple low-dimensional models of the same kind outperform the corresponding single high-dimensional models on link prediction in a certain range and have advantages in training efficiency by using parallel training while the overall numbers of adjustable parameters are same.

* Accepted by the 2021 International Joint Conference on Neural Networks (IJCNN 2021) 
Viaarxiv icon