Alert button
Picture for Bastian Rieck

Bastian Rieck

Alert button

Metric Space Magnitude for Evaluating Unsupervised Representation Learning

Nov 27, 2023
Katharina Limbeck, Rayna Andreeva, Rik Sarkar, Bastian Rieck

The magnitude of a metric space was recently established as a novel invariant, providing a measure of the `effective size' of a space across multiple scales. By capturing both geometrical and topological properties of data, magnitude is poised to address challenges in unsupervised representation learning tasks. We formalise a novel notion of dissimilarity between magnitude functions of finite metric spaces and use them to derive a quality measure for dimensionality reduction tasks. Our measure is provably stable under perturbations of the data, can be efficiently calculated, and enables a rigorous multi-scale comparison of embeddings. We show the utility of our measure in an experimental suite that comprises different domains and tasks, including the comparison of data visualisations.

Viaarxiv icon

Differentiable Euler Characteristic Transforms for Shape Classification

Oct 11, 2023
Ernst Roell, Bastian Rieck

The Euler Characteristic Transform (ECT) has proven to be a powerful representation, combining geometrical and topological characteristics of shapes and graphs. However, the ECT was hitherto unable to learn task-specific representations. We overcome this issue and develop a novel computational layer that enables learning the ECT in an end-to-end fashion. Our method DECT is fast and computationally efficient, while exhibiting performance on a par with more complex models in both graph and point cloud classification tasks. Moreover, we show that this seemingly unexpressive statistic still provides the same topological expressivity as more complex topological deep learning layers provide.

Viaarxiv icon

Filtration Surfaces for Dynamic Graph Classification

Sep 07, 2023
Franz Srambical, Bastian Rieck

Existing approaches for classifying dynamic graphs either lift graph kernels to the temporal domain, or use graph neural networks (GNNs). However, current baselines have scalability issues, cannot handle a changing node set, or do not take edge weight information into account. We propose filtration surfaces, a novel method that is scalable and flexible, to alleviate said restrictions. We experimentally validate the efficacy of our model and show that filtration surfaces outperform previous state-of-the-art baselines on datasets that rely on edge weight information. Our method does so while being either completely parameter-free or having at most one parameter, and yielding the lowest overall standard deviation.

Viaarxiv icon

Topologically-Regularized Multiple Instance Learning for Red Blood Cell Disease Classification

Jul 26, 2023
Salome Kazeminia, Ario Sadafi, Asya Makhro, Anna Bogdanova, Carsten Marr, Bastian Rieck

Figure 1 for Topologically-Regularized Multiple Instance Learning for Red Blood Cell Disease Classification
Figure 2 for Topologically-Regularized Multiple Instance Learning for Red Blood Cell Disease Classification
Figure 3 for Topologically-Regularized Multiple Instance Learning for Red Blood Cell Disease Classification
Figure 4 for Topologically-Regularized Multiple Instance Learning for Red Blood Cell Disease Classification

Diagnosing rare anemia disorders using microscopic images is challenging for skilled specialists and machine-learning methods alike. Due to thousands of disease-relevant cells in a single blood sample, this constitutes a complex multiple-instance learning (MIL) problem. While the spatial neighborhood of red blood cells is not meaningful per se, the topology, i.e., the geometry of blood samples as a whole, contains informative features to remedy typical MIL issues, such as vanishing gradients and overfitting when training on limited data. We thus develop a topology-based approach that extracts multi-scale topological features from bags of single red blood cell images. The topological features are used to regularize the model, enforcing the preservation of characteristic topological properties of the data. Applied to a dataset of 71 patients suffering from rare anemia disorders with 521 microscopic images of red blood cells, our experiments show that topological regularization is an effective method that leads to more than 3% performance improvements for the automated classification of rare anemia disorders based on single-cell images. This is the first approach that uses topological properties for regularizing the MIL process.

Viaarxiv icon

Evaluating the "Learning on Graphs" Conference Experience

Jun 01, 2023
Bastian Rieck, Corinna Coupette

With machine learning conferences growing ever larger, and reviewing processes becoming increasingly elaborate, more data-driven insights into their workings are required. In this report, we present the results of a survey accompanying the first "Learning on Graphs" (LoG) Conference. The survey was directed to evaluate the submission and review process from different perspectives, including authors, reviewers, and area chairs alike.

Viaarxiv icon

MAGNet: Motif-Agnostic Generation of Molecules from Shapes

May 30, 2023
Leon Hetzel, Johanna Sommer, Bastian Rieck, Fabian Theis, Stephan Günnemann

Figure 1 for MAGNet: Motif-Agnostic Generation of Molecules from Shapes
Figure 2 for MAGNet: Motif-Agnostic Generation of Molecules from Shapes
Figure 3 for MAGNet: Motif-Agnostic Generation of Molecules from Shapes
Figure 4 for MAGNet: Motif-Agnostic Generation of Molecules from Shapes

Recent advances in machine learning for molecules exhibit great potential for facilitating drug discovery from in silico predictions. Most models for molecule generation rely on the decomposition of molecules into frequently occurring substructures (motifs), from which they generate novel compounds. While motif representations greatly aid in learning molecular distributions, such methods struggle to represent substructures beyond their known motif set. To alleviate this issue and increase flexibility across datasets, we propose MAGNet, a graph-based model that generates abstract shapes before allocating atom and bond types. To this end, we introduce a novel factorisation of the molecules' data distribution that accounts for the molecules' global context and facilitates learning adequate assignments of atoms and bonds onto shapes. While the abstraction to shapes introduces greater complexity for distribution learning, we show the competitive performance of MAGNet on standard benchmarks. Importantly, we demonstrate that MAGNet's improved expressivity leads to molecules with more topologically distinct structures and, at the same time, diverse atom and bond assignments.

Viaarxiv icon

Metric Space Magnitude and Generalisation in Neural Networks

May 09, 2023
Rayna Andreeva, Katharina Limbeck, Bastian Rieck, Rik Sarkar

Figure 1 for Metric Space Magnitude and Generalisation in Neural Networks
Figure 2 for Metric Space Magnitude and Generalisation in Neural Networks
Figure 3 for Metric Space Magnitude and Generalisation in Neural Networks
Figure 4 for Metric Space Magnitude and Generalisation in Neural Networks

Deep learning models have seen significant successes in numerous applications, but their inner workings remain elusive. The purpose of this work is to quantify the learning process of deep neural networks through the lens of a novel topological invariant called magnitude. Magnitude is an isometry invariant; its properties are an active area of research as it encodes many known invariants of a metric space. We use magnitude to study the internal representations of neural networks and propose a new method for determining their generalisation capabilities. Moreover, we theoretically connect magnitude dimension and the generalisation error, and demonstrate experimentally that the proposed framework can be a good indicator of the latter.

Viaarxiv icon

Euler Characteristic Transform Based Topological Loss for Reconstructing 3D Images from Single 2D Slices

Mar 08, 2023
Kalyan Varma Nadimpalli, Amit Chattopadhyay, Bastian Rieck

Figure 1 for Euler Characteristic Transform Based Topological Loss for Reconstructing 3D Images from Single 2D Slices
Figure 2 for Euler Characteristic Transform Based Topological Loss for Reconstructing 3D Images from Single 2D Slices
Figure 3 for Euler Characteristic Transform Based Topological Loss for Reconstructing 3D Images from Single 2D Slices
Figure 4 for Euler Characteristic Transform Based Topological Loss for Reconstructing 3D Images from Single 2D Slices

The computer vision task of reconstructing 3D images, i.e., shapes, from their single 2D image slices is extremely challenging, more so in the regime of limited data. Deep learning models typically optimize geometric loss functions, which may lead to poor reconstructions as they ignore the structural properties of the shape. To tackle this, we propose a novel topological loss function based on the Euler Characteristic Transform. This loss can be used as an inductive bias to aid the optimization of any neural network toward better reconstructions in the regime of limited data. We show the effectiveness of the proposed loss function by incorporating it into SHAPR, a state-of-the-art shape reconstruction model, and test it on two benchmark datasets, viz., Red Blood Cells and Nuclei datasets. We also show a favourable property, namely injectivity and discuss the stability of the topological loss function based on the Euler Characteristic Transform.

* e-print 
Viaarxiv icon

On the Expressivity of Persistent Homology in Graph Learning

Feb 20, 2023
Bastian Rieck

Figure 1 for On the Expressivity of Persistent Homology in Graph Learning
Figure 2 for On the Expressivity of Persistent Homology in Graph Learning
Figure 3 for On the Expressivity of Persistent Homology in Graph Learning
Figure 4 for On the Expressivity of Persistent Homology in Graph Learning

Persistent homology, a technique from computational topology, has recently shown strong empirical performance in the context of graph classification. Being able to capture long range graph properties via higher-order topological features, such as cycles of arbitrary length, in combination with multi-scale topological descriptors, has improved predictive performance for data sets with prominent topological structures, such as molecules. At the same time, the theoretical properties of persistent homology have not been formally assessed in this context. This paper intends to bridge the gap between computational topology and graph machine learning by providing a brief introduction to persistent homology in the context of graphs, as well as a theoretical discussion and empirical analysis of its expressivity for graph learning tasks.

Viaarxiv icon

Curvature Filtrations for Graph Generative Model Evaluation

Jan 30, 2023
Joshua Southern, Jeremy Wayland, Michael Bronstein, Bastian Rieck

Figure 1 for Curvature Filtrations for Graph Generative Model Evaluation
Figure 2 for Curvature Filtrations for Graph Generative Model Evaluation
Figure 3 for Curvature Filtrations for Graph Generative Model Evaluation
Figure 4 for Curvature Filtrations for Graph Generative Model Evaluation

Graph generative model evaluation necessitates understanding differences between graphs on the distributional level. This entails being able to harness salient attributes of graphs in an efficient manner. Curvature constitutes one such property of graphs, and has recently started to prove useful in characterising graphs. Its expressive properties, stability, and practical utility in model evaluation remain largely unexplored, however. We combine graph curvature descriptors with cutting-edge methods from topological data analysis to obtain robust, expressive descriptors for evaluating graph generative models.

Viaarxiv icon