Alert button
Picture for Lucas G. S. Jeub

Lucas G. S. Jeub

Alert button

Local2Global: A distributed approach for scaling representation learning on graphs

Jan 12, 2022
Lucas G. S. Jeub, Giovanni Colavizza, Xiaowen Dong, Marya Bazzi, Mihai Cucuringu

Figure 1 for Local2Global: A distributed approach for scaling representation learning on graphs
Figure 2 for Local2Global: A distributed approach for scaling representation learning on graphs
Figure 3 for Local2Global: A distributed approach for scaling representation learning on graphs
Figure 4 for Local2Global: A distributed approach for scaling representation learning on graphs

We propose a decentralised "local2global"' approach to graph representation learning, that one can a-priori use to scale any embedding technique. Our local2global approach proceeds by first dividing the input graph into overlapping subgraphs (or "patches") and training local representations for each patch independently. In a second step, we combine the local representations into a globally consistent representation by estimating the set of rigid motions that best align the local representations using information from the patch overlaps, via group synchronization. A key distinguishing feature of local2global relative to existing work is that patches are trained independently without the need for the often costly parameter synchronization during distributed training. This allows local2global to scale to large-scale industrial applications, where the input graph may not even fit into memory and may be stored in a distributed manner. We apply local2global on data sets of different sizes and show that our approach achieves a good trade-off between scale and accuracy on edge reconstruction and semi-supervised classification. We also consider the downstream task of anomaly detection and show how one can use local2global to highlight anomalies in cybersecurity networks.

* 25 pages, 6 figures. arXiv admin note: substantial text overlap with arXiv:2107.12224 
Viaarxiv icon

Local2Global: Scaling global representation learning on graphs via local training

Jul 26, 2021
Lucas G. S. Jeub, Giovanni Colavizza, Xiaowen Dong, Marya Bazzi, Mihai Cucuringu

Figure 1 for Local2Global: Scaling global representation learning on graphs via local training

We propose a decentralised "local2global" approach to graph representation learning, that one can a-priori use to scale any embedding technique. Our local2global approach proceeds by first dividing the input graph into overlapping subgraphs (or "patches") and training local representations for each patch independently. In a second step, we combine the local representations into a globally consistent representation by estimating the set of rigid motions that best align the local representations using information from the patch overlaps, via group synchronization. A key distinguishing feature of local2global relative to existing work is that patches are trained independently without the need for the often costly parameter synchronisation during distributed training. This allows local2global to scale to large-scale industrial applications, where the input graph may not even fit into memory and may be stored in a distributed manner. Preliminary results on medium-scale data sets (up to $\sim$7K nodes and $\sim$200K edges) are promising, with a graph reconstruction performance for local2global that is comparable to that of globally trained embeddings. A thorough evaluation of local2global on large scale data and applications to downstream tasks, such as node classification and link prediction, constitutes ongoing work.

* 5 pages, 1 figure, to appear at DLG-KDD '21 
Viaarxiv icon