Alert button
Picture for Sean Donegan

Sean Donegan

Alert button

A Lightweight Transformer for Faster and Robust EBSD Data Collection

Aug 18, 2023
Harry Dong, Sean Donegan, Megna Shah, Yuejie Chi

Three dimensional electron back-scattered diffraction (EBSD) microscopy is a critical tool in many applications in materials science, yet its data quality can fluctuate greatly during the arduous collection process, particularly via serial-sectioning. Fortunately, 3D EBSD data is inherently sequential, opening up the opportunity to use transformers, state-of-the-art deep learning architectures that have made breakthroughs in a plethora of domains, for data processing and recovery. To be more robust to errors and accelerate this 3D EBSD data collection, we introduce a two step method that recovers missing slices in an 3D EBSD volume, using an efficient transformer model and a projection algorithm to process the transformer's outputs. Overcoming the computational and practical hurdles of deep learning with scarce high dimensional data, we train this model using only synthetic 3D EBSD data with self-supervision and obtain superior recovery accuracy on real 3D EBSD data, compared to existing methods.

Viaarxiv icon

Deep Unfolded Tensor Robust PCA with Self-supervised Learning

Dec 21, 2022
Harry Dong, Megna Shah, Sean Donegan, Yuejie Chi

Figure 1 for Deep Unfolded Tensor Robust PCA with Self-supervised Learning
Figure 2 for Deep Unfolded Tensor Robust PCA with Self-supervised Learning
Figure 3 for Deep Unfolded Tensor Robust PCA with Self-supervised Learning
Figure 4 for Deep Unfolded Tensor Robust PCA with Self-supervised Learning

Tensor robust principal component analysis (RPCA), which seeks to separate a low-rank tensor from its sparse corruptions, has been crucial in data science and machine learning where tensor structures are becoming more prevalent. While powerful, existing tensor RPCA algorithms can be difficult to use in practice, as their performance can be sensitive to the choice of additional hyperparameters, which are not straightforward to tune. In this paper, we describe a fast and simple self-supervised model for tensor RPCA using deep unfolding by only learning four hyperparameters. Despite its simplicity, our model expunges the need for ground truth labels while maintaining competitive or even greater performance compared to supervised deep unfolding. Furthermore, our model is capable of operating in extreme data-starved scenarios. We demonstrate these claims on a mix of synthetic data and real-world tasks, comparing performance against previously studied supervised deep unfolding methods and Bayesian optimization baselines.

Viaarxiv icon