Picture for Mikhail Yurochkin

Mikhail Yurochkin

Understanding new tasks through the lens of training data via exponential tilting

Add code
May 26, 2022
Figure 1 for Understanding new tasks through the lens of training data via exponential tilting
Figure 2 for Understanding new tasks through the lens of training data via exponential tilting
Figure 3 for Understanding new tasks through the lens of training data via exponential tilting
Figure 4 for Understanding new tasks through the lens of training data via exponential tilting
Viaarxiv icon

Domain Adaptation meets Individual Fairness. And they get along

Add code
May 01, 2022
Figure 1 for Domain Adaptation meets Individual Fairness. And they get along
Figure 2 for Domain Adaptation meets Individual Fairness. And they get along
Viaarxiv icon

Log-Euclidean Signatures for Intrinsic Distances Between Unaligned Datasets

Add code
Feb 03, 2022
Figure 1 for Log-Euclidean Signatures for Intrinsic Distances Between Unaligned Datasets
Figure 2 for Log-Euclidean Signatures for Intrinsic Distances Between Unaligned Datasets
Figure 3 for Log-Euclidean Signatures for Intrinsic Distances Between Unaligned Datasets
Figure 4 for Log-Euclidean Signatures for Intrinsic Distances Between Unaligned Datasets
Viaarxiv icon

Rewiring with Positional Encodings for Graph Neural Networks

Add code
Feb 02, 2022
Figure 1 for Rewiring with Positional Encodings for Graph Neural Networks
Figure 2 for Rewiring with Positional Encodings for Graph Neural Networks
Figure 3 for Rewiring with Positional Encodings for Graph Neural Networks
Figure 4 for Rewiring with Positional Encodings for Graph Neural Networks
Viaarxiv icon

Learning Proximal Operators to Discover Multiple Optima

Add code
Jan 28, 2022
Figure 1 for Learning Proximal Operators to Discover Multiple Optima
Figure 2 for Learning Proximal Operators to Discover Multiple Optima
Figure 3 for Learning Proximal Operators to Discover Multiple Optima
Figure 4 for Learning Proximal Operators to Discover Multiple Optima
Viaarxiv icon

On sensitivity of meta-learning to support data

Add code
Oct 26, 2021
Figure 1 for On sensitivity of meta-learning to support data
Figure 2 for On sensitivity of meta-learning to support data
Figure 3 for On sensitivity of meta-learning to support data
Figure 4 for On sensitivity of meta-learning to support data
Viaarxiv icon

Post-processing for Individual Fairness

Add code
Oct 26, 2021
Figure 1 for Post-processing for Individual Fairness
Figure 2 for Post-processing for Individual Fairness
Figure 3 for Post-processing for Individual Fairness
Figure 4 for Post-processing for Individual Fairness
Viaarxiv icon

Your fairness may vary: Group fairness of pretrained language models in toxic text classification

Add code
Aug 03, 2021
Figure 1 for Your fairness may vary: Group fairness of pretrained language models in toxic text classification
Figure 2 for Your fairness may vary: Group fairness of pretrained language models in toxic text classification
Figure 3 for Your fairness may vary: Group fairness of pretrained language models in toxic text classification
Figure 4 for Your fairness may vary: Group fairness of pretrained language models in toxic text classification
Viaarxiv icon

Measuring the sensitivity of Gaussian processes to kernel choice

Add code
Jun 11, 2021
Figure 1 for Measuring the sensitivity of Gaussian processes to kernel choice
Figure 2 for Measuring the sensitivity of Gaussian processes to kernel choice
Figure 3 for Measuring the sensitivity of Gaussian processes to kernel choice
Figure 4 for Measuring the sensitivity of Gaussian processes to kernel choice
Viaarxiv icon

k-Mixup Regularization for Deep Learning via Optimal Transport

Add code
Jun 05, 2021
Figure 1 for k-Mixup Regularization for Deep Learning via Optimal Transport
Figure 2 for k-Mixup Regularization for Deep Learning via Optimal Transport
Figure 3 for k-Mixup Regularization for Deep Learning via Optimal Transport
Figure 4 for k-Mixup Regularization for Deep Learning via Optimal Transport
Viaarxiv icon