Alert button
Picture for Michael W. Mahoney

Michael W. Mahoney

Alert button

Heavy-Tailed Universality Predicts Trends in Test Accuracies for Very Large Pre-Trained Deep Neural Networks

Add code
Bookmark button
Alert button
Jan 24, 2019
Charles H. Martin, Michael W. Mahoney

Figure 1 for Heavy-Tailed Universality Predicts Trends in Test Accuracies for Very Large Pre-Trained Deep Neural Networks
Figure 2 for Heavy-Tailed Universality Predicts Trends in Test Accuracies for Very Large Pre-Trained Deep Neural Networks
Figure 3 for Heavy-Tailed Universality Predicts Trends in Test Accuracies for Very Large Pre-Trained Deep Neural Networks
Figure 4 for Heavy-Tailed Universality Predicts Trends in Test Accuracies for Very Large Pre-Trained Deep Neural Networks
Viaarxiv icon

Traditional and Heavy-Tailed Self Regularization in Neural Network Models

Add code
Bookmark button
Alert button
Jan 24, 2019
Charles H. Martin, Michael W. Mahoney

Figure 1 for Traditional and Heavy-Tailed Self Regularization in Neural Network Models
Figure 2 for Traditional and Heavy-Tailed Self Regularization in Neural Network Models
Figure 3 for Traditional and Heavy-Tailed Self Regularization in Neural Network Models
Figure 4 for Traditional and Heavy-Tailed Self Regularization in Neural Network Models
Viaarxiv icon

On the Computational Inefficiency of Large Batch Sizes for Stochastic Gradient Descent

Add code
Bookmark button
Alert button
Nov 30, 2018
Noah Golmant, Nikita Vemuri, Zhewei Yao, Vladimir Feinberg, Amir Gholami, Kai Rothauge, Michael W. Mahoney, Joseph Gonzalez

Figure 1 for On the Computational Inefficiency of Large Batch Sizes for Stochastic Gradient Descent
Figure 2 for On the Computational Inefficiency of Large Batch Sizes for Stochastic Gradient Descent
Figure 3 for On the Computational Inefficiency of Large Batch Sizes for Stochastic Gradient Descent
Figure 4 for On the Computational Inefficiency of Large Batch Sizes for Stochastic Gradient Descent
Viaarxiv icon

Implicit Self-Regularization in Deep Neural Networks: Evidence from Random Matrix Theory and Implications for Learning

Add code
Bookmark button
Alert button
Oct 02, 2018
Charles H. Martin, Michael W. Mahoney

Figure 1 for Implicit Self-Regularization in Deep Neural Networks: Evidence from Random Matrix Theory and Implications for Learning
Figure 2 for Implicit Self-Regularization in Deep Neural Networks: Evidence from Random Matrix Theory and Implications for Learning
Figure 3 for Implicit Self-Regularization in Deep Neural Networks: Evidence from Random Matrix Theory and Implications for Learning
Figure 4 for Implicit Self-Regularization in Deep Neural Networks: Evidence from Random Matrix Theory and Implications for Learning
Viaarxiv icon

Newton-MR: Newton's Method Without Smoothness or Convexity

Add code
Bookmark button
Alert button
Sep 30, 2018
Fred Roosta, Yang Liu, Peng Xu, Michael W. Mahoney

Figure 1 for Newton-MR: Newton's Method Without Smoothness or Convexity
Figure 2 for Newton-MR: Newton's Method Without Smoothness or Convexity
Figure 3 for Newton-MR: Newton's Method Without Smoothness or Convexity
Figure 4 for Newton-MR: Newton's Method Without Smoothness or Convexity
Viaarxiv icon

Block Basis Factorization for Scalable Kernel Matrix Evaluation

Add code
Bookmark button
Alert button
Sep 12, 2018
Ruoxi Wang, Yingzhou Li, Michael W. Mahoney, Eric Darve

Figure 1 for Block Basis Factorization for Scalable Kernel Matrix Evaluation
Figure 2 for Block Basis Factorization for Scalable Kernel Matrix Evaluation
Figure 3 for Block Basis Factorization for Scalable Kernel Matrix Evaluation
Figure 4 for Block Basis Factorization for Scalable Kernel Matrix Evaluation
Viaarxiv icon

GIANT: Globally Improved Approximate Newton Method for Distributed Optimization

Add code
Bookmark button
Alert button
Sep 11, 2018
Shusen Wang, Farbod Roosta-Khorasani, Peng Xu, Michael W. Mahoney

Figure 1 for GIANT: Globally Improved Approximate Newton Method for Distributed Optimization
Figure 2 for GIANT: Globally Improved Approximate Newton Method for Distributed Optimization
Viaarxiv icon

Error Estimation for Randomized Least-Squares Algorithms via the Bootstrap

Add code
Bookmark button
Alert button
Sep 06, 2018
Miles E. Lopes, Shusen Wang, Michael W. Mahoney

Figure 1 for Error Estimation for Randomized Least-Squares Algorithms via the Bootstrap
Figure 2 for Error Estimation for Randomized Least-Squares Algorithms via the Bootstrap
Viaarxiv icon

Distributed Second-order Convex Optimization

Add code
Bookmark button
Alert button
Jul 18, 2018
Chih-Hao Fang, Sudhir B Kylasa, Farbod Roosta-Khorasani, Michael W. Mahoney, Ananth Grama

Figure 1 for Distributed Second-order Convex Optimization
Figure 2 for Distributed Second-order Convex Optimization
Figure 3 for Distributed Second-order Convex Optimization
Figure 4 for Distributed Second-order Convex Optimization
Viaarxiv icon

Hessian-based Analysis of Large Batch Training and Robustness to Adversaries

Add code
Bookmark button
Alert button
Jun 18, 2018
Zhewei Yao, Amir Gholami, Qi Lei, Kurt Keutzer, Michael W. Mahoney

Figure 1 for Hessian-based Analysis of Large Batch Training and Robustness to Adversaries
Figure 2 for Hessian-based Analysis of Large Batch Training and Robustness to Adversaries
Figure 3 for Hessian-based Analysis of Large Batch Training and Robustness to Adversaries
Figure 4 for Hessian-based Analysis of Large Batch Training and Robustness to Adversaries
Viaarxiv icon