Alert button
Picture for Gal Kaplun

Gal Kaplun

Alert button

Corgi^2: A Hybrid Offline-Online Approach To Storage-Aware Data Shuffling For SGD

Sep 04, 2023
Etay Livne, Gal Kaplun, Eran Malach, Shai Shalev-Schwatz

Figure 1 for Corgi^2: A Hybrid Offline-Online Approach To Storage-Aware Data Shuffling For SGD
Figure 2 for Corgi^2: A Hybrid Offline-Online Approach To Storage-Aware Data Shuffling For SGD
Figure 3 for Corgi^2: A Hybrid Offline-Online Approach To Storage-Aware Data Shuffling For SGD
Figure 4 for Corgi^2: A Hybrid Offline-Online Approach To Storage-Aware Data Shuffling For SGD
Viaarxiv icon

Beyond Implicit Bias: The Insignificance of SGD Noise in Online Learning

Jun 14, 2023
Nikhil Vyas, Depen Morwani, Rosie Zhao, Gal Kaplun, Sham Kakade, Boaz Barak

Figure 1 for Beyond Implicit Bias: The Insignificance of SGD Noise in Online Learning
Figure 2 for Beyond Implicit Bias: The Insignificance of SGD Noise in Online Learning
Figure 3 for Beyond Implicit Bias: The Insignificance of SGD Noise in Online Learning
Figure 4 for Beyond Implicit Bias: The Insignificance of SGD Noise in Online Learning
Viaarxiv icon

SubTuning: Efficient Finetuning for Multi-Task Learning

Feb 14, 2023
Gal Kaplun, Andrey Gurevich, Tal Swisa, Mazor David, Shai Shalev-Shwartz, Eran Malach

Figure 1 for SubTuning: Efficient Finetuning for Multi-Task Learning
Figure 2 for SubTuning: Efficient Finetuning for Multi-Task Learning
Figure 3 for SubTuning: Efficient Finetuning for Multi-Task Learning
Figure 4 for SubTuning: Efficient Finetuning for Multi-Task Learning
Viaarxiv icon

Knowledge Distillation: Bad Models Can Be Good Role Models

Mar 28, 2022
Gal Kaplun, Eran Malach, Preetum Nakkiran, Shai Shalev-Shwartz

Figure 1 for Knowledge Distillation: Bad Models Can Be Good Role Models
Figure 2 for Knowledge Distillation: Bad Models Can Be Good Role Models
Viaarxiv icon

Deconstructing Distributions: A Pointwise Framework of Learning

Feb 20, 2022
Gal Kaplun, Nikhil Ghosh, Saurabh Garg, Boaz Barak, Preetum Nakkiran

Figure 1 for Deconstructing Distributions: A Pointwise Framework of Learning
Figure 2 for Deconstructing Distributions: A Pointwise Framework of Learning
Figure 3 for Deconstructing Distributions: A Pointwise Framework of Learning
Figure 4 for Deconstructing Distributions: A Pointwise Framework of Learning
Viaarxiv icon

For Manifold Learning, Deep Neural Networks can be Locality Sensitive Hash Functions

Mar 11, 2021
Nishanth Dikkala, Gal Kaplun, Rina Panigrahy

Figure 1 for For Manifold Learning, Deep Neural Networks can be Locality Sensitive Hash Functions
Figure 2 for For Manifold Learning, Deep Neural Networks can be Locality Sensitive Hash Functions
Figure 3 for For Manifold Learning, Deep Neural Networks can be Locality Sensitive Hash Functions
Figure 4 for For Manifold Learning, Deep Neural Networks can be Locality Sensitive Hash Functions
Viaarxiv icon

For self-supervised learning, Rationality implies generalization, provably

Oct 16, 2020
Yamini Bansal, Gal Kaplun, Boaz Barak

Figure 1 for For self-supervised learning, Rationality implies generalization, provably
Figure 2 for For self-supervised learning, Rationality implies generalization, provably
Figure 3 for For self-supervised learning, Rationality implies generalization, provably
Figure 4 for For self-supervised learning, Rationality implies generalization, provably
Viaarxiv icon

Robustness from Simple Classifiers

Feb 21, 2020
Sharon Qian, Dimitris Kalimeris, Gal Kaplun, Yaron Singer

Figure 1 for Robustness from Simple Classifiers
Figure 2 for Robustness from Simple Classifiers
Figure 3 for Robustness from Simple Classifiers
Figure 4 for Robustness from Simple Classifiers
Viaarxiv icon

Deep Double Descent: Where Bigger Models and More Data Hurt

Dec 04, 2019
Preetum Nakkiran, Gal Kaplun, Yamini Bansal, Tristan Yang, Boaz Barak, Ilya Sutskever

Figure 1 for Deep Double Descent: Where Bigger Models and More Data Hurt
Figure 2 for Deep Double Descent: Where Bigger Models and More Data Hurt
Figure 3 for Deep Double Descent: Where Bigger Models and More Data Hurt
Figure 4 for Deep Double Descent: Where Bigger Models and More Data Hurt
Viaarxiv icon

SGD on Neural Networks Learns Functions of Increasing Complexity

May 28, 2019
Preetum Nakkiran, Gal Kaplun, Dimitris Kalimeris, Tristan Yang, Benjamin L. Edelman, Fred Zhang, Boaz Barak

Figure 1 for SGD on Neural Networks Learns Functions of Increasing Complexity
Figure 2 for SGD on Neural Networks Learns Functions of Increasing Complexity
Figure 3 for SGD on Neural Networks Learns Functions of Increasing Complexity
Figure 4 for SGD on Neural Networks Learns Functions of Increasing Complexity
Viaarxiv icon