Alert button
Picture for Yann N. Dauphin

Yann N. Dauphin

Alert button

Neglected Hessian component explains mysteries in Sharpness regularization

Add code
Bookmark button
Alert button
Jan 24, 2024
Yann N. Dauphin, Atish Agarwala, Hossein Mobahi

Viaarxiv icon

Has the Machine Learning Review Process Become More Arbitrary as the Field Has Grown? The NeurIPS 2021 Consistency Experiment

Add code
Bookmark button
Alert button
Jun 05, 2023
Alina Beygelzimer, Yann N. Dauphin, Percy Liang, Jennifer Wortman Vaughan

Figure 1 for Has the Machine Learning Review Process Become More Arbitrary as the Field Has Grown? The NeurIPS 2021 Consistency Experiment
Figure 2 for Has the Machine Learning Review Process Become More Arbitrary as the Field Has Grown? The NeurIPS 2021 Consistency Experiment
Figure 3 for Has the Machine Learning Review Process Become More Arbitrary as the Field Has Grown? The NeurIPS 2021 Consistency Experiment
Figure 4 for Has the Machine Learning Review Process Become More Arbitrary as the Field Has Grown? The NeurIPS 2021 Consistency Experiment
Viaarxiv icon

SAM operates far from home: eigenvalue regularization as a dynamical phenomenon

Add code
Bookmark button
Alert button
Feb 17, 2023
Atish Agarwala, Yann N. Dauphin

Figure 1 for SAM operates far from home: eigenvalue regularization as a dynamical phenomenon
Figure 2 for SAM operates far from home: eigenvalue regularization as a dynamical phenomenon
Figure 3 for SAM operates far from home: eigenvalue regularization as a dynamical phenomenon
Figure 4 for SAM operates far from home: eigenvalue regularization as a dynamical phenomenon
Viaarxiv icon

How do Authors' Perceptions of their Papers Compare with Co-authors' Perceptions and Peer-review Decisions?

Add code
Bookmark button
Alert button
Nov 22, 2022
Charvi Rastogi, Ivan Stelmakh, Alina Beygelzimer, Yann N. Dauphin, Percy Liang, Jennifer Wortman Vaughan, Zhenyu Xue, Hal Daumé III, Emma Pierson, Nihar B. Shah

Figure 1 for How do Authors' Perceptions of their Papers Compare with Co-authors' Perceptions and Peer-review Decisions?
Figure 2 for How do Authors' Perceptions of their Papers Compare with Co-authors' Perceptions and Peer-review Decisions?
Figure 3 for How do Authors' Perceptions of their Papers Compare with Co-authors' Perceptions and Peer-review Decisions?
Figure 4 for How do Authors' Perceptions of their Papers Compare with Co-authors' Perceptions and Peer-review Decisions?
Viaarxiv icon

Simple and Effective Noisy Channel Modeling for Neural Machine Translation

Add code
Bookmark button
Alert button
Aug 15, 2019
Kyra Yee, Nathan Ng, Yann N. Dauphin, Michael Auli

Figure 1 for Simple and Effective Noisy Channel Modeling for Neural Machine Translation
Figure 2 for Simple and Effective Noisy Channel Modeling for Neural Machine Translation
Figure 3 for Simple and Effective Noisy Channel Modeling for Neural Machine Translation
Figure 4 for Simple and Effective Noisy Channel Modeling for Neural Machine Translation
Viaarxiv icon

Pay Less Attention with Lightweight and Dynamic Convolutions

Add code
Bookmark button
Alert button
Jan 29, 2019
Felix Wu, Angela Fan, Alexei Baevski, Yann N. Dauphin, Michael Auli

Figure 1 for Pay Less Attention with Lightweight and Dynamic Convolutions
Figure 2 for Pay Less Attention with Lightweight and Dynamic Convolutions
Figure 3 for Pay Less Attention with Lightweight and Dynamic Convolutions
Figure 4 for Pay Less Attention with Lightweight and Dynamic Convolutions
Viaarxiv icon

Fixup Initialization: Residual Learning Without Normalization

Add code
Bookmark button
Alert button
Jan 27, 2019
Hongyi Zhang, Yann N. Dauphin, Tengyu Ma

Figure 1 for Fixup Initialization: Residual Learning Without Normalization
Figure 2 for Fixup Initialization: Residual Learning Without Normalization
Figure 3 for Fixup Initialization: Residual Learning Without Normalization
Figure 4 for Fixup Initialization: Residual Learning Without Normalization
Viaarxiv icon

mixup: Beyond Empirical Risk Minimization

Add code
Bookmark button
Alert button
Apr 27, 2018
Hongyi Zhang, Moustapha Cisse, Yann N. Dauphin, David Lopez-Paz

Figure 1 for mixup: Beyond Empirical Risk Minimization
Figure 2 for mixup: Beyond Empirical Risk Minimization
Figure 3 for mixup: Beyond Empirical Risk Minimization
Figure 4 for mixup: Beyond Empirical Risk Minimization
Viaarxiv icon

Language Modeling with Gated Convolutional Networks

Add code
Bookmark button
Alert button
Sep 08, 2017
Yann N. Dauphin, Angela Fan, Michael Auli, David Grangier

Figure 1 for Language Modeling with Gated Convolutional Networks
Figure 2 for Language Modeling with Gated Convolutional Networks
Figure 3 for Language Modeling with Gated Convolutional Networks
Figure 4 for Language Modeling with Gated Convolutional Networks
Viaarxiv icon

Convolutional Sequence to Sequence Learning

Add code
Bookmark button
Alert button
Jul 25, 2017
Jonas Gehring, Michael Auli, David Grangier, Denis Yarats, Yann N. Dauphin

Figure 1 for Convolutional Sequence to Sequence Learning
Figure 2 for Convolutional Sequence to Sequence Learning
Figure 3 for Convolutional Sequence to Sequence Learning
Figure 4 for Convolutional Sequence to Sequence Learning
Viaarxiv icon