Alert button
Picture for Richard Zemel

Richard Zemel

Alert button

Flexibly Fair Representation Learning by Disentanglement

Add code
Bookmark button
Alert button
Jun 06, 2019
Elliot Creager, David Madras, Jörn-Henrik Jacobsen, Marissa A. Weis, Kevin Swersky, Toniann Pitassi, Richard Zemel

Figure 1 for Flexibly Fair Representation Learning by Disentanglement
Figure 2 for Flexibly Fair Representation Learning by Disentanglement
Figure 3 for Flexibly Fair Representation Learning by Disentanglement
Figure 4 for Flexibly Fair Representation Learning by Disentanglement
Viaarxiv icon

Conditional Generative Models are not Robust

Add code
Bookmark button
Alert button
Jun 04, 2019
Ethan Fetaya, Jörn-Henrik Jacobsen, Richard Zemel

Figure 1 for Conditional Generative Models are not Robust
Figure 2 for Conditional Generative Models are not Robust
Figure 3 for Conditional Generative Models are not Robust
Figure 4 for Conditional Generative Models are not Robust
Viaarxiv icon

High-Level Perceptual Similarity is Enabled by Learning Diverse Tasks

Add code
Bookmark button
Alert button
Mar 26, 2019
Amir Rosenfeld, Richard Zemel, John K. Tsotsos

Figure 1 for High-Level Perceptual Similarity is Enabled by Learning Diverse Tasks
Figure 2 for High-Level Perceptual Similarity is Enabled by Learning Diverse Tasks
Figure 3 for High-Level Perceptual Similarity is Enabled by Learning Diverse Tasks
Figure 4 for High-Level Perceptual Similarity is Enabled by Learning Diverse Tasks
Viaarxiv icon

Learning Latent Subspaces in Variational Autoencoders

Add code
Bookmark button
Alert button
Dec 14, 2018
Jack Klys, Jake Snell, Richard Zemel

Figure 1 for Learning Latent Subspaces in Variational Autoencoders
Figure 2 for Learning Latent Subspaces in Variational Autoencoders
Figure 3 for Learning Latent Subspaces in Variational Autoencoders
Figure 4 for Learning Latent Subspaces in Variational Autoencoders
Viaarxiv icon

Excessive Invariance Causes Adversarial Vulnerability

Add code
Bookmark button
Alert button
Nov 01, 2018
Jörn-Henrik Jacobsen, Jens Behrmann, Richard Zemel, Matthias Bethge

Figure 1 for Excessive Invariance Causes Adversarial Vulnerability
Figure 2 for Excessive Invariance Causes Adversarial Vulnerability
Figure 3 for Excessive Invariance Causes Adversarial Vulnerability
Figure 4 for Excessive Invariance Causes Adversarial Vulnerability
Viaarxiv icon

Neural Guided Constraint Logic Programming for Program Synthesis

Add code
Bookmark button
Alert button
Oct 26, 2018
Lisa Zhang, Gregory Rosenblatt, Ethan Fetaya, Renjie Liao, William E. Byrd, Matthew Might, Raquel Urtasun, Richard Zemel

Figure 1 for Neural Guided Constraint Logic Programming for Program Synthesis
Figure 2 for Neural Guided Constraint Logic Programming for Program Synthesis
Figure 3 for Neural Guided Constraint Logic Programming for Program Synthesis
Figure 4 for Neural Guided Constraint Logic Programming for Program Synthesis
Viaarxiv icon

Learning Adversarially Fair and Transferable Representations

Add code
Bookmark button
Alert button
Oct 22, 2018
David Madras, Elliot Creager, Toniann Pitassi, Richard Zemel

Figure 1 for Learning Adversarially Fair and Transferable Representations
Figure 2 for Learning Adversarially Fair and Transferable Representations
Figure 3 for Learning Adversarially Fair and Transferable Representations
Figure 4 for Learning Adversarially Fair and Transferable Representations
Viaarxiv icon

Understanding the Origins of Bias in Word Embeddings

Add code
Bookmark button
Alert button
Oct 08, 2018
Marc-Etienne Brunet, Colleen Alkalay-Houlihan, Ashton Anderson, Richard Zemel

Figure 1 for Understanding the Origins of Bias in Word Embeddings
Figure 2 for Understanding the Origins of Bias in Word Embeddings
Figure 3 for Understanding the Origins of Bias in Word Embeddings
Figure 4 for Understanding the Origins of Bias in Word Embeddings
Viaarxiv icon

Fairness Through Causal Awareness: Learning Latent-Variable Models for Biased Data

Add code
Bookmark button
Alert button
Sep 10, 2018
David Madras, Elliot Creager, Toniann Pitassi, Richard Zemel

Figure 1 for Fairness Through Causal Awareness: Learning Latent-Variable Models for Biased Data
Figure 2 for Fairness Through Causal Awareness: Learning Latent-Variable Models for Biased Data
Figure 3 for Fairness Through Causal Awareness: Learning Latent-Variable Models for Biased Data
Figure 4 for Fairness Through Causal Awareness: Learning Latent-Variable Models for Biased Data
Viaarxiv icon

Predict Responsibly: Improving Fairness and Accuracy by Learning to Defer

Add code
Bookmark button
Alert button
Sep 07, 2018
David Madras, Toniann Pitassi, Richard Zemel

Figure 1 for Predict Responsibly: Improving Fairness and Accuracy by Learning to Defer
Figure 2 for Predict Responsibly: Improving Fairness and Accuracy by Learning to Defer
Figure 3 for Predict Responsibly: Improving Fairness and Accuracy by Learning to Defer
Figure 4 for Predict Responsibly: Improving Fairness and Accuracy by Learning to Defer
Viaarxiv icon