Picture for Rayid Ghani

Rayid Ghani

Breaking the Cycle of Incarceration With Targeted Mental Health Outreach: A Case Study in Machine Learning for Public Policy

Add code
Sep 17, 2025
Viaarxiv icon

Towards Automated Scoping of AI for Social Good Projects

Add code
Apr 28, 2025
Viaarxiv icon

Aequitas Flow: Streamlining Fair ML Experimentation

Add code
May 09, 2024
Viaarxiv icon

Preventing Eviction-Caused Homelessness through ML-Informed Distribution of Rental Assistance

Add code
Mar 19, 2024
Figure 1 for Preventing Eviction-Caused Homelessness through ML-Informed Distribution of Rental Assistance
Figure 2 for Preventing Eviction-Caused Homelessness through ML-Informed Distribution of Rental Assistance
Figure 3 for Preventing Eviction-Caused Homelessness through ML-Informed Distribution of Rental Assistance
Figure 4 for Preventing Eviction-Caused Homelessness through ML-Informed Distribution of Rental Assistance
Viaarxiv icon

Toward Operationalizing Pipeline-aware ML Fairness: A Research Agenda for Developing Practical Guidelines and Tools

Add code
Sep 29, 2023
Figure 1 for Toward Operationalizing Pipeline-aware ML Fairness: A Research Agenda for Developing Practical Guidelines and Tools
Figure 2 for Toward Operationalizing Pipeline-aware ML Fairness: A Research Agenda for Developing Practical Guidelines and Tools
Figure 3 for Toward Operationalizing Pipeline-aware ML Fairness: A Research Agenda for Developing Practical Guidelines and Tools
Figure 4 for Toward Operationalizing Pipeline-aware ML Fairness: A Research Agenda for Developing Practical Guidelines and Tools
Viaarxiv icon

A Conceptual Framework for Using Machine Learning to Support Child Welfare Decisions

Add code
Jul 12, 2022
Figure 1 for A Conceptual Framework for Using Machine Learning to Support Child Welfare Decisions
Viaarxiv icon

On the Importance of Application-Grounded Experimental Design for Evaluating Explainable ML Methods

Add code
Jun 30, 2022
Figure 1 for On the Importance of Application-Grounded Experimental Design for Evaluating Explainable ML Methods
Figure 2 for On the Importance of Application-Grounded Experimental Design for Evaluating Explainable ML Methods
Figure 3 for On the Importance of Application-Grounded Experimental Design for Evaluating Explainable ML Methods
Figure 4 for On the Importance of Application-Grounded Experimental Design for Evaluating Explainable ML Methods
Viaarxiv icon

Faking feature importance: A cautionary tale on the use of differentially-private synthetic data

Add code
Mar 02, 2022
Figure 1 for Faking feature importance: A cautionary tale on the use of differentially-private synthetic data
Figure 2 for Faking feature importance: A cautionary tale on the use of differentially-private synthetic data
Figure 3 for Faking feature importance: A cautionary tale on the use of differentially-private synthetic data
Figure 4 for Faking feature importance: A cautionary tale on the use of differentially-private synthetic data
Viaarxiv icon

An Empirical Comparison of Bias Reduction Methods on Real-World Problems in High-Stakes Policy Settings

Add code
May 13, 2021
Figure 1 for An Empirical Comparison of Bias Reduction Methods on Real-World Problems in High-Stakes Policy Settings
Figure 2 for An Empirical Comparison of Bias Reduction Methods on Real-World Problems in High-Stakes Policy Settings
Figure 3 for An Empirical Comparison of Bias Reduction Methods on Real-World Problems in High-Stakes Policy Settings
Figure 4 for An Empirical Comparison of Bias Reduction Methods on Real-World Problems in High-Stakes Policy Settings
Viaarxiv icon

Machine learning for public policy: Do we need to sacrifice accuracy to make models fair?

Add code
Dec 05, 2020
Figure 1 for Machine learning for public policy: Do we need to sacrifice accuracy to make models fair?
Figure 2 for Machine learning for public policy: Do we need to sacrifice accuracy to make models fair?
Figure 3 for Machine learning for public policy: Do we need to sacrifice accuracy to make models fair?
Figure 4 for Machine learning for public policy: Do we need to sacrifice accuracy to make models fair?
Viaarxiv icon