Alert button
Picture for Christoph Molnar

Christoph Molnar

Alert button

Scientific Inference With Interpretable Machine Learning: Analyzing Models to Learn About Real-World Phenomena

Add code
Bookmark button
Alert button
Jun 11, 2022
Timo Freiesleben, Gunnar König, Christoph Molnar, Alvaro Tejero-Cantero

Figure 1 for Scientific Inference With Interpretable Machine Learning: Analyzing Models to Learn About Real-World Phenomena
Figure 2 for Scientific Inference With Interpretable Machine Learning: Analyzing Models to Learn About Real-World Phenomena
Figure 3 for Scientific Inference With Interpretable Machine Learning: Analyzing Models to Learn About Real-World Phenomena
Figure 4 for Scientific Inference With Interpretable Machine Learning: Analyzing Models to Learn About Real-World Phenomena
Viaarxiv icon

Marginal Effects for Non-Linear Prediction Functions

Add code
Bookmark button
Alert button
Jan 21, 2022
Christian A. Scholbeck, Giuseppe Casalicchio, Christoph Molnar, Bernd Bischl, Christian Heumann

Figure 1 for Marginal Effects for Non-Linear Prediction Functions
Figure 2 for Marginal Effects for Non-Linear Prediction Functions
Figure 3 for Marginal Effects for Non-Linear Prediction Functions
Figure 4 for Marginal Effects for Non-Linear Prediction Functions
Viaarxiv icon

Relating the Partial Dependence Plot and Permutation Feature Importance to the Data Generating Process

Add code
Bookmark button
Alert button
Sep 03, 2021
Christoph Molnar, Timo Freiesleben, Gunnar König, Giuseppe Casalicchio, Marvin N. Wright, Bernd Bischl

Figure 1 for Relating the Partial Dependence Plot and Permutation Feature Importance to the Data Generating Process
Figure 2 for Relating the Partial Dependence Plot and Permutation Feature Importance to the Data Generating Process
Figure 3 for Relating the Partial Dependence Plot and Permutation Feature Importance to the Data Generating Process
Figure 4 for Relating the Partial Dependence Plot and Permutation Feature Importance to the Data Generating Process
Viaarxiv icon

Interpretable Machine Learning -- A Brief History, State-of-the-Art and Challenges

Add code
Bookmark button
Alert button
Oct 19, 2020
Christoph Molnar, Giuseppe Casalicchio, Bernd Bischl

Figure 1 for Interpretable Machine Learning -- A Brief History, State-of-the-Art and Challenges
Figure 2 for Interpretable Machine Learning -- A Brief History, State-of-the-Art and Challenges
Viaarxiv icon

Relative Feature Importance

Add code
Bookmark button
Alert button
Jul 16, 2020
Gunnar König, Christoph Molnar, Bernd Bischl, Moritz Grosse-Wentrup

Figure 1 for Relative Feature Importance
Figure 2 for Relative Feature Importance
Figure 3 for Relative Feature Importance
Figure 4 for Relative Feature Importance
Viaarxiv icon

Pitfalls to Avoid when Interpreting Machine Learning Models

Add code
Bookmark button
Alert button
Jul 08, 2020
Christoph Molnar, Gunnar König, Julia Herbinger, Timo Freiesleben, Susanne Dandl, Christian A. Scholbeck, Giuseppe Casalicchio, Moritz Grosse-Wentrup, Bernd Bischl

Figure 1 for Pitfalls to Avoid when Interpreting Machine Learning Models
Figure 2 for Pitfalls to Avoid when Interpreting Machine Learning Models
Figure 3 for Pitfalls to Avoid when Interpreting Machine Learning Models
Figure 4 for Pitfalls to Avoid when Interpreting Machine Learning Models
Viaarxiv icon

Model-agnostic Feature Importance and Effects with Dependent Features -- A Conditional Subgroup Approach

Add code
Bookmark button
Alert button
Jun 08, 2020
Christoph Molnar, Gunnar König, Bernd Bischl, Giuseppe Casalicchio

Figure 1 for Model-agnostic Feature Importance and Effects with Dependent Features -- A Conditional Subgroup Approach
Figure 2 for Model-agnostic Feature Importance and Effects with Dependent Features -- A Conditional Subgroup Approach
Figure 3 for Model-agnostic Feature Importance and Effects with Dependent Features -- A Conditional Subgroup Approach
Figure 4 for Model-agnostic Feature Importance and Effects with Dependent Features -- A Conditional Subgroup Approach
Viaarxiv icon

Multi-Objective Counterfactual Explanations

Add code
Bookmark button
Alert button
Apr 23, 2020
Susanne Dandl, Christoph Molnar, Martin Binder, Bernd Bischl

Figure 1 for Multi-Objective Counterfactual Explanations
Figure 2 for Multi-Objective Counterfactual Explanations
Figure 3 for Multi-Objective Counterfactual Explanations
Figure 4 for Multi-Objective Counterfactual Explanations
Viaarxiv icon

Sampling, Intervention, Prediction, Aggregation: A Generalized Framework for Model Agnostic Interpretations

Add code
Bookmark button
Alert button
Apr 08, 2019
Christian A. Scholbeck, Christoph Molnar, Christian Heumann, Bernd Bischl, Giuseppe Casalicchio

Figure 1 for Sampling, Intervention, Prediction, Aggregation: A Generalized Framework for Model Agnostic Interpretations
Figure 2 for Sampling, Intervention, Prediction, Aggregation: A Generalized Framework for Model Agnostic Interpretations
Viaarxiv icon

Quantifying Interpretability of Arbitrary Machine Learning Models Through Functional Decomposition

Add code
Bookmark button
Alert button
Apr 08, 2019
Christoph Molnar, Giuseppe Casalicchio, Bernd Bischl

Figure 1 for Quantifying Interpretability of Arbitrary Machine Learning Models Through Functional Decomposition
Figure 2 for Quantifying Interpretability of Arbitrary Machine Learning Models Through Functional Decomposition
Figure 3 for Quantifying Interpretability of Arbitrary Machine Learning Models Through Functional Decomposition
Figure 4 for Quantifying Interpretability of Arbitrary Machine Learning Models Through Functional Decomposition
Viaarxiv icon