Machine learning algorithms are widely used in data science applications and have significant potential to improve predictions and understanding of social scientific processes. In many applications researchers need to be able to explain why the model made one prediction over another. In this workshop we introduce several techniques for interpreting black box models using model-agnostic techniques.
Machine learning algorithms are widely used in data science applications and have significant potential to improve predictions and understanding of social scientific processes. However machine learning models generally do not explain their predictions – they simply seek to minimize some loss function and provide for a given observation the probability of an event occurring. In many applications researchers need to be able to explain why the model made one prediction over another. This emphasis on interpretability and explanation is directly relevant to many social scientific questions, and can provide necessary context for decision makers who need to use machine learning models but lack a strong technical background. In this workshop we introduce several techniques for interpreting black box models using model-agnostic techniques.
This workshop is designed for individuals with introductory-to-intermediate knowledge of machine learning algorithms, as well as experience training machine learning models using R. Prior experience with tidymodels
is helpful, but not required.
Room 295 in 1155 E 60th St.
DALEX
which outlines the intuition and methodology of many of the interpretation/explanation methods we discuss in the workshop. Also includes code examples in R and Python.iml
, an alternative package for interpreting and explaining models in R using model-agnostic methods.tidymodels
workflow.
If you see mistakes or want to suggest changes, please create an issue on the source repository.
Text and figures are licensed under Creative Commons Attribution CC BY-NC 4.0. Source code is available at https://github.com/css-skills/interpretable-ml, unless otherwise noted. The figures that have been reused from other sources don't fall under this license and can be recognized by a note in their caption: "Figure from ...".
For attribution, please cite this work as
Soltoff (2022, Feb. 15). Computation Skills Workshop: Interpreting and explaining machine learning models. Retrieved from https://css-skills.uchicago.edu/posts/2022-02-15-interpretable-machine-learning-methods/
BibTeX citation
@misc{soltoff2022interpreting, author = {Soltoff, Benjamin}, title = {Computation Skills Workshop: Interpreting and explaining machine learning models}, url = {https://css-skills.uchicago.edu/posts/2022-02-15-interpretable-machine-learning-methods/}, year = {2022} }