Intelligible ML

Building glass-box models that are both predictive and interpretable

For many reasons (e.g. scientific inquiry, high-stakes decision making), we need AI systems that are both accurate and intelligible.

We find that interaction effects are often a useful lens through which to view intelligibility. Interaction effects (Lengerich et al., 2020) are effects which require two input components to know anything about the output (one component alone tells you nothing). Since humans reason by chunking and hierarchical logic, we struggle to understand interactions of multiple variables. If we can instead represent effects as additive (non-interactive) combinations of components, we can understand the components independently and reason about even very complex concepts.

Toward this end, we have designed new architectures including deep additive models (Agarwal et al., 2021) and contextualized additive models (Lengerich et al., 2022), studied deep learning theory through the lens of interaction effects (Lengerich et al., 2022), studied additive models and identifiability (Chang et al., 2021), (Lengerich et al., 2020), and applied intelligible models to real-world evidence (Lengerich et al., 2024).



References

2024

  1. Interpretable Machine Learning Predicts Postpartum Hemorrhage with Severe Maternal Morbidity in a Lower Risk Laboring Obstetric Population
    Benjamin J LengerichRich Caruana, Ian Painter, and 3 more authors
    American Journal of Obstetrics & Gynecology MFM, 2024

2022

  1. Automated interpretable discovery of heterogeneous treatment effectiveness: A COVID-19 case study
    Benjamin J Lengerich, Mark E Nunnally, Yin Aphinyanaphongs, and 2 more authors
    Journal of biomedical informatics, 2022
  2. Dropout as a Regularizer of Interaction Effects
    In Proceedings of the Twenty Fifth International Conference on Artificial Intelligence and Statistics , 2022

2021

  1. Neural Additive Models: Interpretable Machine Learning with Neural Nets
    Rishabh Agarwal, Levi Melnick, Nicholas Frosst, and 4 more authors
    Advances in Neural Information Processing Systems, 2021
  2. How Interpretable and Trustworthy are GAMs?
    Chun-Hao Chang, Sarah Tan, Ben Lengerich, and 2 more authors
    In Proceedings of the 27th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining , 2021

2020

  1. Purifying Interaction Effects with the Functional ANOVA: An Efficient Algorithm for Recovering Identifiable Additive Models
    Ben Lengerich, Sarah Tan, Chun-Hao Chang, and 2 more authors
    In Proceedings of the Twenty Third International Conference on Artificial Intelligence and Statistics (AISTATS) , 26–28 aug 2020