Counterfactual Explanations for Machine Learning: Challenges Revisited
- Manage
- Copy
- Actions
- Export
- Annotate
- Print Preview
Choose the export format from the list below:
- Office Formats (1)
-
Export as Portable Document Format (PDF) using Apache Formatting Objects Processor (FOP)
-
- Other Formats (1)
-
Export as HyperText Markup Language (HTML)
-
Sahil Verma, John P. Dickerson, Keegan Hines
CoRR abs/2106.07756
2021
Counterfactual explanations (CFEs) are an emerging technique under the umbrella of interpretability of machine learning (ML) models. They provide ``what if'' feedback of the form ``if an input datapoint were x′ instead of x, then an ML model's output would be y′ instead of y.'' Counterfactual explainability for ML models has yet to see widespread adoption in industry. In this short paper, we posit reasons for this slow uptake. Leveraging recent work outlining desirable properties of CFEs and our experience running the ML wing of a model monitoring startup, we identify outstanding obstacles hindering CFE deployment in industry. |
Publications / Personal
Publications / Views
Home
— clouds
tags | authors | editors | journals
— per year
2023 | 2022 | 2021 | 2020 | 2019 | 2018 | 2017 | 2016 | 2015 | 2014–1927
— per sort
in journal | in proc | chapters | books | edited | spec issues | editorials | entries | manuals | tech reps | phd th | others
— per status
online | in press | proof | camera-ready | revised | accepted | revision | submitted | draft | note
— services
ACM Digital Library | DBLP | IEEE Xplore | IRIS | PubMed | Google Scholar | Scopus | Semantic Scholar | Web of Science | DOI
Publication
— authors
Sahil Verma, John P. Dickerson, Keegan Hines
— status
published
— sort
other
— publication date
2021
— journal
CoRR
— volume
abs/2106.07756
URLs
original page | original PDF | open access PDF
files