skip to main content
10.1145/3306618.3314273acmconferencesArticle/Chapter ViewAbstractPublication PagesaiesConference Proceedingsconference-collections
research-article

TED: Teaching AI to Explain its Decisions

Published:27 January 2019Publication History

ABSTRACT

Artificial intelligence systems are being increasingly deployed due to their potential to increase the efficiency, scale, consistency, fairness, and accuracy of decisions. However, as many of these systems are opaque in their operation, there is a growing demand for such systems to provide explanations for their decisions. Conventional approaches to this problem attempt to expose or discover the inner workings of a machine learning model with the hope that the resulting explanations will be meaningful to the consumer. In contrast, this paper suggests a new approach to this problem. It introduces a simple, practical framework, called Teaching Explanations for Decisions (TED), that provides meaningful explanations that match the mental model of the consumer. We illustrate the generality and effectiveness of this approach with two different examples, resulting in highly accurate explanations with no loss of prediction accuracy for these two examples.

References

  1. Yessenalina Ainur, Yejin Choi, and Claire Cardie. 2010. Automatically Generating Annotator Rationales to Improve Sentiment Classification. In Proceedings of the ACL 2010 Conference Short Papers. 336--341. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Osbert Bastani, Carolyn Kim, and Hamsa Bastani. 2018. Interpreting Blackbox Models via Model Extraction. arXiv preprint arXiv:1705.08504 (2018).Google ScholarGoogle Scholar
  3. Or Biran and Courtenay Cotton. 2017. Explanation and Justification in Machine Learning: A Survey. In IJCAI-17 Workshop on Explainable AI (XAI).Google ScholarGoogle Scholar
  4. Alex Campolo, Madelyn Sanfilippo Meredith Whittaker, and Kate Crawford. 2017. 2017 Annual Report. Technical Report. AI NOW.Google ScholarGoogle Scholar
  5. Rich Caruana, Yin Lou, Johannes Gehrke, Paul Koch, Marc Sturm, and Noemie Elhadad. 2015. Intelligible Models for HealthCare: Predicting Pneumonia Risk and Hospital 30-day Readmission. In Proc. ACM SIGKDD Int. Conf. Knowl. Disc. Data Min. Sydney, Australia, 1721--1730. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. Sanjeeb Dash, Oktay Gunluk, and Dennis Wei. 2018. Boolean Decision Rules via Column Generation. In Advances in Neural Information Processing Systems. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Amit Dhurandhar, Vijay Iyengar, Ronny Luss, and Karthikeyan Shanmugam. 2017. A Formal Framework to Characterize Interpretability of Procedures. In Proc. ICML Workshop Human Interp. Mach. Learn. Sydney, Australia, 1--7.Google ScholarGoogle Scholar
  8. Jeff Donahue and Kristen Grauman. 2011. Annotator Rationales for Visual Recognition. In ICCV. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. Finale Doshi-Velez and Been Kim. 2017. Towards A Rigorous Science of Interpretable Machine Learning. In https://arxiv.org/abs/1702.08608v2.Google ScholarGoogle Scholar
  10. Finale Doshi-Velez, Mason Kortz, Ryan Budish, Chris Bavitz, Sam Gershman, David O'Brien, Stuart Schieber, James Waldo, David Weinberger, and Alexandra Wood. 2017. Accountability of AI Under the Law: The Role of Explanation. CoRR, Vol. abs/1711.01134 (2017). arxiv: 1711.01134 http://arxiv.org/abs/1711.01134Google ScholarGoogle Scholar
  11. Kun Duan, Devi Parikh, David Crandall, and Kristen Grauman. 2012. Discovering Localized Attributes for Fine-grained Recognition. In CVPR.Google ScholarGoogle Scholar
  12. FICO. 2018. Explainable Machine Learning Challenge. https://community.fico.com/s/explainable-machine-learning-challenge?tabset-3158a=2Google ScholarGoogle Scholar
  13. Ian Goodfellow, Yoshua Bengio, and Aaron Courville. 2016. Deep Learning .MIT Press. http://www.deeplearningbook.org. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Bryce Goodman and Seth Flaxman. 2016. EU Regulations on Algorithmic Decision-Making and a `Right to Explanation'. In Proc. ICML Workshop Human Interp. Mach. Learn. New York, NY, 26--30.Google ScholarGoogle Scholar
  15. Lisa Anne Hendricks, Zeynep Akata, Marcus Rohrbach, Jeff Donahue, Bernt Schiele, and Trevor Darrell. 2016. Generating Visual Explanations. In European Conference on Computer Vision.Google ScholarGoogle Scholar
  16. Been Kim. 2017. Tutorial on Interpretable machine learning. "http://people.csail.mit.edu/beenkim/papers/BeenK_FinaleDV_ICML2017_tutorial.pdf"Google ScholarGoogle Scholar
  17. Been Kim, Dmitry M. Malioutov, Kush R. Varshney, and Adrian Weller (Eds.). 2017. 2017 ICML Workshop on Human Interpretability in Machine Learning. https://arxiv.org/abs/1708.02666v1Google ScholarGoogle Scholar
  18. Been Kim, Kush R. Varshney, and Adrian Weller (Eds.). 2018. 2018 Workshop on Human Interpretability in Machine Learning. https://sites.google.com/view/whi2018/homeGoogle ScholarGoogle Scholar
  19. Todd Kulesza, Simone Stumpf, Margaret Burnett, Sherry Yang, Irwin Kwan, and Weng-Keen Wong. 2013. Too Much, Too Little, or Just Right? Ways Explanations Impact End Users' Mental Models. In Proc. IEEE Symp. Vis. Lang. Human-Centric Comput. San Jose, CA, 3--10.Google ScholarGoogle ScholarCross RefCross Ref
  20. Tao Lei, Regina Barzilay, and Tommi Jaakkola. 2016. Rationalizing neural predictions. In EMNLP.Google ScholarGoogle Scholar
  21. Zachary C Lipton. 2016. The mythos of model interpretability. In ICML Workshop on Human Interpretability of Machine Learning.Google ScholarGoogle Scholar
  22. Tania Lombrozo. 2007. Simplicity and probability in causal explanation. Cognitive Psychol., Vol. 55, 3 (Nov. 2007), 232--257.Google ScholarGoogle ScholarCross RefCross Ref
  23. Scott Lundberg and Su-In Lee. 2017. A unified approach to interpreting model predictions. In Advances of Neural Inf. Proc. Systems. Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. T. McDonnell, M. Lease, M. Kutlu, and T. Elsayed. 2016. Why Is That Relevant? Collecting Annotator Rationales for Relevance Judgments. In Proc. AAAI Conf. Human Comput. Crowdsourc.Google ScholarGoogle Scholar
  25. Tim Miller. 2017. Explanation in Artificial Intelligence: Insights from the Social Sciences. arXiv preprint arXiv:1706.07269 (June 2017).Google ScholarGoogle Scholar
  26. Tim Miller, Piers Howe, and Liz Sonenberg. 2017. Explainable AI: Beware of Inmates Running the Asylum Or: How I Learnt to Stop Worrying and Love the Social and Behavioural Sciences. In Proc. IJCAI Workshop Explainable Artif. Intell. Melbourne, Australia.Google ScholarGoogle Scholar
  27. Grégoire Montavon, Wojciech Samek, and Klaus-Robert Müller. 2017. Methods for interpreting and understanding deep neural networks. Digital Signal Processing (2017).Google ScholarGoogle Scholar
  28. P. Peng, Y. Tian, T. Xiang, Y. Wang, and T. Huang. 2016. Joint Learning of Semantic and Latent Attributes. In ECCV 2016, Lecture Notes in Computer Science, Vol. 9908.Google ScholarGoogle Scholar
  29. Marco Tulio Ribeiro, Sameer Singh, and Carlos Guestrin. 2016. "Why Should I Trust You?": Explaining the Predictions of Any Classifier. In Proc. ACM SIGKDD Int. Conf. Knowl. Disc. Data Min. San Francisco, CA, 1135--1144. Google ScholarGoogle ScholarDigital LibraryDigital Library
  30. A. D. Selbst and J. Powles. 2017. Meaningful Information and the Right to Explanation. Int. Data Privacy Law, Vol. 7, 4 (Nov. 2017), 233--242.Google ScholarGoogle ScholarCross RefCross Ref
  31. Qiang Sun and Gerald DeJong. 2005. Explanation-Augmented SVM: an Approach to Incorporating Domain Knowledge into SVM Learning. In 22nd International Conference on Machine Learning. Google ScholarGoogle ScholarDigital LibraryDigital Library
  32. James Vacca. 2018. A Local Law in relation to automated decision systems used by agencies. Technical Report. The New York City Council.Google ScholarGoogle Scholar
  33. Kush R. Varshney. 2016. Engineering Safety in Machine Learning. In Information Theory and Applications Workshop.Google ScholarGoogle Scholar
  34. S. Wachter, B. Mittelstadt, and L. Floridi. 2017. Transparent, explainable, and accountable AI for robotics. Science Robotics, Vol. 2 (May 2017). Issue 6.Google ScholarGoogle Scholar
  35. Sandra Wachter, Brent Mittelstadt, and Luciano Floridi. 2017. Why a Right to Explanation of Automated Decision-Making Does Not Exist in the General Data Protection Regulation. Int. Data Privacy Law, Vol. 7, 2 (May 2017), 76--99.Google ScholarGoogle Scholar
  36. Omar F. Zaidan and Jason Eisner. 2007. Using 'annotator rationales' to improve machine learning for text categorization. In In NAACL-HLT. 260--267.Google ScholarGoogle Scholar
  37. Omar F. Zaidan and Jason Eisner. 2008. Modeling Annotators: A Generative Approach to Learning from Annotator Rationales. In Proceedings of EMNLP 2008. 31--40. Google ScholarGoogle ScholarDigital LibraryDigital Library
  38. Ye Zhang, Iain James Marshall, and Byron C. Wallace. 2016. Rationale-Augmented Convolutional Neural Networks for Text Classification. In Conference on Empirical Methods in Natural Language Processing (EMNLP).Google ScholarGoogle Scholar

Index Terms

  1. TED: Teaching AI to Explain its Decisions

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in
      • Published in

        cover image ACM Conferences
        AIES '19: Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society
        January 2019
        577 pages
        ISBN:9781450363242
        DOI:10.1145/3306618

        Copyright © 2019 ACM

        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 27 January 2019

        Permissions

        Request permissions about this article.

        Request Permissions

        Check for updates

        Qualifiers

        • research-article

        Acceptance Rates

        Overall Acceptance Rate61of162submissions,38%

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader