UCL Discovery
UCL home » Library Services » Electronic resources » UCL Discovery

Interpretability of deep learning models: A survey of results

Chakraborty, S; Tomsett, R; Raghavendra, R; Harborne, D; Alzantot, M; Cerutti, F; Srivastava, M; ... Gurram, P; + view all (2017) Interpretability of deep learning models: A survey of results. In: 2017 IEEE SmartWorld: conference proceedings. IEEE Green open access

[thumbnail of Chakraborty_Interpretability of deep learning models.pdf]
Preview
Text
Chakraborty_Interpretability of deep learning models.pdf - Accepted Version

Download (112kB) | Preview

Abstract

Deep neural networks have achieved near-human accuracy levels in various types of classification and prediction tasks including images, text, speech, and video data. However, the networks continue to be treated mostly as black-box function approximators, mapping a given input to a classification output. The next step in this human-machine evolutionary process-incorporating these networks into mission critical processes such as medical diagnosis, planning and control-requires a level of trust association with the machine output. Typically, statistical metrics are used to quantify the uncertainty of an output. However, the notion of trust also depends on the visibility that a human has into the working of the machine. In other words, the neural network should provide human-understandable justifications for its output leading to insights about the inner workings. We call such models as interpretable deep networks. Interpretability is not a monolithic notion. In fact, the subjectivity of an interpretation, due to different levels of human understanding, implies that there must be a multitude of dimensions that together constitute interpretability. In addition, the interpretation itself can be provided either in terms of the low-level network parameters, or in terms of input features used by the model. In this paper, we outline some of the dimensions that are useful for model interpretability, and categorize prior work along those dimensions. In the process, we perform a gap analysis of what needs to be done to improve model interpretability.

Type: Proceedings paper
Title: Interpretability of deep learning models: A survey of results
Event: 2017 IEEE SmartWorld, 4-8 August 2017, San Fransisco, California, USA
ISBN-13: 9781538604342
Open access status: An open access version is available from UCL Discovery
DOI: 10.1109/UIC-ATC.2017.8397411
Publisher version: https://doi.org/10.1109/UIC-ATC.2017.8397411
Language: English
Additional information: This version is the author accepted manuscript. For information on re-use, please refer to the publisher’s terms and conditions.
UCL classification: UCL
UCL > Provost and Vice Provost Offices > UCL BEAMS
UCL > Provost and Vice Provost Offices > UCL BEAMS > Faculty of Engineering Science
UCL > Provost and Vice Provost Offices > UCL BEAMS > Faculty of Engineering Science > Dept of Computer Science
URI: https://discovery.ucl.ac.uk/id/eprint/10059575
Downloads since deposit
3,353Downloads
Download activity - last month
Download activity - last 12 months
Downloads by country - last 12 months

Archive Staff Only

View Item View Item