Research Connect
Research PapersAboutContact

Interpretable Machine Learning

Published in (2019-11-07)
aionlincourseaionlincourseaionlincourseaionlincourse
Generate GraphDownload

On This Page

  • TL;DR
  • Abstract
  • Authors
  • Datasets
  • References
TL

TL;DR

This project introduces Robust T CAV, which builds on TCAV and experimentally determines best practices for this method and is a step in the direction of making TCAVs, an already impactful algorithm in interpretability, more reliable and useful for practitioners.

Abstract

Interpretable machine learning has become a popular research direction as deep neural networks (DNNs) have become more powerful and their applications more mainstream, yet DNNs remain difficult to understand. Testing with Concept Activation Vectors, TCAV, (Kim et al. 2017) is an approach to interpreting DNNs in a human-friendly way and has recently received significant attention in the machine learning community. The TCAV algorithm achieves a degree of global interpretability for DNNs through human-defined concepts as explanations. This project introduces Robust TCAV, which builds on TCAV and experimentally determines best practices for this method. The objectives for Robust TCAV are 1) Making TCAV more consistent by reducing variance in the TCAV score distribution and 2) Increasing CAV and TCAV score resistance to perturbations. A difference of means method for CAV generation was determined to be the best practice to achieve both objectives. Many areas of the TCAV process are explored including CAV visualization in low dimensions, negative class selection, and activation perturbation in the direction of a CAV. Finally, a thresholding technique is considered to remove noise in TCAV scores. This project is a step in the direction of making TCAV, an already impactful algorithm in interpretability, more reliable and useful for practitioners.

Authors

Bradley C. Boehmke

1 Paper

Brandon M. Greenwell

1 Paper

References165 items

1

Deep Residual Learning for Image Recognition

2

Unpaired Image-to-Image Translation Using Cycle-Consistent Adversarial Networks

3

Going deeper with convolutions

4

ImageNet classification with deep convolutional neural networks

5

Deep learning in neural networks: An overview

6

Research Impact

2391

Citations

165

References

0

Datasets

2

ImageNet Large Scale Visual Recognition Challenge

7

Datasheets for datasets

8

Intriguing properties of neural networks

9

Machine learning: Trends, perspectives, and prospects

Medicine
10

Gender Shades: Intersectional Accuracy Disparities in Commercial Gender Classification

11

Runaway Feedback Loops in Predictive Policing

12

Automated Experiments on Ad Privacy Settings

13

Explainable Artificial Intelligence (XAI): Concepts, Taxonomies, Opportunities and Challenges toward Responsible AI

14

Understanding artificial intelligence ethics and safety

15

Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization

16

European Union Regulations on Algorithmic Decision-Making and a "Right to Explanation"

17

The mythos of model interpretability

18

Synthesizing the preferred inputs for neurons in neural networks via deep generator networks

19

Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning

20

Explainable AI: Interpreting, Explaining and Visualizing Deep Learning

21

“Why Should I Trust You?”: Explaining the Predictions of Any Classifier

22

Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead

MedicineComputer Science
23

Explaining and Harnessing Adversarial Examples

24

InterpretML: A Unified Framework for Machine Learning Interpretability

25

Explainable machine learning in deployment

26

Counterfactual Explanations Without Opening the Black Box: Automated Decisions and the GDPR

27

The global landscape of AI ethics guidelines

28

A Survey of Methods for Explaining Black Box Models

29

Interpretable Explanations of Black Boxes by Meaningful Perturbation

30

SmoothGrad: removing noise by adding noise

31

Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps

32

AUTO-ENCODING VARIATIONAL BAYES

33

Model Cards for Model Reporting

34

Describing Textures in the Wild

35

beta-VAE: Learning Basic Visual Concepts with a Constrained Variational Framework

36

Interpretability Beyond Feature Attribution: Quantitative Testing with Concept Activation Vectors (TCAV)

37

Real Time Image Saliency for Black Box Classifiers

38

Image Style Transfer Using Convolutional Neural Networks

39

Explaining Explanations in AI

40

This looks like that: deep learning for interpretable image recognition

41

Why a Right to Explanation of Automated Decision-Making Does Not Exist in the General Data Protection Regulation

42

A Human-Centered Agenda for Intelligible Machine Learning

43

Machine Learning Explainability for External Stakeholders

44

The EU Commission

45

Interpreting Interpretability: Understanding Data Scientists' Use of Interpretability Tools for Machine Learning

46

Toward Trustworthy AI Development: Mechanisms for Supporting Verifiable Claims

47

Closing the AI accountability gap: defining an end-to-end framework for internal algorithmic auditing

48

Face recognition vendor test part 3:

49

Algorithmic Decision-Making and the Control Problem

50

Millions of black people affected by racial bias in health-care algorithms

51

Dissecting racial bias in an algorithm used to manage the health of populations

52

A systematic review of algorithm aversion in augmented decision making

53

AI-Assisted Decision-making in Healthcare

54

Machine Learning Interpretability: A Survey on Methods and Metrics

55

How model accuracy and explanation fidelity influence user trust

56

Administrative law and the machines of government: judicial review of automated public-sector decision-making

57

FactSheets: Increasing trust in AI services through supplier's declarations of conformity

58

Shaping the State of Machine Learning Algorithms within Policing: Workshop Report

59

Principles alone cannot guarantee ethical AI

60

Affinity Profiling and Discrimination by Association in Online Behavioural Advertising

61

From What to How: An Initial Review of Publicly Available AI Ethics Tools, Methods and Research to Translate Principles into Practices

62

The Ethics of AI Ethics: An Evaluation of Guidelines

63

Dirty Data, Bad Predictions: How Civil Rights Violations Impact Police Data, Predictive Policing Systems, and Justice

64

Human-Centered Tools for Coping with Imperfect Algorithms During Medical Decision-Making

65

Automating Interpretability: Discovering and Testing Visual Concepts Learned by Neural Networks

66

Towards a Definition of Disentangled Representations

67

Opening the black box of machine learning.

68

Response to centre for data ethics and innovation consultation

69

Algorithm-assisted decision-making in the public sector: framing the issues using administrative law rules governing discretionary power

70

Explaining Image Classifiers by Counterfactual Generation

71

General Data Protection Regulation

72

Defining Locality for Surrogates in Post-hoc Interpretablity

73

Transparent to whom? No algorithmic accountability without a critical audience

74

Explaining Explanations: An Overview of Interpretability of Machine Learning

75

What About Us?

76

The General Data Protection Regulation (GDPR)

77

Rights related to automated decision making including profiling

78

Enslaving the Algorithm: From a “Right to an Explanation” to a “Right to Better Decisions”?

79

Manipulating and Measuring Model Interpretability

80

Net2Vec: Quantifying and Explaining How Concepts are Encoded by Filters in Deep Neural Networks

81

"Meaningful Information" and the Right to Explanation

82

Industrial strategy: building a Britain fit for the future

83

Accountability of AI Under the Law: The Role of Explanation

84

The (Un)reliability of saliency methods

85

Algorithmic risk assessment policing models: lessons from the Durham HART model and ‘Experimental’ proportionality

86

Algorithmic Bias in Autonomous Systems

87

The Equality Act 2010

88

Challenges for Transparency

89

Deep Text Classification Can be Fooled

90

Artificial Neural Networks

91

Towards A Rigorous Science of Interpretable Machine Learning

Computer ScienceMathematics
92

The ethics of algorithms: Mapping the debate

93

Understanding intermediate layers using linear classifier probes

94

Can we open the black box of AI?

95

To predict and serve?

96

Semantics derived automatically from language corpora contain human-like biases

97

Making Tree Ensembles Interpretable

98

Model-Agnostic Interpretability of Machine Learning

99

An FDA for Algorithms

100

Interpretable classifiers using rules and Bayesian analysis: Building a better stroke prediction model

101

Algorithm Aversion: People Erroneously Avoid Algorithms after Seeing Them Err

102

Trade-offs

103

How to Discriminate between Computer-Aided and Computer-Hindered Decisions

104

Data Protection Act

105

The relative influence of advice from human experts and statistical methods on forecast adjustments

106

Discrimination-aware data mining

107

Sparse Principal Component Analysis

108

Weighted support vector machine for data classification

109

Detection of Influential Observation in Linear Regression

110

Explaining Decisions Made with AI

111

White Paper on Artificial Intelligence

112

IEEE Invites Companies, Governments and Other Stakeholders Globally to Expand on Ethics Certification Program for Autonomous and Intelligent Systems (ECPAIS) Work

113

Commitee on Standards in Public Life

114

Explainable AI: the basics

115

Examining-the Black Box

116

Guidance: Medical device stand-alone software including apps (including IVDMDs)

117

Algorithmic Impact Assessment -Évaluation de l'Incidence Algorithmique

118

Transparency: Motivations and Challenges

119

「マテリアルズインフォマティクス」Interpretable Machine Learningによる新材料開発

120

High-Stakes AI Decisions Need to Be Automatically Audited

121

Discriminating systems

122

This is how AI bias really happens — and why it’s so hard to fix

123

Ethically Aligned Design

124

TED”. In: Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society - AIES ’19

125

Snapshot Paper -AI and Personal Insurance

126

The balance: Accuracy vs. Interpretability. Data Science Ninja

127

Project explAIn -interim report

128

Big Brother Watch - submission to the CDEI Bias in Algorithmic Decision Making review

129

The Dstl Biscuit Book

130

Human-AI Collaboration Trust Literature Review - Key Insights and Bibliography

131

AI sector deal one year on

132

AI used for first time in job interviews in UK to find best applicants

133

Algorithms for decision making

134

House of Lords Select Committee on AI

135

Algorithmic Impact Assessments. AI Now

136

Debates, awareness, and projects about GDPR and data protection

137

Sanity Checks for Saliency Maps. arXiv: 1810.03292 [cs.CV

138

Factsheets for AI Services: Building Trusted AI - IBM Research

139

Audit the algorithms that are ruling our lives

140

10 principles for public sector use of algorithmic decision making

141

Artificial intelligence and automation in the UK

142

Algorithms in the Criminal Justice System: Assessing the Use of Risk Assessments in Sentencing

143

SLAVE TO THE ALGORITHM ? WHY A ‘ RIGHT TO AN EXPLANATION ’ IS PROBABLY NOT THE REMEDY YOU ARE LOOKING FOR

144

AI Now 2017 Report

145

Machine learning: the power and promise of computers that learn by example

146

AI analysis: sizing the prize

147

Learning certifiably optimal rule lists for categorical data

148

Economic

149

Artificial intelligence: opportunities and implications for the future of decision making

150

Machine Bias

151

Rethinking the Inception Architecture

152

Financial Conduct Authority

153

Head of maize Content What is the AI Black Box Problem?

154

To appear: IEEE Transactions on Information Forensics and Security Face Recognition Performance: Role of Demographic Information

155

An Applied Perspective on Fairness, Accountability, Transparency, and Explainable AI An Introduction to Machine Learning Interpretability

156

Information Commissioner's Office (2020)

157

Data Ethics Framework

158

ICO Big data, artificial intelligence, machine learning and data protection

159

AI Explanations Whitepaper

160

ICO Explaining

161

GitHub interpretml/interpret

162

PricewaterhouseCoopers Explainable AI

163

The Challenges and Opportunities of Explainable AI

164

PricewaterhouseCoopers Opening AI’s black box will become a priority

165

Responsible AI principles from Microsoft

Authors

Field of Study

Computer Science

Journal Information

Name

Hands-On Machine Learning with R