Generating Human Interpretable Rules from Convolutional Neural Networks
Listed in
This article is not in any list yet, why not save it to one of your lists.Abstract
Advancements in the field of Artificial Intelligence has been rapid in recent years and has revolutionized various industries. Various deep neural network architectures capable of handling both text and images, covering code generation from natural language, producing machine translation and text summarizations have been proposed. For example, Convolutional Neural Networks or CNNs perform image classification at a level equivalent to that of humans on many image datasets. These state-of-the-art networks have reached unprecedented levels of success by using complex architectures with billions of parameters, numerous kernel configurations, weight initialization and regularization methods. Unfortunately to reach this level of success, the models that CNNs use are essentially black box in nature with little or no human interpretable information on the decision-making process. This lack of transparency in decision making gave rise to concerns amongst some sectors of the user community such as healthcare, finance, justice, defense, among others. This challenge motivated our research where we successfully produced human interpretable influential features from CNNs for image classification and captured the interactions between these features by producing a concise decision tree making accurate classification decisions. The proposed methodology made use of a pre-trained VGG16 with finetuning to extract feature maps produced by learnt filters. On the CelebA image benchmark dataset, we successfully produced human interpretable rules that captured the main facial landmarks responsible for segmenting males from females with 89.6% accuracy, while on the more challenging Cats vs Dogs dataset the decision tree produced 87.6% accuracy.