One of the main challenges for broad adoption of deep learning based models such as convolutional neural networks (CNN), is the lack of understanding of their decisions. In many applications, a simpler, less capable model that can be easily understood is favorable to a black-box model that has superior performance. In this paper, we present an approach for designing CNNs based on visualization of the internal activations of the model. We visualize the model's response through attentive response maps obtained using a fractional stride convolution technique and compare the results with known imaging landmarks from the medical literature. We show that sufficiently deep and capable models can be successfully trained to use the same medical landmarks a human expert would use. Our approach allows for communicating the model decision process well, but also offers insight towards detecting biases.
翻译:广泛采用以深层次学习为基础的模型,如进化神经网络(CNN)的主要挑战之一是缺乏对其决定的理解。在许多应用中,一个容易理解的更简单、能力较弱的模型优于一个表现优异的黑盒模型。在本文中,我们介绍了一种基于模型内部激活的可视化设计CNN的方法。我们通过使用小步进进化技术获得的细微反应图来想象出该模型的反应,并将结果与医学文献中已知的成像标志进行比较。我们表明,足够深、能力较弱的模型可以成功地被训练,以便使用人类专家将使用的相同医学标志。我们的方法可以很好地传达示范决策进程,但也为发现偏差提供了洞察力。