We propose Self-Supervised Implicit Attention (SSIA), a new approach that adaptively guides deep neural network models to gain attention by exploiting the properties of the models themselves. SSIA is a novel attention mechanism that does not require any extra parameters, computation, or memory access costs during inference, which is in contrast to existing attention mechanism. In short, by considering attention weights as higher-level semantic information, we reconsidered the implementation of existing attention mechanisms and further propose generating supervisory signals from higher network layers to guide lower network layers for parameter updates. We achieved this by building a self-supervised learning task using the hierarchical features of the network itself, which only works at the training stage. To verify the effectiveness of SSIA, we performed a particular implementation (called an SSIA block) in convolutional neural network models and validated it on several image classification datasets. The experimental results show that an SSIA block can significantly improve the model performance, even outperforms many popular attention methods that require additional parameters and computation costs, such as Squeeze-and-Excitation and Convolutional Block Attention Module. Our implementation will be available on GitHub.
翻译:我们提出“自我增强隐含注意”这一新方针,通过利用模型本身的特性来适应性地指导深神经网络模型,以引起人们的注意。SSIA是一个新的关注机制,在推断过程中不需要额外的参数、计算或记忆存取费用,这与现有的注意机制形成对照。简而言之,通过将注意力权重视为更高层次的语义学信息,我们重新考虑了现有关注机制的实施,并进一步提议从更高的网络层产生监督信号,以指导低网络层的参数更新。我们通过利用网络本身的等级特征来建立自我监督的学习任务来实现这一目标,而这种任务只是在培训阶段运作。为了核查SSIA的有效性,我们进行了一个特殊的演进神经网络模型(称为SSIA区块),并在若干图像分类数据集上验证了它。实验结果表明,SSIA区块可以大大改进模型的性能,甚至超越许多需要额外参数和计算成本的公众关注方法。我们将在GiHUB上实施。