储岳中(通讯作者), 石玉金, 张学峰, 刘恒. 基于切分通道注意力网络的图像分类算法[J]. 工程科学学报. DOI: 10.13374/j.issn2095-9389.2023.12.21.002
引用本文: 储岳中(通讯作者), 石玉金, 张学峰, 刘恒. 基于切分通道注意力网络的图像分类算法[J]. 工程科学学报. DOI: 10.13374/j.issn2095-9389.2023.12.21.002
Image classification algorithm based on split channel attention network[J]. Chinese Journal of Engineering. DOI: 10.13374/j.issn2095-9389.2023.12.21.002
Citation: Image classification algorithm based on split channel attention network[J]. Chinese Journal of Engineering. DOI: 10.13374/j.issn2095-9389.2023.12.21.002

基于切分通道注意力网络的图像分类算法

Image classification algorithm based on split channel attention network

  • 摘要: 通道注意力机制可以有效利用不同的特征通道,以提高卷积神经网络的分类能力。然而,对于使用全局平均池化来获取通道全局特征的方法,特征图中不同的通道有极大概率出现相同的均值,使得全局平均池化后的特征缺乏多样性,进一步影响网络分类性能。针对此问题,提出一种切分通道注意力机制来构建模块,该模块对全局平均池化的输出维度进行了扩展,增强了通道注意力中全局平均池化层的特征多样性,然后使用一维卷积计算通道注意力。将切分通道注意力机制与多种残差网络相结合,在CIFAR-100和ImageNet数据集上进行了图像分类实验。实验结果表明,切分通道注意力机制能有效提高模型的精度,并且与其它注意力机制相比也表现出较好的优势。

     

    Abstract: Channel attention mechanism can effectively utilize different feature channels to improve the classification ability of convolutional neural networks. However, for the method using global average pooling to obtain the global features of channels, there is a high probability that different channels in the feature graph have the same mean, which makes the features after global average pooling lack diversity, and further affects the classification performance of the network. To solve this problem, a shred channel attention mechanism is proposed to construct a module, which extends the output dimension of global average pooling, enhances the feature diversity of global average pooling layer in channel attention, and then uses one-dimensional convolution to calculate channel attention. Image classification experiments are performed on CIFAR-100 and ImageNet datasets by combining the split-channel attention mechanism with multiple residual networks. The experimental results show that the segmentation channel attention mechanism can effectively improve the accuracy of the model, and it also shows better advantages compared with other attention mechanisms.

     

/

返回文章
返回