基于自适应步幅卷积的细粒度视觉识别

FINE-GRAINED VISUAL CLASSIFICATION BASED ON ADAPTIVE STRIDE CONVOLUTION

  • 摘要: 平均池化等下采样方法已被广泛用于降低计算成本、防止过拟合和提高卷积神经网络的性能。然而,在细粒度的识别任务中,这些均匀采样方法不能很好地关注细微的辨别区域。提出一个自适应步幅卷积网络,JP+1其中自适应步幅卷积模块被用来专注于提取细微的特征。具体来说,给定一个图像,使用注意力图提取器获得一个注意力图,以突出物体的有判别性的部分。基于注意力图的步幅向量生成器产生步幅向量,它表示卷积核每次的移动步幅。自适应步幅卷积在输入图像上以不同的步幅提取信息。在CUB-200-2011、Stanford Cars和FGVC-Aircraft三个具有挑战性的细粒度数据集上,对该方法的有效性进行实验评估,结果达到先进的性能。

     

    Abstract: Down-sampling methods such as average pooling have been widely used to reduce computation cost, prevent overfitting, and improve the performance of convolutional neural networks. However, in fine-grained recognition tasks, these uniform sampling methods cannot focus well on subtle discriminative regions. In this paper, we propose an Adaptive Stride Convolution Network (ASCNet) in which the ASC module is used to focus on extracting subtle features. Specifically, given an image, we obtained an attention map to highlight the discriminative parts of object, where the attention map extractor was used. The attention map-based stride generator produced stride vectors which indicated the moving steps of convolutional kernels every time. The adaptive stride convolution extracted information over the input image or features with varying strides. We experimentally evaluated the effectiveness of our method on three challenging JP3fine-grained benchmarks, i.e.KG-*3, CUB-200-2011, Stanford Cars, and FGVC-Aircraft, and advanced performance is achieved.

     

/

返回文章
返回