搜索

x

聚合双流特征的高分遥感影像场景分类模型

A high-resolution remote sensing image scene classification model with aggregated dual-stream features

  • 摘要: 针对当前基于深度学习的场景分类方法所提取特征鉴别能力有限、分类效果不理想的问题,为了促进场景代表性特征的有效学习,本研究提出一种包含卷积流和转换器流的双流架构SC-ETNet用于遥感图像场景分类。具体而言,卷积流采用空间和通道重建卷积对卷积层提取的特征进行分离重建;转换器流采用LightViT对全局标记与图像标记进行交互以实现局部-全局注意力计算。SC-ETNet在UC-Merced、AID和NWPU-RESISC45三个数据集的实验评估分别达到了99.61%、97.81%和95.33%的平均分类准确率,这一结果表明,相比现有的先进场景分类方法,SC-ETNet具有更加优越的分类性能。

     

    Abstract: Aimed at the limitation of the discriminative capability of the extracted features and unsatisfactory classification performance of the scene classification methods based on the deep learning, in the report, in order to improve the effective learning of the scene-representative features, a dual-stream architecture named SC-ETNet, which include the convolution stream and the transformation stream, was proposed for the remote sensing image scene classification. The convolution stream employed the spatial and channel reconstruction convolutions to separate and reconstruct features extracted by convolutional layers. The transformation stream used LightViT for the interaction between global tokens and image tokens to achieve local-global attention computation. The mean classification accuracy of the evaluation on the UC-Merced, AID, and NWPU-RESISC45 datasets was 99.61%, 97.81%, and 95.33%, respectively. These data suggested that compared with the existing advanced scene classification methods, SC-ETNet demonstrates superior classification performance.

     

/

返回文章
返回