系统工程与电子技术 ›› 2023, Vol. 45 ›› Issue (11): 3363-3373.doi: 10.12305/j.issn.1001-506X.2023.11.01

• 电子技术 • 上一篇    下一篇

三分支空间变换注意力机制的图像匹配算法

黄妍妍1,2, 盖绍彦1,2,*, 达飞鹏1,2   

  1. 1. 东南大学自动化学院, 江苏 南京 210096
    2. 东南大学复杂工程系统测量与控制教育部重点实验室, 江苏 南京 210096
  • 收稿日期:2022-08-15 出版日期:2023-10-25 发布日期:2023-10-31
  • 通讯作者: 盖绍彦
  • 作者简介:黄妍妍 (1997—), 女, 硕士研究生, 主要研究方向为深度学习、图像处理
    盖绍彦 (1979—), 男, 副教授, 博士, 主要研究方向为计算机视觉、模式识别、三维测量
    达飞鹏 (1968—), 男, 教授, 博士, 主要研究方向为三维精密测量、三维精确识别、三维优化控制理论与技术
  • 基金资助:
    江苏省前沿引领技术基础研究专项(BK20192004C);江苏省高校优势学科建设工程资助课题

Image matching algorithm based on attention mechanism of three branch spatial transformation

Yanyan HUANG1,2, Shaoyan GAI1,2,*, Feipeng DA1,2   

  1. 1. School of Automation, Southeast University, Nanjing 210096, China
    2. Key Laboratory of Measurement and Control of Complex Systems of Engineering, Ministry of Education, Southeast University, Nanjing 210096, China
  • Received:2022-08-15 Online:2023-10-25 Published:2023-10-31
  • Contact: Shaoyan GAI

摘要:

对于待匹配图像具有旋转、缩放、平移等空间几何变换的图像模板匹配任务, 现有的算法耗时较长, 且准确率不高。针对该问题提出一种高准确率、低运算成本的图像匹配算法, 首先根据中心点与邻域点的像素差来寻找特征点, 进行快速特征检测, 然后以这些特征点为中心, 并以快速特征检测所计算出来的旋转角截取出一定尺寸的图像块。再将这些图像块输入空间变换注意力模块的特征描述子提取网络, 最后使用K最邻近算法计算两张待匹配图像特征描述子中匹配的特征。特征描述子提取网络中引入了空间变换注意力模块, 网络在训练的时候着重对空间信息进行学习, 故所提算法提高了具有较大空间变化图像匹配任务的准确率。在匹配时间方面, 所提的匹配算法仅次于检测和匹配都使用快速特征检测算法的方法。在匹配准确率方面, 所提算法匹配的准确率远远优于实验所比较的其他算法。

关键词: 计算机视觉, 图像匹配, 空间变换不变性, 卷积神经网络, 注意力机制

Abstract:

For the image template matching task in which the image to be matched has spatial transformations such as rotation transformation, scale transformation, and translation, the existing algorithms are time-consuming and have low accuracy. In this paper, an image matching algorithm with high accuracy is proposed. Firstly, oriented fast feature detection algorithm is used to find the feature points according to the pixel difference between the center point and the neighboring point in order to detect the feature points. With these feature points taken as the center, patches of a certain size are cropped according to the rotation angle calculated by the fast feauture detection algorithm. These image patches are fed into the feature descriptor extraction network of the spatial transform attention module. Finally, K-nearest neighbor (KNN) algorithm is used to calculate the matched features in the feature descriptors of the two images to be matched. The spatial transformation attention module is introduced into the feature descriptor extraction network. The network focuses on learning spatial information during training, so the entire algorithm improves the accuracy of image matching tasks with large spatial changes. In terms of matching time, the matching algorithm proposed in this paper is second only to the method that uses fast feauture detection algorithm algorithmfor both detection and matching tasks. In terms of matching accuracy, the matching accuracy of the algorithm in this paper is far superior to other algorithms compared in the experiment.

Key words: computer vision, image matching, space transformation invariance, convolutional neural network (CNN), attention mechanism

中图分类号: