{"title":"CGR-Net: Consistency Guided ResFormer for Two-View Correspondence Learning","authors":"Changcai Yang;Xiaojie Li;Jiayi Ma;Fengyuan Zhuang;Lifang Wei;Riqing Chen;Guodong Chen","doi":"10.1109/TCSVT.2024.3439348","DOIUrl":null,"url":null,"abstract":"Accurately identifying correct correspondences (inliers) in two-view images is a fundamental task in computer vision. Recent studies usually adopt Graph Neural Networks or stack local graphs into global ones to establish neighborhood relations. However, the smoothing properties of Graph Convolutional Neural network (GCN) cause the model to fall into local extreme, which leads to the issue of indistinguishability between inliers and outliers. Especially when the initial correspondences contain a large number of incorrect correspondences (outliers), these studies suffer from severe performance degradation. To address the above issues and refocus perspective information on distinct features, we design a Consistency Guided ResFormer Network (CGR-Net) that uses consistent correspondences to guide model perspective focusing, thereby avoiding the negative impact of outliers. Specifically, we design an efficient Graph Score Calculation module, which aims to compute global graph scores by enhancing the representation of important features and comprehensively capturing the contextual relationships between correspondences. Then, we propose a Consistency Guided Correspondences Selection module to dynamically fuse global graph scores and consistency graphs and construct a novel consistency matrix to accurately recognize inliers. Extensive experiments on various challenging tasks demonstrate that our CGR-Net outperforms state-of-the-art methods. Our code is released at \n<uri>https://github.com/XiaojieLi11/CGR-Net</uri>\n.","PeriodicalId":13082,"journal":{"name":"IEEE Transactions on Circuits and Systems for Video Technology","volume":"34 12","pages":"12450-12465"},"PeriodicalIF":11.1000,"publicationDate":"2024-08-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Circuits and Systems for Video Technology","FirstCategoryId":"5","ListUrlMain":"https://ieeexplore.ieee.org/document/10623710/","RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0
Abstract
Accurately identifying correct correspondences (inliers) in two-view images is a fundamental task in computer vision. Recent studies usually adopt Graph Neural Networks or stack local graphs into global ones to establish neighborhood relations. However, the smoothing properties of Graph Convolutional Neural network (GCN) cause the model to fall into local extreme, which leads to the issue of indistinguishability between inliers and outliers. Especially when the initial correspondences contain a large number of incorrect correspondences (outliers), these studies suffer from severe performance degradation. To address the above issues and refocus perspective information on distinct features, we design a Consistency Guided ResFormer Network (CGR-Net) that uses consistent correspondences to guide model perspective focusing, thereby avoiding the negative impact of outliers. Specifically, we design an efficient Graph Score Calculation module, which aims to compute global graph scores by enhancing the representation of important features and comprehensively capturing the contextual relationships between correspondences. Then, we propose a Consistency Guided Correspondences Selection module to dynamically fuse global graph scores and consistency graphs and construct a novel consistency matrix to accurately recognize inliers. Extensive experiments on various challenging tasks demonstrate that our CGR-Net outperforms state-of-the-art methods. Our code is released at
https://github.com/XiaojieLi11/CGR-Net
.
期刊介绍:
The IEEE Transactions on Circuits and Systems for Video Technology (TCSVT) is dedicated to covering all aspects of video technologies from a circuits and systems perspective. We encourage submissions of general, theoretical, and application-oriented papers related to image and video acquisition, representation, presentation, and display. Additionally, we welcome contributions in areas such as processing, filtering, and transforms; analysis and synthesis; learning and understanding; compression, transmission, communication, and networking; as well as storage, retrieval, indexing, and search. Furthermore, papers focusing on hardware and software design and implementation are highly valued. Join us in advancing the field of video technology through innovative research and insights.