Abstract
As a recently emerging problem, cross-view geo-localization aims at finding image pairs captured from different views (e.g., drone and satellite views) or domains yet same location, which can be widely employed in various applications. However, unlike traditional scene classification problem, it faces several challenges, including large intra-class distance and small inter-class distance caused by domain gap, as well as redundant contextual information and visual distractors across views. To address the concerns, we propose a novel cross-domain matching framework to handle this task, which measures the similarity for query and candidate images from two different domains. Comparing to prior classification based framework, our matching based framework is better suited for the task by forcing the model to learn discriminative features for scenes. Moreover, to aid cross-domain matching, we propose a matching-oriented feature modulation scheme, in which we not only apply a large-view attention module to enhance spatial features but also employ channel shuffling to loose the correlation of key feature semantics and distractors in the respective domains. Last, we conduct experiments to show that our model achieves the state-of-the-art performance and surpasses the competing method by a large margin on the public benchmarks.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Liu, L., Li, H.: Lending orientation to neural networks for cross-view geo-localization. In: CVPR
Shi, Y., Liu, L., Yu, X., Li, H.: Spatial-aware feature aggregation for image based cross-view geo-localization. NIPS 32 (2019)
Zheng, Z., Wei, Y., Yang, Y.: University-1652: a multi-view multi-source benchmark for drone-based geo-localization. In: ACM MM
Workman, S., Jacobs, N.: On the location dependence of convolutional neural network features. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops
Babenko, A., Lempitsky, V.: Aggregating local deep features for image retrieval. In: IEEE
Zheng, L., Yang, Y., Tian, Q.: Sift meets cnn: a decade survey of instance retrieval. IEEE 40(5)
Zheng, L., Shen, L., Tian, L., Wang, S., Wang, J., Tian, Q.: Scalable person re-identification: a benchmark. In: Proceedings of the IEEE International Conference on Computer Vision
Chaudhuri, U., Banerjee, B., Bhattacharya, A.: Siamese graph convolutional network for content based remote sensing image retrieval. Comput. Vision Image Understand. 184
Nair, L.R., Subramaniam, K., Prasannavenkatesan, G.: A review on multiple approaches to medical image retrieval system. In: Intelligent Computing in Engineering
Shi, Y., Yu, X., Liu, L., Zhang, T., Li, H.: Optimal feature transport for cross-view image geo-localization. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34
Shi, Y., Yu, X., Campbell, D., Li, H.: Where am I looking at? joint location and orientation estimation by cross-view matching. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition
Hu, S., Feng, M., Nguyen, R.M., Lee, G.H.: Cvm-net: cross-view matching network for image-based ground-to-aerial geo-localization. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition
Lin, T.Y., Cui, Y., Belongie, S., Hays, J.: Learning deep representations for ground-to-aerial geolocalization. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition
Tian, Y., Chen, C., Shah, M.: Cross-view image matching for geo-localization in urban environments. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition
Vo, N.N., Hays, J.: Localizing and orienting street views using overhead imagery. In: European Conference on Computer Vision
Workman, S., Souvenir, R., Jacobs, N.: Wide-area image geolocalization with aerial reference imagery. In: Proceedings of the IEEE International Conference on Computer Vision
Zhai, M., Bessinger, Z., Workman, S., Jacobs, N.: Predicting ground-level scene layout from aerial imagery. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition
Wang, T., Zheng, Z., Yan, C., Zhang, J., Sun, Y., Zheng, B., Yang, Y.: Each part matters: Local patterns facilitate cross-view geo-localization. IEEE (2021)
Zhang, X., Zhou, X., Lin, M., Sun, J.: Shufflenet: an extremely efficient convolutional neural network for mobile devices. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition
He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770–778 (2016)
Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Fei-Fei, L.: Imagenet: a large-scale hierarchical image database. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 248–255. IEEE (2009)
Chechik, G., Sharma, V., Shalit, U., Bengio, S.: Large scale online learning of image similarity through ranking. J. Mach. Learn. Res. 11(3) (2010)
Acknowledgements
This work was supported by National Natural Science Foundation of China (NSFC) under grant 61873067, and University-Industry Cooperation Project of Fujian Provincial Department of Science and Technology under grant 2020H6101.
Author information
Authors and Affiliations
Corresponding authors
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Wu, X., Ma, Q., Li, Q., Yu, Y., Liu, W. (2023). Cross-view Geo-localization Based on Cross-domain Matching. In: Xiong, N., Li, M., Li, K., Xiao, Z., Liao, L., Wang, L. (eds) Advances in Natural Computation, Fuzzy Systems and Knowledge Discovery. ICNC-FSKD 2022. Lecture Notes on Data Engineering and Communications Technologies, vol 153. Springer, Cham. https://doi.org/10.1007/978-3-031-20738-9_81
Download citation
DOI: https://doi.org/10.1007/978-3-031-20738-9_81
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-20737-2
Online ISBN: 978-3-031-20738-9
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)