Skip to main content
Log in

M-PFGMNet: multi-pose feature generation mapping network for visual object tracking

  • Published:
Multimedia Tools and Applications Aims and scope Submit manuscript

Abstract

The occlusion and low-resolution environments lead to insufficient feature data and redundant calculations, which affects the accuracy and timeliness of the unmanned detection system in visual object tracking. To solve this issue, this paper proposes the multi-pose feature generation mapping network for visual object tracking (M-PFGMNet). Firstly, M-PFGMNet suggests a distributed generative adversarial network to increase the accuracy of target detection and reconstruct the local-to-global features of the data with insufficient target features. Secondly, M-PFGMNet proposes a multi-pose feature mapping method for real-time target detection, this method can migrate sample data with high similarity between dynamic target and unmanned system. The experimental results show that the algorithm is effective compared with other similar algorithms in the OTB-50/100 public dataset.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4

Similar content being viewed by others

References

  1. Cai L, Luo P, Zhou G et al (2019) Dynamic target detection method based on multi-perspective light field reconstruction. International Journal of Distributed Sensor Networks 15(8):1–12

    Article  Google Scholar 

  2. Chang J, Scherer S (2017) Learning representations of emotional speech with deep convolutional generative adversarial networks. In: IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), New Orleans, LA , vol 2017, pp 2746–2750

  3. Chen H, Wang Y, Wang G et al (2020) Progressive object transfer detection. IEEE Trans Image Process 29:986–1000

    Article  MathSciNet  Google Scholar 

  4. Chen X, Yu J, Kong S et al (2019) Towards Real-Time advancement of underwater visual quality with GAN. IEEE Trans Ind Electron 66 (12):9350–9359

    Article  Google Scholar 

  5. Fan Q, Zhuo W, Tang CK et al (2020) Few-shot Object Detection With attention-RPN and Multi-Relation Detector. In: 2020 IEEE/CVF conference on computer vision and pattern recognition (CVPR), vol 2020. IEEE, pp 4012–4021

  6. Goodfellow IJ, Pouget-Abadie J, Mirza M et al (2014) Generative adversarial nets. In: International conference on neural information processing systems, vol 2014. MIT Press, pp 2672–2680

  7. Hu Q, Zhai L (2019) RGB-D image Multi-Target detection method based on 3D DSF r- CNN. International Journal of Pattern Detection and Artificial Intelligence 33(8):1–15

    Google Scholar 

  8. Huang B, Chen W, Wu X, et al. (2018) High-quality face image generated with conditional boundary equilibrium generative adversarial networks. Pattern Detection Letters 111:72–79

    Article  Google Scholar 

  9. Karbalayghareh A, Qian X, Dougherty ER (2018) Optimal bayesian transfer learning. IEEE Transactions on Signal Processing 66(14):3724–3739

    Article  MathSciNet  Google Scholar 

  10. Lu H, Fu X, Liu C et al (2017) Cultivated land information extraction in UAV imagery based on deep convolutional neural network and transfer learning. Journal of Mountain Science 14(4):731–741

    Article  Google Scholar 

  11. Martin D, Gustav H, Fahad SK et al (2015) Learning spatially regularized correlation filters for visual tracking. In: IEEE International conference on computer vision (ICCV), Santiago, vol 2015, pp 4310–4318

  12. Mo K, Li S, Zhang Y et al (2016) Personalizing a Dialogue System with Transfer Reinforcement Learning. In: 32nd AAAI conference on artificial intelligence, vol 2016, pp 5317–5324

  13. Nam H, Han B (2016) Learning Multi-Domain convolutional neural networks for visual tracking. In: IEEE conference on computer vision and pattern detection (CVPR), las vegas, NV, vol 2016, pp 4293–4302

  14. Pan SJ, Yang Q (2010) A survey on transfer learning. IEEE Trans Knowl Data Eng 22(10):1345–1359

    Article  Google Scholar 

  15. Qin F, Xu D, Zhang D et al (2019) Robotic skill learning for precision assembly with microscopic vision and force feedback. IEEE/ASME Transactions on Mechatronics 24(3):1117–1128

    Article  Google Scholar 

  16. Ravishankar H, Sudhakar P, Venkataramani R et al (Sept 2016) Understanding the mechanisms of deep transfer learning for medical images. Deep Learning and Data Labeling for Medical Applications 10008:188–196

  17. Saha B, Gupta S, Phung D et al (2016) Multiple task transfer learning with small sample sizes. Knowl Inf Syst 46(2):315–342

    Article  Google Scholar 

  18. Shi W, Ragunathan R (2020) Point-GNN: Graph Neural Network for 3D Object Detection in a Point Cloud. In: 2020 IEEE/CVF conference on computer vision and pattern recognition (CVPR), vol 2020. IEEE, pp 1708–1716

  19. Sun Y, Liu X, Cong P et al (2018) Digital radiography image denoising using a generative adversarial network. Journal of X-Ray Science and Technology 26(2):1–12

    Google Scholar 

  20. Wu S, Wang S, Laganière R, et al. (2018) Exploiting target data to learn deep convolutional networks for Scene-Adapted human detection. IEEE Trans Image Process 27(3):1418–1432

    Article  MathSciNet  Google Scholar 

  21. Yin X, Liu G, Ma X (2020) Fast Scale Estimation Method in Object Tracking. IEEE Access 2020(8):31057–31068

    Article  Google Scholar 

  22. Yuan Z, Li H, Liu J et al (2020) Multiview scene image inpainting based on conditional generative adversarial networks. IEEE Transactions on Intelligent Vehicles 5(2):314–323

    Article  Google Scholar 

  23. Zhang D, Shao J, Hu G et al (2017) Sharp and real image Super-Resolution using generative adversarial network. Neural Information Processing 10636:217–226

    Google Scholar 

  24. Zhao J, Mathieu M, Lecun Y (2016) Energy-based generative adversarial network. In: International conference on learning representations (ICLR), vol 2016, pp 1–17

  25. Zheng X, Ravishankar S, Long Y et al (2018) PWLS-ULTRA: An efficient clustering and Learning-Based approach for Low-Dose 3D CT image reconstruction. IEEE Trans Med Imaging 37(6):1498–1510

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Lei Cai.

Ethics declarations

Competing interests/Competing interests

The authors declared no potential Conflicts of Interests/Competing interests with respect to the research, authorship, and publication of this article.

Additional information

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

This work was supported by National Key R&D Program of China (2019YFB1311002).

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Cai, L., Luo, P., Xu, T. et al. M-PFGMNet: multi-pose feature generation mapping network for visual object tracking. Multimed Tools Appl 81, 38803–38816 (2022). https://doi.org/10.1007/s11042-022-12875-3

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11042-022-12875-3

Keywords

Navigation