The Enriched Feature Enhancement Technique for Satellite Image Based on Transforms Using PCNN

  • K. HariharanEmail author
  • N. R. Rajaan
  • Peter Pethuru Raj Chelliah
  • Malini Deepika


The features of the satellite images can be improved by fusing or combining two images with complementary property. By fusing these two images the spatial property of the resultant image is improved. Satellite images are one of the agents that give the features of the earth’s surface. Processing these satellite images will provide more geographical information hidden in the images. This research paper have an detailed insight study of two types of the satellite images one is Panchromatic (PAN) and other Multispectral (MS). The PAN image with high spatial resolution and MS image with spectral resolution are fused to get better resultant output. For fusion process Nonsubsampled Contour let Transform is used to decompose the images into low and high frequency values. Pulse Coupled Neural Network is used to motivate the low frequency pixel and Morphological filter is applied to the edge detected image for finding the features in the images. This is an real time transformations which will give better results in SAR image processing, video processing, stereo based reconstruction of depth and width of the features present in the image.


Subsampled contourlet transform (SCT) Pulse Coupled Neural Network (PCNN) Panchromatic (PAN) Multispectral (MS) 



Author would like to thank Mr. Manickavasagam, Director, DRDL, Hydrabad and Mr Peter Pethuru Raj, CEO, SRE, Reliance Jio Cloud, Banglore, for their continuous support throughout the work.


  1. 1.
    Do, M. N., & Vetterli, M. (2005). The contourlet transform: An efficient directional multi-resolution image representation. IEEE Transaction on Image Processing,14(12), 2001–2106.CrossRefGoogle Scholar
  2. 2.
    Tu, T. M., Cheng, W. C., Cheng, C. P., & Hirang, P. S. (2007). Best trade off for high resolution image fusion to preserve spatial details and minimize colour distortion. IEEE Geoscience and Remote Sensing Letters,4(2), 302–306.CrossRefGoogle Scholar
  3. 3.
    Vaithyanathan, V. (2012). An efficient method to improve the spatial property of the medical images. Journal of Theoretical and Applied Information Technology,35(2), 141–148.Google Scholar
  4. 4.
    Do, M. N., & Vetterli, M. (2005). The contourlet transform: An efficient directional multiresolution image representation. IEEE Transactions on Image Processing,14(12), 2091–2106.CrossRefGoogle Scholar
  5. 5.
    Krishnamoorthy, S., & Soman, K. P. (2010). Implementation and comparative study of image fusion algorithms. International Journal of Computer Applications,9(2), 25–35.CrossRefGoogle Scholar
  6. 6.
    Naraghi, M. G., Koohi, M., & Shakery, A. (2011). Edge detection in multispectral images based on structural elements. The International Journal of Multimedia and Its Application,3(1), 90–99.CrossRefGoogle Scholar
  7. 7.
    Zhang, W., Yang, J., Wang, X., & Yang, Q. (2009). The fusion of remote sensing images based on lifting wavelet transformation. Computer and Information Science,2(1), 69–75.Google Scholar
  8. 8.
    Plellar, G. (2003). A general framework for multi-resolution image fusion from pixel to region image fusion. Information Fusion,4, 259–280.CrossRefGoogle Scholar
  9. 9.
    Ravichandran, C. G., Rubesh Selvakumar, R., & Goutham, S. (2011). Analysis and comparison of medical image fusion techniques: Wavelet based Transform and contourlet based transforms. International Journal of Computer Science and Information Security,9(3), 70.Google Scholar
  10. 10.
    Deshmukh, M., & Bhosale, U. (2010). Image fusion and image quality assessment of fused images. International Journal of Image Processing,4(5), 484–505.Google Scholar
  11. 11.
    Maini, R., & Aggarwal, H. (2008). Study and comparison of various image edge detection techniques. International Journal of Image Processing,3(1), 1–12.Google Scholar
  12. 12.
    Bacher, U., & Mayer, H. (2005). Automatic road extraction from multispectral high resolution satellite images (Vol. XXXVI, pp. 29–34). Part 3/W24 Vienna, Austria, August 29–30, 2005.Google Scholar
  13. 13.
    Nadernejad, E., Sharifzadeh, S., & Hassanpour, H. (2008). Edge detection techniques: evaluations and comparisons. Applied Mathematical Sciences,2(31), 1507–1520.MathSciNetzbMATHGoogle Scholar
  14. 14.
    Naraghi, M. G., Koohi, M., & Shakery, A. (2011). Edge detection in multispectral images based on structural elements. The International Journal of Multimedia and Its Applications,3(1), 90–99.CrossRefGoogle Scholar
  15. 15.
    Johnson, J. L., & Padgett, M. L. (1999). PCNN models and applications. IEEE Transactions on Neural Networks,10(3), 480–498.CrossRefGoogle Scholar
  16. 16.
    He, X., Li, J., Wei, D., Jia, W., & Wu, Q. (2009). Canny edge detection on a virtual hexagonal image structure. In: Advanced concepts for intelligent vision systems (pp. 233–244), 978-1-4244-5228-6/09/$26.00 c2009 IEEE.Google Scholar
  17. 17.
    Mallat, S. G. (1989). Mutifrequency channel decomposition of images and wavelet models. IEEE Transaction on Acoustic, Speech and Signal Processing,37(12), 2091–2110.CrossRefGoogle Scholar
  18. 18.
    Cotterill, R. M. J. (Ed.). (1989). Models of brain function. Cambridge: Cambridge University Press.zbMATHGoogle Scholar
  19. 19.
    Cannady, J. F. (1983). Finding edges and lines in images. M.S. thesis, Massachusetts Institute of Technology, Artificial Intelligence Lab, Cambridge.Google Scholar

Copyright information

© Springer Science+Business Media, LLC, part of Springer Nature 2020

Authors and Affiliations

  • K. Hariharan
    • 1
    Email author
  • N. R. Rajaan
    • 2
  • Peter Pethuru Raj Chelliah
    • 3
  • Malini Deepika
    • 2
  1. 1.School of ComputingSASTRA Deemed UniversityThanjavurIndia
  2. 2.School of Electrical and Electronics Communication EngineeringSASTRA Deemed UniversityThanjavurIndia
  3. 3.(SRE) DivisionReliance Jio CloudBangloreIndia

Personalised recommendations