Skip to main content

An Efficient Image Dehazing Technique Using DSRGAN and VGG19

  • Conference paper
  • First Online:
Applied Machine Learning and Data Analytics (AMLDA 2023)

Abstract

Haze and clouds can cause the weather to affect the clarity and contrast of photos that are taken by cameras, air quality, and other reasons. To address the issue of haze, image dehazing has become an area of significant importance. However, many current techniques for unsupervised picture dehazing rely on simplified atmospheric scattering models and a priori knowledge, which can result in inaccuracies and poor dehazing performance. The study of image-dehazing techniques has shown promise Generative Adversarial Networks (GANs) were developed. Unfortunately, because haze is so complicated, the bi-directional mappings domain translation techniques currently used in unsupervised GANs are not suitable for dehazing image work. With haze-free photos, the image may become distorted, lose image details, or retain image features poorly as a result. The paper makes recommendations for an end-to-end unsupervised image-dehazing system that is used to address these problems. Deep Super-Resolution Generative Adversarial Network (DSRGAN) that uses VGG19 for feature extraction. The problem addressed in this project is to create a novel picture dehazing method that uses the DSRGAN and VGG19 models to successfully remove haze from photographs. The specific objectives of this project are to implement the DSRGAN and VGG19 models, modify them for image dehazing, acquire a dataset of hazy images, preprocess the dataset to remove any anomalies, train the models on the dataset to learn the features and characteristics of hazy images and assess the performance of the models using both qualitative and quantitative metrics. The proposed method showed an improvement in both PSNR and SSIM metrics with 24.06 and 0.912. The proposed system achieves negligible generator loss and discriminator loss, offering a promising solution to the challenge of image dehazing in complex and dynamic environments.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 59.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 79.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Negru, M., Nedevschi, S., Peter, R.I.: Exponential contrast restoration in fog conditions for driving assistance. IEEE Trans. Intell. Transp. Syst. 16(4), 2257–2268 (2015)

    Article  Google Scholar 

  2. Min, X., Zhai, G., Gu, K., Yang, X., Guan, X.: Objective quality evaluation of dehazed images. IEEE Trans. Intell. Transp. Syst. 20(8), 2879–2892 (2019)

    Article  Google Scholar 

  3. Peters, J.R., Surana, A., Taylor, G.S., Turpin, T.S., Bullo, F.: UAV surveillance under visibility and dwell-time constraints: a sampling based approach (2019). arXiv:1908.05347

  4. Choi, D.-Y., Choi, J.-H., Choi, J., Song, B.C.: Sharpness enhancement and super-resolution of around-view monitor images. IEEE Trans. Intell. Transp. Syst. 19(8), 2650–2662 (2018)

    Article  Google Scholar 

  5. Nasir, M., Muhammad, K., Lloret, J., Sangaiah, A.K., Sajjad, M.: Fog computing enabled cost-effective distributed summarization of surveillance videos for smart cities. J. Parallel Distrib. Comput. 126, 161–170 (2019)

    Article  Google Scholar 

  6. Makwana, Y., Iyer, S.S., Tiwari, S.: The food recognition and nutrition assessment from images using artificial intelligence: a survey. ECS Trans. 107(1), 3547 (2022)

    Article  Google Scholar 

  7. Zhao, S., Fang, Y., Qiu, L.: Deep learning-based channel estimation with SRGAN in OFDM systems. In: 2021 IEEE Wireless Communications and Networking Conference (WCNC), China, pp. 1–6 (2021). https://doi.org/10.1109/WCNC49053.2021.9417242

  8. Kim, J., Lee, J.K., Lee, K.M.: Accurate image super-resolution using very deep convolutional networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1646–1654 (2016)

    Google Scholar 

  9. Xu, Z., Liu, X., Chen, X.: Fog removal from video sequences using contrast limited adaptive histogram equalization. In: Proceedings of the 2009 International Conference on Computational Intelligence and Software Engineering, Washington, DC, USA, 11–14 December 2009, pp. 1–4 (2009)

    Google Scholar 

  10. Nayar, S.K., Narasimhan, S.G.: Vision in bad weather. In: Proceedings of the Proceedings of the Seventh IEEE International Conference on Computer Vision, Corfu, Greece, 20–27 September 1999, vol. 2, pp. 820–827 (1999)

    Google Scholar 

  11. Khan, H., et al.: Localization of radiance transformation for image dehazing in wavelet domain. Neurocomputing 381, 141–151 (2020)

    Article  Google Scholar 

  12. Cai, B., Xu, X., Jia, K., Qing, C., Tao, D.: Dehazenet: an end-to-end system for single image haze removal. IEEE Trans. Image Process. 25, 5187–5198 (2016)

    Article  MathSciNet  Google Scholar 

  13. Tan, R.T.: Visibility in bad weather from a single image. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Anchorage, AK, USA, 23–28 June 2008; IEEE: Piscataway, NJ, USA, 2008; pp. 1–8 (2008)

    Google Scholar 

  14. He, K., Sun, J., Tang, X.: Single image haze removal using dark channel prior. IEEE Trans. Pattern Anal. Mach. Intell. 33, 2341–2353 (2010)

    Google Scholar 

  15. Berman, D., Treibitz, T., Avidan, S.: Non-local image dehazing. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2016, pp. 1674–1682 (2016)

    Google Scholar 

  16. Liu, X., Ma, Y., Shi, Z., Chen, J.: Griddehazenet: attention-based multi-scale network for image Dehazing. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; IEEE/CVF: Piscataway, NJ, USA, pp. 7314–7323 (2019)

    Google Scholar 

  17. Li, B., Peng, X., Wang, Z., Xu, J., Feng, D.: Aod-net: all-in-one Dehazing network. In: Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; IEEE: Piscataway, NJ, USA, pp. 4770–4778 (2017)

    Google Scholar 

  18. Zhang, H., Patel, V.M.: Densely connected pyramid Dehazing network. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; IEEE: Piscataway, NJ, USA, pp. 3194–3203 (2018)

    Google Scholar 

  19. Huang, G., Liu, Z., Van Der Maaten, L., Weinberger, K.Q.: Densely connected convolutional networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; IEEE: Piscataway, NJ, USA, pp. 4700–4708 (2017)

    Google Scholar 

  20. Ronneberger, O., Fischer, P., Brox, T.: U-Net: convolutional networks for biomedical image segmentation. In: Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F. (eds.) MICCAI 2015. LNCS, vol. 9351, pp. 234–241. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-24574-4_28

    Chapter  Google Scholar 

  21. Goodfellow, I., et al.: Generative adversarial nets. In: Proceedings of the Twenty-Eighth Conference on Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014, pp. 2672–2680 (2014)

    Google Scholar 

  22. Wang, K., Gou, C., Duan, Y., Lin, Y., Zheng, X., Wang, F.-Y.: Generative adversarial networks: introduction and outlook. IEEE/CAA J. Autom. Sinica 4(4), 588598 (2017). https://doi.org/10.1109/jas.2017.751058310.1109/JAS.2017.7510583

    Article  MathSciNet  Google Scholar 

  23. Arjovsky, M., Chintala, S., Bottou, L.: Wasserstein generative adversarial networks. In: Proceedings of the International Conference on Machine Learning, PMLR: Sydney, Australia, 6–11 August 2017, pp. 214–223 (2017)

    Google Scholar 

  24. Miyato, T., Kataoka, T., Koyama, M., Yoshida, Y.: Spectral Normalization for Generative Adversarial Networks. arXiv arXiv:1802.05957 (2018)

  25. Mao, X., Li, Q., Xie, H., Lau, R.Y., Wang, Z., Paul Smolley, S.: Least squares generative adversarial networks. In: Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017, pp. 2794–2802 (2017)

    Google Scholar 

  26. Qu, Y., Chen, Y., Huang, J., Xie, Y.: Enhanced pix2pix dehazing network. In: Proceedings of IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2019, pp. 8160–8168 (2019)

    Google Scholar 

  27. Creswell, A., et al.: Generative adversarial networks: an overview. IEEE Signal Process. Mag. 35(1), 53–65 (2018)

    Article  Google Scholar 

  28. Ledig, C., et al.: Photo-realistic single image super-resolution using a generative adversarial network. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2017)

    Google Scholar 

  29. Goodfellow, I., et al.: Generative adversarial networks. In: Advances in Neural Information Processing Systems, pp. 2672–2680 (2014)

    Google Scholar 

  30. Lv, B., Liu, Y., Zhang, S., Zeng, H., Zhu, G.: Super Resolution with Generative Adversarial Networks (2018)

    Google Scholar 

  31. Tanwar, R., Phukan, O.C., Singh, G., Tiwari, S.: CNN-LSTM Based Stress Recognition Using Wearables (2022)

    Google Scholar 

  32. Nalwar, S., et al.: EffResUNet: encoder decoder architecture for cloud-type segmentation. Big Data Cogn. Comput. 6(4), 150 (2022)

    Article  Google Scholar 

  33. Mane, D., Shah, K., Solapure, R., Bidwe, R., Shah, S.: Image-based plant seedling classification using ensemble learning. In: Pati, B., Panigrahi, C.R., Mohapatra, P., Li, K.C. (eds.) ICACIE 2021, pp. 433–447. Springer, Cham (2022). https://doi.org/10.1007/978-981-19-2225-1_39

  34. https://www.kaggle.com/datasets/balraj98/indoor-training-set-its-residestandard

  35. Hotkar, O., Radhakrishnan, P., Singh, A., Jhamnani, N., Bidwe, R.V.: U-net and YOLO: AIML models for lane and object detection in real-time. In: Proceedings of the 2023 Fifteenth International Conference on Contemporary Computing, pp. 467–473 (2023)

    Google Scholar 

  36. Agrawal, G., Jha, U., Bidwe, R.: Automatic facial expression recognition using advanced transfer learning. In: Proceedings of the 2023 Fifteenth International Conference on Contemporary Computing, pp. 450–458 (2023)

    Google Scholar 

  37. Bidwe, R.V., Mishra, S., Bajaj, S.: Performance evaluation of transfer learning models for ASD prediction using non-clinical analysis. In: Proceedings of the 2023 Fifteenth International Conference on Contemporary Computing, pp. 474–483 (2023)

    Google Scholar 

  38. Bidwe, R.V., et al.: Deep learning approaches for video compression: a bibliometric analysis. Big Data Cogn. Comput. 6(2), 44 (2022)

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Ranjeet Vasant Bidwe .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2024 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Jadav, B., Mishra, S., Bagane, P., Bidwe, R.V. (2024). An Efficient Image Dehazing Technique Using DSRGAN and VGG19. In: Jabbar, M.A., Tiwari, S., Ortiz-Rodríguez, F., Groppe, S., Bano Rehman, T. (eds) Applied Machine Learning and Data Analytics. AMLDA 2023. Communications in Computer and Information Science, vol 2047. Springer, Cham. https://doi.org/10.1007/978-3-031-55486-5_7

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-55486-5_7

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-55485-8

  • Online ISBN: 978-3-031-55486-5

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics