Abstract
Waste management systems and their inherent problems are still a matter of great concern amid this cutting edge of science and technology. Piles of untreated waste in open locations contribute to a variety of issues, including air and water pollution, messy and unpleasant workplaces, waste of recyclable products, potential health risks for waste management workers, and so on. The root cause of this problem points to a single fact—too much manual labour involved in the garbage treatment process (collection, separation and recycling)—that can’t keep up to the pace with which garbage generation happens. An efficient recycling method is imperative to solve this problem, which can be achieved by a fast, real-time garbage detection and classification system. In this research, we will propose a novel Deep Learning-based approach for automatic detection and classification of five kinds of waste materials, namely Kitchen Waste, Glass Waste, Metal Waste, Paper Waste and Plastic Waste, from the garbage dump for an efficient recycling process, which not only improves the efficiency of the current manual approach but also provides a scalable solution to the problem. The contributions of this paper include a fully human labelled dataset consisting of 2,200 images of a garbage dump with 135,541 annotated objects from the aforementioned categories and a real-time garbage object localisation and classification framework based on a lightning-fast, fairly accurate and end-to-end deep learning algorithm. For the baseline, we have articulated a single-stage object detection and classification framework and initialised the detector training process with pre-trained weights (trained on MS COCO Dataset) of the feature extractor. Then, with some fine-tuning and employing a few transfer learning tricks, we have proposed a waste object detection framework that yields a mAP of 66.08% at an IoU threshold of 0.35 with an inference speed of roughly 55–58 milliseconds in a single GPU environment on both images and videos. It surpasses the performances of all the contemporary frameworks which deal with waste separation tasks.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Everingham, M., Van Gool, L., Williams, C.K., Winn, J., Zisserman, A.: The pascal visual object classes (voc) challenge. Int. J. Comput. Vision 88(2), 303–338 (2010)
Bahauddin, K., Uddin, M.: Prospect of solid waste situation and an approach of environ-mental management measure (emm) model for sustainable solid waste management: Case study of dhaka city. J. Environ. Sci. Nat. Resour. 5(1), 99–111 (2012)
Russel, M.M., Chowdhury, M., Shekh, M., Uddin, N., Newaz, A., Mehdi, M., Talukder, M.M.M.: Development of automatic smart waste sorter machine (2013)
Concern, W.: Bangladesh waste database 2014. Dhaka: Waste Concern 29, 2019 (2014). Accessed June
Girshick, R., Donahue, J., Darrell, T., Malik, J.: Rich feature hierarchies for accurate object detection and semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 580–587 (2014)
Lin, T.Y., Maire, M., Belongie, S., Hays, J., Perona, P., Ramanan, D., Dollár, P., Zitnick, C.L.: Microsoft coco: common objects in context. In: European Conference on Computer Vision, pp. 740–755. Springer (2014)
Abedin, M.A., Jahiruddin, M.: Waste generation and management in Bangladesh: an overview. Asian J. Med. Biol. Res. 1(1), 114–120 (2015)
Girshick, R.: Fast r-cnn. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1440–1448 (2015)
Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C.Y., Berg, A.C.: Ssd: single shot multibox detector. In: European Conference on Computer Vision, pp. 21–37. Springer (2016)
Mittal, G., Yagnik, K.B., Garg, M., Krishnan, N.C.: Spotgarbage: smartphone app to detect garbage using deep learning. In: Proceedings of the 2016 ACM International Joint Conference on Pervasive and Ubiquitous Computing, pp. 940–945 (2016)
Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: United, real-time object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 779–788 (2016)
Sudha, S., Vidhyalakshmi, M., Pavithra, K., Sangeetha, K., Swaathi, V.: An automatic classification method for environment: friendly waste segregation using deep learning. In: IEEE Technological Innovations in ICT for Agriculture and Rural Development (TIAR), vol 2016, pp. 65–70 . IEEE (2016)
Yang, M., Thung, G.: Classification of trash for recyclability status, CS229 Project Report, vol. 2016 (2016)
Awe, O., Mengistu, R., Sreedhar, V.: Smart trash net: waste localization and classifica-tion (2017)
He, K., Gkioxari, G., Dollár, P., Girshick, R.: Mask r-cnn. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2961–2969 (2017)
Lin, T.-Y., Goyal, P., Girshick, R., He, K., Dollár, P.: Focal loss for dense object detection. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2980–2988 (2017)
Rad, M.S., von Kaenel, A., Droux, A., Tieche, F., Ouerhani, N., Ekenel, H.K., Thiran, J.-P.: A computer vision system to localize and classify wastes on the streets. In: International Conference on Computer Vision Systems, pp. 195–204. Springer (2017)
Redmon, J., Farhadi, A.: Yolo9000: better, faster, stronger. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7263–7271 (2017)
Alfarrarjeh, A., Kim, S.H., Agrawal, S., Ashok, M., Kim, S.Y., Shahabi, C.: Image classification to determine the level of street cleanliness: a case study. In: 2018 IEEE Fourth International Conference on Multimedia Big Data (BigMM), pp. 1–5. IEEE (2018)
Aral, R.A., Keskin, Ş.R., Kaya, M., Hacıömeroğlu, M.: Classification of trashnet dataset based on deep learning models. In: 2018 IEEE International Conference on Big Data (Big Data), pp. 2058–2062. IEEE (2018)
Chu, Y., Huang, C., Xie, X., Tan, B., Kamal, S., Xiong, X.: Multilayer hybrid deep-learning method for waste classification and recycling. Comput. Intell. Neurosci. 2018 (2018)
Redmon, J., Farhadi, A.: Yolov3: an incremental improvement (2018). arXiv preprint arXiv:1804.02767
Kulkarni, H.N., Raman, N.K.S.: Waste object detection and classification (2019)
Sousa, J., Rebelo, A., Cardoso, J.S.: Automation of waste sorting with deep learning. In: XV Workshop de Visão Computacional (WVC), vol 2019, pp. 43–48. IEEE (2019)
Bochkovskiy, A., Wang, C.-Y., Liao, H.-Y.M.: Yolov4: optimal speed and accuracy of object detection (2020). arXiv preprint arXiv:2004.10934
Proença, P.F., Simões, P.: Taco: trash annotations in context for litter detection (2020). arXiv preprint arXiv:2003.06975
Wang, C.Y., Liao, H.Y.M., Wu, Y.H., Chen, P.Y., Hsieh, J.W., Yeh, I.H.: Cspnet: a new backbone that can enhance learning capability of cnn. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 390–391 (2020)
Tzutalin, Labelimg. https://github.com/tzutalin/labelImg
Vinodha, D., Sangeetha. J., Sherin, B.C., Renukadevi, M.: Smart garbage system with garbage separation using object detection
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Islam, S.M.Y., Alam, M.G.R. (2022). Computer Vision-based Waste Detection and Classification for Garbage Management and Recycling. In: Hossain, S., Hossain, M.S., Kaiser, M.S., Majumder, S.P., Ray, K. (eds) Proceedings of International Conference on Fourth Industrial Revolution and Beyond 2021 . Lecture Notes in Networks and Systems, vol 437. Springer, Singapore. https://doi.org/10.1007/978-981-19-2445-3_10
Download citation
DOI: https://doi.org/10.1007/978-981-19-2445-3_10
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-19-2444-6
Online ISBN: 978-981-19-2445-3
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)