Advertisement

Computer Vision and Deep Learning-Enabled UAVs: Proposed Use Cases for Visually Impaired People in a Smart City

Conference paper
Part of the Communications in Computer and Information Science book series (CCIS, volume 1089)

Abstract

Technological research and innovation have advanced at a rapid pace in recent years, and one group hoping to benefit from this, is visually impaired people (VIP). Technology may enable them to find new ways of travelling around smart cities, thus improving their quality of life (QoL). Currently, there are approximately 110 million VIP worldwide, and continuous research is crucial to find innovative solutions to their mobility problems. Recent advances such as the increase in Unmanned Aerial Vehicles (UAVs), smartphones and wearable devices, together with an ever-growing uptake of deep learning, computer vision, the Internet of Things (IoT), and virtual and augmented reality (VR)/(AR), have provided VIP with the hope of having an improved QoL. In particular, indoor and outdoor spaces could be improved with the use of such technologies to make them suitable for VIP. This paper examines use cases both indoors and outdoors and provides recommendations of how deep learning and computer vision-enabled UAVs could be employed in smart cities to improve the QoL for VIP in the coming years.

Keywords

Deep learning Computer vision UAVs Drone Visually impaired people (VIP) Smart city 

Notes

Acknowledgement

Dr. Moustafa Nasralla would like to acknowledge the Department of Communications and Networks Engineering at Prince Sultan University (PSU) for the valued support and research environmental provision which have led to completing this work. Dr. Ikram Ur Rehman would like to thank the School of Computing and Engineering, University of West London for its support and research provision. Dr. Drishty Sobnath would like to thank the Research Innovation and Enterprise department at Solent University for its continuous support that led to the completion of this study. Dr. Sara Paiva would like to thank the Instituto Politécnico de Viana do Castelo for its continuous support that led to the completion of this study.

References

  1. 1.
    Texeira, C., Toledo, A.S., da Silva Amorim, A., Kofuji, S.T., Rogério dos Santos, V.: Visual impairment and smart cities: perspectives on mobility. JOJ Ophthalmol. 3 (2017)Google Scholar
  2. 2.
    Ramadhan, A.J.: Wearable smart system for visually impaired people. Sensors 13(3), 834 (2018)Google Scholar
  3. 3.
    Gibson, D.V., Kozmetsky, G., Smilor, R.W.: The Technopolis Phenomenon: Smart Cities, Fast Systems, Global Networks. Rowman & Littlefield Publishers, Lanham (1992)Google Scholar
  4. 4.
    Skouby, K.E., Kivimäki, A., Haukipuro, L., Lynggaard, P., Windekilde, I.: Smart cities and the ageing population. In: 32nd Meeting of WWRF, Marrakech, Morocco. https://pdfs.semanticscholar.org/d7a5/84f867996dbdf78a34697523c537dae218bc.pdf
  5. 5.
    European Commission: General Assembly of the European Innovation Partnership on Smart Cities and Communities (EIP-SCC) — European Commission (2018)Google Scholar
  6. 6.
    Philip, N.Y., Rehman, I.U.: Towards 5G health for medical video streaming over small cells. In: Kyriacou, E., Christofides, S., Pattichis, C. (eds.) XIV Mediterranean Conference on Medical and Biological Engineering and Computing 2016. IFMBE Proceedings, vol. 57. Springer, Cham (2016).  https://doi.org/10.1007/978-3-319-32703-7_215CrossRefGoogle Scholar
  7. 7.
    Avila, M., Funk, M., Henze, N.: DroneNavigator: using drones for navigating visually impaired persons. In: Proceedings of the 17th International ACM SIGACCESS Conference on Computers & Accessibility, pp. 327–328 (2015)Google Scholar
  8. 8.
    Voulodimos, A., Doulamis, N., Doulamis, A., Protopapadakis, E.: Deep learning for computer vision: a brief review. Comput. Intell. Neurosci. Article ID 7068349 (2018).  https://doi.org/10.1155/2018/7068349Google Scholar
  9. 9.
    Liang, M., Hu, X.: Recurrent convolutional neural network for object recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3367–3375 (2015)Google Scholar
  10. 10.
    Hinton, G.E., Sejnowski, T.J.: Learning and relearning in Boltzmann machines. In: Parallel Distributed Processing: Explorations in the Microstructure of Cognition, pp. 282–317 (1986)Google Scholar
  11. 11.
    Baltrušaitis, T., Ahuja, C., Morency, L.P.: Multimodal machine learning: a survey and taxonomy. IEEE Trans. Pattern Anal. Mach. Intell. 41(2), 423–443 (2018)CrossRefGoogle Scholar
  12. 12.
    Carrio, A., Sampedro, C., Rodriguez-Ramos, A., Campoy, P.: A review of deep learning methods and applications for unmanned aerial vehicles. J. Sens. 2017, Article ID 3296874, 13 p. (2017).  https://doi.org/10.1155/2017/3296874CrossRefGoogle Scholar
  13. 13.
    Hung, C., Xu, Z., Sukkarieh, S.: Feature learning based approach for weed classification using high resolution aerial images from a digital camera mounted on a UAV. Remote Sens. 6(12), 12037–12054 (2014)CrossRefGoogle Scholar
  14. 14.
    Jung, S., Hwang, S., Shin, H., Shim, D.H.: Perception, guidance, and navigation for indoor autonomous drone racing using deep learning. IEEE Robot. Autom. Lett. 3(3), 2539–2544 (2018)CrossRefGoogle Scholar
  15. 15.
    Cardin, S., Thalmann, D., Vexo, F.: A wearable system for mobility improvement of visually impaired people. Vis. Comput. 23(2), 109–118 (2007)CrossRefGoogle Scholar
  16. 16.
    Dunai, L.D., Lengua, I.L., Tortajada, I., Simon, F.B.: Obstacle detectors for visually impaired people. In: 2014 International Conference on Optimization of Electrical and Electronic Equipment, OPTIM (2014)Google Scholar
  17. 17.
    Gharani, P., Karimi, H.A.: Context-aware obstacle detection for navigation by visually impaired. Image Vis. Comput. 64, 103–115 (2017)CrossRefGoogle Scholar
  18. 18.
    Jonnalagedda, A., et al.: Enhancing the Safety of Visually Impaired Travelers in and around Transit Stations (2014)Google Scholar
  19. 19.
    Lee, C.L., Chen, C.Y., Sung, P.C., Lu, S.Y.: Assessment of a simple obstacle detection device for the visually impaired. Appl. Ergon. 45(4), 817–824 (2014)CrossRefGoogle Scholar
  20. 20.
    Poggi, M., Mattoccia, S.: A wearable mobility aid for the visually impaired based on embedded 3D vision and deep learning. In: Proceedings - IEEE Symposium on Computers and Communications (2016)Google Scholar
  21. 21.
    Rahman, M.M., Poon, B., Amin, M.A., Yan, H.: Support system using microsoft kinect and mobile phone for daily activity of visually impaired. In: Yang, G.-C., Ao, S.-I., Huang, X., Castillo, O. (eds.) Transactions on Engineering Technologies, pp. 425–440. Springer, Dordrecht (2015).  https://doi.org/10.1007/978-94-017-9588-3_32CrossRefGoogle Scholar
  22. 22.
    Patricio, M., Haidee, L., Ciro, L., Telma, R., Felipe, F.: Analysis and proposed improvements in the support for the visually impaired in the use of public transportation. In: SMART 2015: The Fourth International Conference on Smart Systems, Devices and Technologies (2015). ISBN 978-1-61208-414-5Google Scholar

Copyright information

© Springer Nature Switzerland AG 2019

Authors and Affiliations

  1. 1.Department of Communications and Networks EngineeringPrince Sultan UniversityRiyadhSaudi Arabia
  2. 2.School of Computing and EngineeringUniversity of West LondonLondonUK
  3. 3.Research, Innovation and EnterpriseSolent UniversitySouthamptonUK
  4. 4.ARC4DigiTInstituto Politécnico de Viana do CasteloViana do CasteloPortugal

Personalised recommendations