Advertisement

Detection and Classification of Urban Actors Through TensorFlow with an Android Device

  • Andres CampoverdeEmail author
  • Gabriel Barros
Conference paper
Part of the Advances in Intelligent Systems and Computing book series (AISC, volume 1099)

Abstract

With the advancement of Artificial Intelligence, now it is possible to perform Neural Networks for detection and classification inside a device with limited hardware. We compare the performance of Deep Neural Networks (DNN). Model selected is Single Shot Detector (SSD) by default and re-trained, and frameworks are TensorFlow mobile versus TensorFlow lite. Default model contains 80 different classes of objects and the second one is our re-trained model with only 6 different classes based on urban actors (car, bus, truck, bicycle, motorcycle, person). The main goal is to build an object tracker for urban mobility with an Android device. Results are based in two metrics for object detection: mean Average Precision (mAP) and log-average miss rate. And for classification we report two metrics: precision (PRE) and recall (REC). We report inference time as an additional metric which is strongly related to hardware used. TensorFlow mobile is much slower than TensorFlow lite in terms of time of inference. Finally, re-trained model allows the integration of new scenarios, improving the detection rate.

Keywords

Object detection and classification Performance comparison Android middleware Convolutional Neural Networks TensoFlow Single Shot Detector 

References

  1. 1.
    Alsing, O.: Mobile Object Detection using TensorFlow Lite and Transfer Learning. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-233775 (2018)
  2. 2.
    Barros-Gavilanes, G.: Persons counter through Wi-Fi’s passive sniffing for IoT. In: 2018 IEEE Third Ecuador Technical Chapters Meeting (ETCM), pp. 1–6, October 2018.  https://doi.org/10.1109/ETCM.2018.8580283
  3. 3.
    Chen, Z., Ellis, T., Velastin, S.A.: Vehicle detection, tracking and classification in urban traffic. In: 2012 15th International IEEE Conference on Intelligent Transportation Systems, pp. 951–956, September 2012.  https://doi.org/10.1109/ITSC.2012.6338852
  4. 4.
    Dollár, P., Wojek, C., Schiele, B., Perona, P.: Pedestrian detection: an evaluation of the state of the art. IEEE Trans. Pattern Anal. Mach. Intell. 34(4), 743–761 (2012).  https://doi.org/10.1109/TPAMI.2011.155CrossRefGoogle Scholar
  5. 5.
    EdjeElectronics: How to train an object detection classifier for multiple objects using TensorFlow (GPU) on windows 10 (2018). https://github.com/EdjeElectronics/
  6. 6.
  7. 7.
    Huang, J., et al.: Speed/accuracy trade-offs for modern convolutional object detectors. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3296–3297, July 2017.  https://doi.org/10.1109/CVPR.2017.351
  8. 8.
    Ignatov, A., Timofte, R., Chou, W., Wang, K., Wu, M., Hartley, T., Van Gool, L.: AI benchmark: running deep neural networks on android smartphones. In: Leal-Taixé, L., Roth, S. (eds.) Computer Vision - ECCV 2018 Workshops, pp. 288–314. Springer International Publishing, Cham (2019)CrossRefGoogle Scholar
  9. 9.
    Lin, T.: Labelimg (2019). https://github.com/tzutalin/labelImg
  10. 10.
    Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C.Y., Berg, A.C.: SSD: single shot multibox detector. Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics). LNCS, vol. 9905, pp. 21–37 (2016).  https://doi.org/10.1007/978-3-319-46448-0_2CrossRefGoogle Scholar
  11. 11.
    Python Programming: Training custom object detector - TensorFlow object detection API tutorial (2017). https://pythonprogramming.net/introduction-use-tensorflow-object-detection-api-tutorial/
  12. 12.
    Raval, S.: Best laptop for machine learning (2018). https://www.youtube.com/watch?v=dtFZrFKMiPI
  13. 13.
    Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: unified, real-time object detection. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 779–788. IEEE Computer Society, Los Alamitos, June 2016.  https://doi.org/10.1109/CVPR.2016.91, https://www.computer.org/
  14. 14.
    Ren, S., He, K., Girshick, R., Sun, J.: Faster R-CNN: towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 39(6), 1137–1149 (2017).  https://doi.org/10.1109/TPAMI.2016.2577031CrossRefGoogle Scholar
  15. 15.
    Sha, Q.: Flutter tflite (2018). https://pub.dev/packages/tflite
  16. 16.
    Siekkinen, M.: Latency and Throughput Characterization of Convolutional Neural Networks for Mobile Computer Vision (2018). CoRR ABS/1803.0Google Scholar
  17. 17.
    Sobolevsky, S., Levitskaya, E., Chan, H., Postle, M., Kontokosta, C.: Impact Of Bike Sharing In New York City. arXiv preprint arXiv:1808.06606, pp. 1–26 (2018)
  18. 18.
    Takala, V., Pietikäinen, M.: Multi-object tracking using color, texture and motion. In: Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (2007).  https://doi.org/10.1109/CVPR.2007.383506
  19. 19.
    de TensorFlow, E.: Object detection (2019). https://www.tensorflow.org/lite/models/object_detection/overview
  20. 20.
    Torres, F., Barros, G., Barros, M.J.: Computer vision classifier and platform for automatic counting: more than cars. In: 2017 IEEE 2nd Ecuador Technical Chapters Meeting, ETCM 2017, June 2017 (2018).  https://doi.org/10.1109/ETCM.2017.8247454
  21. 21.
    Torres, F.A., Barros, G.: Sound noise monitoring platform: smart-phones as sensors. In: European Wireless 2017 - 23rd European Wireless Conference (2017)Google Scholar
  22. 22.
    Verdugo-Romero, W., González-Delgado, L., Sacoto-Cabrera, E.J., Verdugo-Cabrera, A., Verdugo-Cabrera, O.: Mathematical and statistical analysis for the simulation of the vehicular flow in a specific sector of the city of cuenca-ecuador. In: 2018 IEEE Third Ecuador Technical Chapters Meeting (ETCM), pp. 1–4, October 2018.  https://doi.org/10.1109/ETCM.2018.8580275

Copyright information

© Springer Nature Switzerland AG 2020

Authors and Affiliations

  1. 1.LIDI, Universidad del AzuayCuencaEcuador
  2. 2.tivo.ec Research Institute. Don Bosco 2-07CuencaEcuador

Personalised recommendations