Skip to main content

Feature Synthesization for Real-Time Pedestrian Detection in Urban Environment

  • Conference paper
  • First Online:
Advances in Multimedia Information Processing – PCM 2018 (PCM 2018)

Part of the book series: Lecture Notes in Computer Science ((LNISA,volume 11165))

Included in the following conference series:

  • 2399 Accesses

Abstract

Real-time pedestrian detection is very essential for auto assisted driving system. For improving the accuracy, more and more complicate features are proposed. However, most of them are impracticable for the real-world application because of high computation complexity and memory consumption, especially for onboard embedding system in the unmanned vehicle. In this paper, a novel framework that utilizes reconstruction sparsity to synthesize the feature map online is proposed for real-time pedestrian detection for the early warning system of the unmanned vehicle in real world. In this framework, the feature map is computed by sparse line combination of the representative coefficient and the feature response of trained basis which is learned offline. The efficiency of our method only depends on the dictionary decomposition no matter how complicated the feature is. Moreover, our method is suitable for most of the known complicate features. Experiments on four challenging datasets: Caltech, INRIA, ETH and TUD-Brussels, demonstrate that our proposed method is much efficient (more than 10 times acceleration) than the state-of-the-art approaches with comparable accuracy.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Navneet, D., Bill, T.: Histograms of oriented gradients for human detection. In: Proceedings of the 22nd IEEE Conference on Computer Vision and Pattern Recognition, pp. 886–893, June 2005

    Google Scholar 

  2. Sabzmeydani, P., Greg, M.: Detecting pedestrians by learning shapelet features. In: Proceedings of the 24th IEEE Conference on Computer Vision and Pattern Recognition, pp. 1–8, June 2007

    Google Scholar 

  3. Bo, W., Ram, N.: Detection of multiple, partially occluded humans in a single image by bayesian combination of edgelet part detectors. In: Proceedings of the Tenth IEEE International Conference on Computer Vision, pp. 90–97, June 2005

    Google Scholar 

  4. Pedro, F., David, M., et al.: A discriminatively trained, multiscale, deformable part model. In: Proceedings of the 25th IEEE Conference on Computer Vision and Pattern Recognition, pp. 1–8, June 2008

    Google Scholar 

  5. Piotr, D., Zhuowen, T., et al.: Integral channel features. In: Proceedings of the 20th British Machine Vision Conference, pp. 250–258, September 2009

    Google Scholar 

  6. Ahonen, T., Hadid, A., Pietikäinen, M.: Face recognition with local binary patterns. In: Pajdla, T., Matas, J. (eds.) ECCV 2004. LNCS, vol. 3021, pp. 469–481. Springer, Heidelberg (2004). https://doi.org/10.1007/978-3-540-24670-1_36

    Chapter  Google Scholar 

  7. Xiaoyu, W., Tony, X., et al.: An HOG-LBP human detector with partial occlusion handling. In: Proceedings of the 26th IEEE Conference on Computer Vision and Pattern Recognition, pp. 32–39, June 2009

    Google Scholar 

  8. Junjie, Y., Zhen, L., et al.: The fastest deformable part model for object detection. In: Proceedings of the 32nd IEEE Conference on Computer Vision and Pattern Recognition, pp. 2497–2504, June 2014

    Google Scholar 

  9. Pedro, F., Ross, B., et al.: Cascade object detection with deformable part models. In: Proceedings of the 28th IEEE Conference on Computer Vision and Pattern Recognition, pp. 2241–2248, June 2010

    Google Scholar 

  10. Uijlings, J. R., Van De Sande, K. E., et al.: Selective search for object recognition. Int. J. Comput. Vis. 104(2), 154–171 (2013)

    Article  Google Scholar 

  11. Zitnick, C.L., Dollár, P.: Edge boxes: locating object proposals from edges. In: Proceedings of the 18th European Conference on Computer Vision, pp. 391–405, September 2014

    Google Scholar 

  12. Cheng, M.M., Zhang, Z., et al.: Bing: binarized normed gradients for objectness estimation at 300 fps. In: Proceedings of the 32nd IEEE Conference on Computer Vision and Pattern Recognition, pp. 3286–3293, June 2014

    Google Scholar 

  13. Ren, S., He, K., et al.: Faster R-CNN: towards real-time object detection with region proposal networks. In: Proceedings of the 33rd IEEE Transactions on Pattern Analysis and Machine Intelligence, pp. 1–8, June 2015

    Google Scholar 

  14. Li, J., Liang, X., et al.: Scale-aware fast R-CNN for pedestrian detection. Comput. Sci. 25–32 (2015)

    Google Scholar 

  15. Song, H.O., et al.: Sparselet models for efficient multiclass object detection. In: Fitzgibbon, A., Lazebnik, S., Perona, P., Sato, Y., Schmid, C. (eds.) ECCV 2012. LNCS, pp. 802–815. Springer, Heidelberg (2012). https://doi.org/10.1007/978-3-642-33709-3_57

    Chapter  Google Scholar 

  16. Girshick, R., Song, H.O., et al.: Discriminatively activated sparselets. In: Proceedings of the 30th International Conference on Machine Learning, pp. 196–204, June 2013

    Google Scholar 

  17. Dollr, P., Tu, Z., et al.: Integral channel features. In: Proceedings of the 20th British Machine Vision Conference, pp. 7–10, September 2009

    Google Scholar 

  18. Dollar, P., Appel, R., et al.: Fast feature pyramids for object detection. IEEE Trans. Pattern Anal. Mach. Intell. 36(8), 1532–1545, May 2014

    Article  Google Scholar 

  19. Aharon, M., Elad, M., et al.: K-SVD: an algorithm for designing overcomplete dictionaries for sparse representation. IEEE Trans. Signal Process. 54(11), 4311–4322, October 2006

    Article  Google Scholar 

  20. Dollar, P., Wojek, C., et al.: Pedestrian detection: an evaluation of the state of the art. IEEE Trans. Pattern Anal. Mach. Intell. 43(4), 743–761, March 2011

    Article  Google Scholar 

  21. Andreas, E., Bastian, L., et al.: Depth and appearance for mobile scene analysis. In: Proceedings of the 25th IEEE International Conference on Computer Vision, pp. 1–8, June 2007

    Google Scholar 

  22. Wojek, C., Walk, S., et al.: Multi-cue onboard pedestrian detection. In: Proceedings of the 27th IEEE Conference on Computer Vision and Pattern Recognition, pp. 794–801, June 2009

    Google Scholar 

  23. Benenson, R., Mathias, M., et al.: Pedestrian detection at 100 frames per second. In: Proceedings of 30th IEEE Conference on Computer Vision and Pattern Recognition, pp. 2903–2910, June 2012

    Google Scholar 

  24. Hosang, J., Benenson, R., Dollar, P., et al.: What makes for effec-tive detection proposals? IEEE Trans. Pattern Anal. Mach. Intell. 38(4), 814–830, March 2015

    Google Scholar 

  25. Zitnick, C.L., Dollr, P.: Edge boxes: locating object proposals from edges. In: Proceedings of the 18th European Conference on Computer Vision, pp. 391–405, September 2014

    Google Scholar 

  26. Cotter, S.F., Rao, B.D., et al.: Forward sequential algorithms for best basis selection. IEEE Vis. Image Signal Process. 146(5), 235 (1999)

    Article  Google Scholar 

  27. Mallat, S.G., Zhang, Z.: Matching pursuits with time-frequency dictionaries. IEEE Trans. Signal Process. 41(12), 3397–3415 (1993)

    Article  Google Scholar 

  28. Tibshirani, R.: Regression shrinkage and selection via the lasso. J. Roy. Stat. Soc. 58(1), 267–288 (1996)

    MathSciNet  MATH  Google Scholar 

  29. Zou, H., Hastie, T.: Regularization and variable selection via the elastic net. J. Roy. Stat. Soc. 67(2), 301–320 (2005)

    Article  MathSciNet  Google Scholar 

Download references

Aknowledgment

This research is based upon work supported by National Nature Science Founda- tion of China (No. U1736206), National Nature Science Foundation of China (61671336), National Nature Science Foundation of China (61671332), Technology Research 10 F. Author et al. Program of Ministry of Public Security (No. 2016JSYJA12), Hubei Province Technological Innovation Major Project (No. 2016AAA015), Hubei Province Tech- nological Innovation Major Project (2017AAA123), The National Key Research and Development Program of China (No.2016YFB0100901), Nature Science Foun- dation of Jiangsu Province (No. BK20160386) and National Nature Science Foundation of China (61502354).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Wenhua Fang .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2018 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Fang, W., Chen, J., Lu, T., Hu, R. (2018). Feature Synthesization for Real-Time Pedestrian Detection in Urban Environment. In: Hong, R., Cheng, WH., Yamasaki, T., Wang, M., Ngo, CW. (eds) Advances in Multimedia Information Processing – PCM 2018. PCM 2018. Lecture Notes in Computer Science(), vol 11165. Springer, Cham. https://doi.org/10.1007/978-3-030-00767-6_10

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-00767-6_10

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-00766-9

  • Online ISBN: 978-3-030-00767-6

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics