Advertisement

Evaluating CNNs on the Gestalt Principle of Closure

  • Gregor EhrenspergerEmail author
  • Sebastian Stabinger
  • Antonio Rodríguez Sánchez
Conference paper
Part of the Lecture Notes in Computer Science book series (LNCS, volume 11727)

Abstract

Deep convolutional neural networks (CNNs) are widely known for their outstanding performance in classification and regression tasks over high-dimensional data. This made them a popular and powerful tool for a large variety of applications in industry and academia. Recent publications show that seemingly easy classification tasks (for humans) can be very challenging for state of the art CNNs. An attempt to describe how humans perceive visual elements is given by the Gestalt principles. In this paper we evaluate AlexNet and GoogLeNet regarding their performance on classifying the correctness of the well known Kanizsa triangles and triangles where sections of the edges were removed. Both types heavily rely on the Gestalt principle of closure. Therefore we created various datasets containing valid as well as invalid variants of the described triangles. Our findings suggest that perceiving objects by utilizing the principle of closure is very challenging for the applied network architectures but they appear to adapt to the effect of closure.

Keywords

Convolutional neural network CNN Gestalt principles Principle of closure 

Notes

Acknowledgment

We want to thank the anonymous reviewers for their constructive suggestions and helpful comments.

References

  1. 1.
    Kim, B., Reif, E., Wattenberg, M., Bengio, S.: Do neural networks show gestalt phenomena? An exploration of the law of closure. arXiv preprint (2019)Google Scholar
  2. 2.
    Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. In: Advances in Neural Information Processing Systems (2012)Google Scholar
  3. 3.
    LeCun, Y., et al.: Backpropagation applied to handwritten zip code recognition. Neural Comput. 1(4), 541–551 (1989).  https://doi.org/10.1162/neco.1989.1.4.541CrossRefGoogle Scholar
  4. 4.
    LeCun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition. Proc. IEEE 86(11), 2278–2324 (1998).  https://doi.org/10.1109/5.726791CrossRefGoogle Scholar
  5. 5.
    Stabinger, S., Rodríguez-Sánchez, A., Piater, J.: 25 years of CNNs: can we compare to human abstraction capabilities? In: Villa, A.E.P., Masulli, P., Pons Rivero, A.J. (eds.) ICANN 2016. LNCS, vol. 9887, pp. 380–387. Springer, Cham (2016).  https://doi.org/10.1007/978-3-319-44781-0_45CrossRefGoogle Scholar
  6. 6.
    Stabinger, S., Rodríguez-Sánchez, A.: Evaluation of deep learning on an abstract image classification dataset. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2767–2772 (2017).  https://doi.org/10.1109/iccvw.2017.325
  7. 7.
    Szegedy, C., et al.: Going deeper with convolutions. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1–9 (2015).  https://doi.org/10.1109/cvpr.2015.7298594
  8. 8.
    Wertheimer, M.: Laws of organization in perceptual forms. In: A Source Book of Gestalt Psychology, pp. 71–88, London (1938).  https://doi.org/10.1037/11496-005

Copyright information

© Springer Nature Switzerland AG 2019

Authors and Affiliations

  1. 1.University of InnsbruckInnsbruckAustria

Personalised recommendations