Modeling Paired Objects and Their Interaction

Part of the Cognitive Systems Monographs book series (COSMOS, volume 23)

Abstract

Object categorization and human action recognition are two important capabilities for an intelligent robot. Traditionally, they are treated separately. Recently, more researchers started to model the object features, object affordance, and human action at the same time. Most of the works build a relation model between single object features and human action or object affordance and uses the models to improve object recognition accuracies [16, 21, 12].

Keywords

Bayesian Network Mirror Neuron Mirror Neuron System Bayesian Network Model Paired Object 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. 1.
    Aksoy, E., Abramov, A., Worgotter, F., Dellen, B.: Categorizing object-action relations from semantic scene graphs. In: IEEE Intl. Conference on Robotics and Automation, pp. 398–405 (2010)Google Scholar
  2. 2.
    Argyros, A.A., Lourakis, M.I.A.: Real-time tracking of multiple skin-colored objects with a possibly moving camera. In: Pajdla, T., Matas, J(G.) (eds.) ECCV 2004. LNCS, vol. 3023, pp. 368–379. Springer, Heidelberg (2004)CrossRefGoogle Scholar
  3. 3.
    Borghi, A., Flumini, A., Natraj, N., Wheaton, L.: One hand, two objects: emergence of affordance in contexts. Brain and Cognition 80(1), 64–73 (2012)CrossRefGoogle Scholar
  4. 4.
    Chang, C., Lin, C.: Libsvm: A library for support vector machines. ACM Transactions on Intelligent Systems and Technology 2(3), 1–27 (2011)CrossRefGoogle Scholar
  5. 5.
    Conaire, C., O’Connor, N.E., Smeaton, A.F.: Detector adaptation by maximising agreement between independent data sources. In: IEEE International Workshop on Object Tracking and Classification Beyond the Visible Spectrum, pp. 1–6 (2007)Google Scholar
  6. 6.
    Dai, W., Sun, Y., Qian, X.: Functional analysis of grasping motion. In: IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 1–7 (2013)Google Scholar
  7. 7.
    Dalal, N., Triggs, B.: Histograms of oriented gradients for human detection. In: Conference on Computer Vision and Pattern Recognition, pp. 886–893 (2005)Google Scholar
  8. 8.
    Deng, J., Dong, W., Socher, R., Li, L., Li, K., Fei-Fei, L.: Imagenet: A large-scale hierarchical image database. In: Int. Conf. on Computer Vision and Pattern Recognition, pp. 248–255 (2009)Google Scholar
  9. 9.
    Di Pellegrino, G., Fadiga, L., Fogassi, L., Gallese, V., Rizzolatti, G.: Understanding motor events: A neurophysiological study. Exp. Brain Res. 91, 176–180 (1992)CrossRefGoogle Scholar
  10. 10.
    Felzenszwalb, P.F., McAllester, D., Ramanan, D.: A discriminatively trained, multiscale, deformable part model. In: Conference on Computer Vision and Pattern Recognition, pp. 1–8 (2008)Google Scholar
  11. 11.
    Fogassi, L., et al.: Parietal lobe: From action organization to intention understanding. Science 29(308), 662–667 (2005)CrossRefGoogle Scholar
  12. 12.
    Gall, J., Fossati, A., Gool, L.: Functional categorization of objects using real-time markerless motion capture. In: Conference on Computer Vision and Pattern Recognition, pp. 1969–1976 (2011)Google Scholar
  13. 13.
    Gallese, V., Fogassi, L., Fadiga, L., Rizzolatti, G.: Action representation and the inferior parietal lobule. In: Prinz, W., Hommel, B. (eds.) Attention and Performance XIX. Common mechanisms in perception and action. Oxford University Press, Oxford (2002)Google Scholar
  14. 14.
    Gibson, J.: The theory of affordances. In: Shaw, R., Bransford, J. (eds.) Perceiving, Acting and Knowing. Erlbaum, Hillsdale (1977)Google Scholar
  15. 15.
    Grabner, H., Gall, J., Van Gool, L.: What makes a chair a chair? In: Conference on Computer Vision and Pattern Recognition, pp. 1529–1536 (2011)Google Scholar
  16. 16.
    Gupta, A., Davis, L.: Objects in action: An approach for combining action understanding and object perception. In: Conference on Computer Vision and Pattern Recognition, pp. 1–8 (2007)Google Scholar
  17. 17.
    Hamilton, A., Grafton, S.: The motor hierarchy: from kinematics to goals and intentions. In: Haggard, P., Rosetti, Y., Kawato, M. (eds.) Attention and Performance, ch. 22 (2007)Google Scholar
  18. 18.
    Iacoboni, M., Molnar-Szakacs, I., Gallese, V., Buccino, G., Mazziotta, J., et al.: Grasping the intentions of others with one’s own mirror neuron system. PLoS Biol. 3(3) (2005)Google Scholar
  19. 19.
    Jax, S.A., Buxbaum, L.J.: Response interference between functional and structural actions linked to the same familiar object. Cognition 115(2), 350–355 (2010)CrossRefGoogle Scholar
  20. 20.
    Kala, Z., Matas, J., Mikolajczyk, K.: P-n learning: Bootstrapping binary classifiers by structural constraints. In: Conference on Computer Vision and Pattern Recognition, pp. 49–56 (2010)Google Scholar
  21. 21.
    Kjellstrom, H., Romero, J., Kragic, D.: Visual object-action recognition: Inferring object affordances from human demonstration. Computer Vision and Image Understanding 115, 81–90 (2010)CrossRefGoogle Scholar
  22. 22.
    Lin, Y., Shaogang, R., Clevenger, M., Sun, Y.: Learning grasping force from demonstration. In: IEEE Intl. Conference on Robotics and Automation, pp. 1526–1531 (2012)Google Scholar
  23. 23.
    Lin, Y., Sun, Y.: Task-oriented grasp planning based on disturbance distribution. In: International Symposium on Robotics Research (ISRR), pp. 1–16 (2013)Google Scholar
  24. 24.
    Oztop, E., Kawato, M., Arbib, M.: Mirror neurons and imitation: a computationally guided review. Epub Neural Networks 19, 254–271 (2006)CrossRefMATHGoogle Scholar
  25. 25.
    Pearl, J.: Probabilistic Reasoning in Intelligent Systems: Network and Plausible Inference. Morgan Kaufmann (1988)Google Scholar
  26. 26.
    Prasad, V.S.N., Kellokumpu, V., Davis, L.S.: Ballistic hand movements. In: Perales, F.J., Fisher, R.B. (eds.) AMDO 2006. LNCS, vol. 4069, pp. 153–164. Springer, Heidelberg (2006)CrossRefGoogle Scholar
  27. 27.
    Ren, S., Sun, Y.: Human-object-object-interaction affordance. In: IEEE Workshop in Robot Vision, pp. 1–6 (2013)Google Scholar
  28. 28.
    Rizzolatti, G., Craighero, L.: The mirror neuron system. Ann. Rev. Neurosci. 27, 169–192 (2004)CrossRefGoogle Scholar
  29. 29.
    Rizzolatti, G., Craighero, L.: Mirror neuron: A neurological approach to empathy. In: Changeux, J.P., Damasio, A.R., Singer, W., Christen, Y. (eds.) Neurobiology of Human Values, Springer, Heidelberg (2005)Google Scholar
  30. 30.
    Stark, M., Lies, P., Zillich, M., Wyatt, J.C., Schiele, B.: Functional object class detection based on learned affordance cues. In: Gasteratos, A., Vincze, M., Tsotsos, J.K. (eds.) ICVS 2008. LNCS, vol. 5008, pp. 435–444. Springer, Heidelberg (2008)CrossRefGoogle Scholar
  31. 31.
    Sun, Y., Ren, S., Lin, Y.: Object-object interaction affordance learning. Robotics and Autonomous Systems (in Press)Google Scholar
  32. 32.
    Yoon, E., Humphreys, W., Riddoch, M.: The paired-object affordance effect. J. Exp. Psychol. Human 36, 812–824 (2010)CrossRefGoogle Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2015

Authors and Affiliations

  1. 1.Computer Science and EngineeringUniversity of South FloridaTampaU.S.A

Personalised recommendations