Skip to main content
Log in

Towards affordance detection for robot manipulation using affordance for parts and parts for affordance

  • Published:
Autonomous Robots Aims and scope Submit manuscript

Abstract

As robots start to interact with their environments, they need to reason about the affordances of objects in those environments. In most cases, affordances can be inferred only from parts of objects, such as the blade of a knife for cutting or the head of a hammer for pounding. We propose an RGB-D part-based affordance detection method where the parts are obtained based on the affordances as well. We show that affordance detection benefits from a part-based object representation since parts are distinctive and generalizable to novel objects. We compare our method with other state-of-the-art affordance detection methods on a benchmark dataset (Myers et al. in International conference on robotics and automation (ICRA), 2015), outperforming these methods by an average of 14% on novel object instances. Furthermore, we apply our affordance detection method to a robotic grasping scenario to demonstrate that the robot is able to perform grasps after detecting the affordances.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Institutional subscriptions

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13
Fig. 14
Fig. 15

Similar content being viewed by others

Notes

  1. http://pointclouds.org/.

  2. http://pointclouds.org/.

  3. Please see Sect. 6 for a complete list of object instances and their corresponding splits.

  4. The reader may refer to Sect. 6 for a complete list of objects and their corresponding category splits.

References

  • Aldoma, A., Tombari, F., & Vincze, M. (2012). Supervised learning of hidden and non-hidden 0-order affordances and detection in real scenes. In: 2012 IEEE International Conference on Robotics and Automation (ICRA), IEEE, pp 1732–1739

  • Bo, L., Ren, X., & Fox, D. (2013). Unsupervised feature learning for RGB-D based object recognition. In J. P. Desai, G. Dudek, O. Khatib, & V. Kumar (Eds.), Experimental robotics (pp. 387–402). Springer.

  • Desai, C & Ramanan, D. (2013). Predicting functional regions on objects. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops.

  • Fei-Fei, L & Perona, P. (2005). A bayesian hierarchical model for learning natural scene categories. In IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2005. CVPR 2005. IEEE, vol 2, pp. 524–531.

  • Felzenszwalb, P. F., Girshick, R. B., McAllester, D., & Ramanan, D. (2010). Object detection with discriminatively trained part-based models. IEEE Transactions on Pattern Analysis and Machine Intelligence (PAMI), 32(9), 1627–1645.

    Article  Google Scholar 

  • Fidler, S & Leonardis, A. (2007). Towards scalable representations of object categories: Learning a hierarchy of parts. In: IEEE Conference on Computer Vision and Pattern Recognition, IEEE, pp. 1–8.

  • Fu, H., Cohen-Or, D., Dror, G., & Sheffer, A. (2008). Upright orientation of man-made objects. In ACM transactions on graphics (TOG) (Vol. 27, p. 42). ACM.

  • Gibson, J. J. (1977). The theory of affordances. Perceiving, Acting, and Knowing: Toward an Ecological Psychology, 67–82.

  • Gibson, J. J. (1979). The ecological approach to visual perception. Hove: Psychology Press.

    Google Scholar 

  • Hart, S., Dinh, P., & Hambuchen, K. (2014). Affordance templates for shared robot control. In: Artificial Intelligence and Human-Robot Interaction, AAAI Fall Symposium Series, Arlington, VA, USA.

  • Hart, S., Dinh, P., & Hambuchen, K. (2015). The affordance template ros package for robot task programming. In: IEEE International Conference on Robotics and Automation (ICRA), 2015, IEEE, pp. 6227–6234.

  • Hermans, T., Rehg, J. M. & Bobick, A. (2011). Affordance prediction via learned object attributes. In: IEEE International Conference on Robotics and Automation (ICRA): Workshop on Semantic Perception, Mapping, and Exploration, pp. 181–184.

  • Katz, D., Venkatraman, A., Kazemi, M., Bagnell, J. A., & Stentz, A. (2014). Perceiving, learning, and exploiting object affordances for autonomous pile manipulation. Autonomous Robots, 37(4), 369–382.

    Article  Google Scholar 

  • Koppula, H. S. & Saxena, A. (2014). Physically grounded spatio-temporal object affordances. In: European Conference on Computer Vision, Springer, pp. 831–847.

  • Laga, H., Mortara, M., & Spagnuolo, M. (2013). Geometry and context for semantic correspondences and functionality recognition in man-made 3D shapes. ACM Transactions on Graphics (TOG), 32(5), 150.

    Article  Google Scholar 

  • Lazebnik, S., Schmid, C & Ponce, J. (2006). Beyond bags of features: Spatial pyramid matching for recognizing natural scene categories. In: IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2006, IEEE, vol. 2, pp. 2169–2178.

  • Leung, T., & Malik, J. (2001). Representing and recognizing the visual appearance of materials using three-dimensional textons. International Journal of Computer Vision, 43(1), 29–44.

    Article  MATH  Google Scholar 

  • Margolin, R., Zelnik-Manor, L., Tal, A. (2014). How to evaluate foreground maps? In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 248–255.

  • Myers, A., Teo, CL., Fermüller, C., & Aloimonos, Y. (2015). Affordance detection of tool parts from geometric features. In: International Conference on Robotics and Automation (ICRA).

  • Nguyen, A., Kanoulas, D., Caldwell, D. G., & Tsagarakis, N. G. (2016). Detecting object affordances with convolutional neural networks. In: IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), IEEE, pp. 2765–2770.

  • Norman, D. A. (1988). The psychology of everyday things. New York: Basic Books.

    Google Scholar 

  • Omrčen, D., Böge, C., Asfour, T., Ude, A., & Dillmann, R. (2009). Autonomous acquisition of pushing actions to support object grasping with a humanoid robot. In: 9th IEEE-RAS International Conference on Humanoid Robots, IEEE, pp. 277–283.

  • Platt, J., et al. (1999). Probabilistic outputs for support vector machines and comparisons to regularized likelihood methods. Advances in large margin classifiers, 10(3), 61–74.

    Google Scholar 

  • Rabbani, T., Van Den Heuvel, F., & Vosselmann, G. (2006). Segmentation of point clouds using smoothness constraint. International Archives of Photogrammetry, Remote Sensing and Spatial Information Sciences, 36(5), 248–253.

    Google Scholar 

  • Rezapour Lakani, S., Rodríguez-Sánchez, A., & Piater, J. (2017). Can affordances guide object decomposition into semantically meaningful parts? In: IEEE Winter Conference on Applications of Computer Vision (WACV).

  • Richtsfeld, A., Mörwald, T., Prankl, J., Zillich, M., & Vincze, M. (2014). Learning of perceptual grouping for object segmentation on rgb-d data. Journal of visual communication and image representation, 25(1), 64–73.

    Article  Google Scholar 

  • Rivlin, E., Dickinson, S. J., & Rosenfeld, A. (1995). Recognition by functional parts. Computer Vision and Image Understanding, 62(2), 164–176.

    Article  MATH  Google Scholar 

  • Rumelhart, D. E., Hinton, G. E., & Williams, R. J. (1985). Learning internal representations by error propagation. DTIC Document: Tech. rep.

  • Rusu, R. B., & Cousins, S. (2011). 3D is here: Point cloud library (PCL). In: IEEE International Conference on Robotics and Automation (ICRA), IEEE, pp. 1–4.

  • Sawatzky, J., Srikantha, A & Gall, J. (2017). Weakly supervised affordance detection. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

  • Schmidt, M. (2007). UGM: A matlab toolbox for probabilistic undirected graphical models. http://www.cs.ubc.ca/~schmidtm/Software/UGM.html.

  • Stark, L., & Bowyer, K. (1991). Achieving generalized object recognition through reasoning about association of function to structure. IEEE Transactions on Pattern Analysis and Machine Intelligence, 13(10), 1097–1104.

    Article  Google Scholar 

  • Stark, M., Lies, P., Zillich, M., Wyatt, J., & Schiele, B. (2008). Functional object class detection based on learned affordance cues. Computer Vision Systems, 5008, 435–444.

    Article  Google Scholar 

  • Stein, C. S., Schoeler, M., Papon, J., & Wörgötter, F. (2014). Object partitioning using local convexity. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

  • Varadarajan, K. M., & Vincze, M. (2011). Affordance based part recognition for grasping and manipulation. In: Workshop on Autonomous Grasping, ICRA.

  • Wang, J., & Yuille, A. L. (2015). Semantic part segmentation using compositional model combining shape and appearance. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1788–1797.

  • Yao, B., Ma, J., & Fei-Fei, L. (2013). Discovering object functionality. In: The IEEE International Conference on Computer Vision (ICCV).

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Safoura Rezapour Lakani.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Electronic supplementary material

Below is the link to the electronic supplementary material.

Supplementary material 1 (mp4 32348 KB)

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Rezapour Lakani, S., Rodríguez-Sánchez, A.J. & Piater, J. Towards affordance detection for robot manipulation using affordance for parts and parts for affordance. Auton Robot 43, 1155–1172 (2019). https://doi.org/10.1007/s10514-018-9787-5

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10514-018-9787-5

Keywords

Navigation