Abstract
As robots start to interact with their environments, they need to reason about the affordances of objects in those environments. In most cases, affordances can be inferred only from parts of objects, such as the blade of a knife for cutting or the head of a hammer for pounding. We propose an RGB-D part-based affordance detection method where the parts are obtained based on the affordances as well. We show that affordance detection benefits from a part-based object representation since parts are distinctive and generalizable to novel objects. We compare our method with other state-of-the-art affordance detection methods on a benchmark dataset (Myers et al. in International conference on robotics and automation (ICRA), 2015), outperforming these methods by an average of 14% on novel object instances. Furthermore, we apply our affordance detection method to a robotic grasping scenario to demonstrate that the robot is able to perform grasps after detecting the affordances.
Similar content being viewed by others
Notes
References
Aldoma, A., Tombari, F., & Vincze, M. (2012). Supervised learning of hidden and non-hidden 0-order affordances and detection in real scenes. In: 2012 IEEE International Conference on Robotics and Automation (ICRA), IEEE, pp 1732–1739
Bo, L., Ren, X., & Fox, D. (2013). Unsupervised feature learning for RGB-D based object recognition. In J. P. Desai, G. Dudek, O. Khatib, & V. Kumar (Eds.), Experimental robotics (pp. 387–402). Springer.
Desai, C & Ramanan, D. (2013). Predicting functional regions on objects. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops.
Fei-Fei, L & Perona, P. (2005). A bayesian hierarchical model for learning natural scene categories. In IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2005. CVPR 2005. IEEE, vol 2, pp. 524–531.
Felzenszwalb, P. F., Girshick, R. B., McAllester, D., & Ramanan, D. (2010). Object detection with discriminatively trained part-based models. IEEE Transactions on Pattern Analysis and Machine Intelligence (PAMI), 32(9), 1627–1645.
Fidler, S & Leonardis, A. (2007). Towards scalable representations of object categories: Learning a hierarchy of parts. In: IEEE Conference on Computer Vision and Pattern Recognition, IEEE, pp. 1–8.
Fu, H., Cohen-Or, D., Dror, G., & Sheffer, A. (2008). Upright orientation of man-made objects. In ACM transactions on graphics (TOG) (Vol. 27, p. 42). ACM.
Gibson, J. J. (1977). The theory of affordances. Perceiving, Acting, and Knowing: Toward an Ecological Psychology, 67–82.
Gibson, J. J. (1979). The ecological approach to visual perception. Hove: Psychology Press.
Hart, S., Dinh, P., & Hambuchen, K. (2014). Affordance templates for shared robot control. In: Artificial Intelligence and Human-Robot Interaction, AAAI Fall Symposium Series, Arlington, VA, USA.
Hart, S., Dinh, P., & Hambuchen, K. (2015). The affordance template ros package for robot task programming. In: IEEE International Conference on Robotics and Automation (ICRA), 2015, IEEE, pp. 6227–6234.
Hermans, T., Rehg, J. M. & Bobick, A. (2011). Affordance prediction via learned object attributes. In: IEEE International Conference on Robotics and Automation (ICRA): Workshop on Semantic Perception, Mapping, and Exploration, pp. 181–184.
Katz, D., Venkatraman, A., Kazemi, M., Bagnell, J. A., & Stentz, A. (2014). Perceiving, learning, and exploiting object affordances for autonomous pile manipulation. Autonomous Robots, 37(4), 369–382.
Koppula, H. S. & Saxena, A. (2014). Physically grounded spatio-temporal object affordances. In: European Conference on Computer Vision, Springer, pp. 831–847.
Laga, H., Mortara, M., & Spagnuolo, M. (2013). Geometry and context for semantic correspondences and functionality recognition in man-made 3D shapes. ACM Transactions on Graphics (TOG), 32(5), 150.
Lazebnik, S., Schmid, C & Ponce, J. (2006). Beyond bags of features: Spatial pyramid matching for recognizing natural scene categories. In: IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2006, IEEE, vol. 2, pp. 2169–2178.
Leung, T., & Malik, J. (2001). Representing and recognizing the visual appearance of materials using three-dimensional textons. International Journal of Computer Vision, 43(1), 29–44.
Margolin, R., Zelnik-Manor, L., Tal, A. (2014). How to evaluate foreground maps? In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 248–255.
Myers, A., Teo, CL., Fermüller, C., & Aloimonos, Y. (2015). Affordance detection of tool parts from geometric features. In: International Conference on Robotics and Automation (ICRA).
Nguyen, A., Kanoulas, D., Caldwell, D. G., & Tsagarakis, N. G. (2016). Detecting object affordances with convolutional neural networks. In: IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), IEEE, pp. 2765–2770.
Norman, D. A. (1988). The psychology of everyday things. New York: Basic Books.
Omrčen, D., Böge, C., Asfour, T., Ude, A., & Dillmann, R. (2009). Autonomous acquisition of pushing actions to support object grasping with a humanoid robot. In: 9th IEEE-RAS International Conference on Humanoid Robots, IEEE, pp. 277–283.
Platt, J., et al. (1999). Probabilistic outputs for support vector machines and comparisons to regularized likelihood methods. Advances in large margin classifiers, 10(3), 61–74.
Rabbani, T., Van Den Heuvel, F., & Vosselmann, G. (2006). Segmentation of point clouds using smoothness constraint. International Archives of Photogrammetry, Remote Sensing and Spatial Information Sciences, 36(5), 248–253.
Rezapour Lakani, S., Rodríguez-Sánchez, A., & Piater, J. (2017). Can affordances guide object decomposition into semantically meaningful parts? In: IEEE Winter Conference on Applications of Computer Vision (WACV).
Richtsfeld, A., Mörwald, T., Prankl, J., Zillich, M., & Vincze, M. (2014). Learning of perceptual grouping for object segmentation on rgb-d data. Journal of visual communication and image representation, 25(1), 64–73.
Rivlin, E., Dickinson, S. J., & Rosenfeld, A. (1995). Recognition by functional parts. Computer Vision and Image Understanding, 62(2), 164–176.
Rumelhart, D. E., Hinton, G. E., & Williams, R. J. (1985). Learning internal representations by error propagation. DTIC Document: Tech. rep.
Rusu, R. B., & Cousins, S. (2011). 3D is here: Point cloud library (PCL). In: IEEE International Conference on Robotics and Automation (ICRA), IEEE, pp. 1–4.
Sawatzky, J., Srikantha, A & Gall, J. (2017). Weakly supervised affordance detection. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
Schmidt, M. (2007). UGM: A matlab toolbox for probabilistic undirected graphical models. http://www.cs.ubc.ca/~schmidtm/Software/UGM.html.
Stark, L., & Bowyer, K. (1991). Achieving generalized object recognition through reasoning about association of function to structure. IEEE Transactions on Pattern Analysis and Machine Intelligence, 13(10), 1097–1104.
Stark, M., Lies, P., Zillich, M., Wyatt, J., & Schiele, B. (2008). Functional object class detection based on learned affordance cues. Computer Vision Systems, 5008, 435–444.
Stein, C. S., Schoeler, M., Papon, J., & Wörgötter, F. (2014). Object partitioning using local convexity. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
Varadarajan, K. M., & Vincze, M. (2011). Affordance based part recognition for grasping and manipulation. In: Workshop on Autonomous Grasping, ICRA.
Wang, J., & Yuille, A. L. (2015). Semantic part segmentation using compositional model combining shape and appearance. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1788–1797.
Yao, B., Ma, J., & Fei-Fei, L. (2013). Discovering object functionality. In: The IEEE International Conference on Computer Vision (ICCV).
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Electronic supplementary material
Below is the link to the electronic supplementary material.
Rights and permissions
About this article
Cite this article
Rezapour Lakani, S., Rodríguez-Sánchez, A.J. & Piater, J. Towards affordance detection for robot manipulation using affordance for parts and parts for affordance. Auton Robot 43, 1155–1172 (2019). https://doi.org/10.1007/s10514-018-9787-5
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10514-018-9787-5