Publication: 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)
Abstract: We present a novel and real-time method to detect object affordances from RGB-D images. Our method trains a deep Convolutional Neural Network (CNN) to learn deep features from the input data in an end-to-end manner. The CNN has an encoder-decoder architecture in order to obtain smooth label predictions. The input data are represented as multiple modalities to let the network learn the features more effectively. Our method sets a new benchmark on detecting object affordances, improving the accuracy by 20% in comparison with the state-of-the-art methods that use hand-designed geometric features. Furthermore, we apply our detection method on a full-size humanoid robot (WALK-MAN) to demonstrate that the robot is able to perform grasps after efficiently detecting the object affordances.
Bibtex:
@inproceedings{nguyen2016detecting, author = {Nguyen, Anh and Kanoulas, Dimitrios and Caldwell, Darwin G. and Tsagarakis, Nikos G.}, booktitle = {2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)}, title = {Detecting object affordances with Convolutional Neural Networks}, year = {2016}, volume = {}, number = {}, pages = {2765-2770}, keywords = {}, doi = {10.1109/IROS.2016.7759429}, issn = {2153-0866}, month = oct }