Abstract
Purpose This paper aims to present an object detection methodology to categorize 3D object models in an efficient manner. The authors propose a dynamically generated hierarchical architecture to compute very fast objects’ 3D pose for mobile service robots to grasp them. Design/methodology/approach The methodology used in this study is based on a dynamic pyramid search and fast template representation, metadata and context-free grammars. In the experiments, the authors use an omnidirectional KUKA mobile manipulator equipped with an RGBD camera, to localize objects requested by humans. The proposed architecture is based on efficient object detection and visual servoing. In the experiments, the robot successfully finds 3D poses. The present proposal is not restricted to specific robots or objects and can grow as much as needed. Findings The authors present the dynamic categorization using context-free grammars and 3D object detection, and through several experiments, the authors perform a proof of concept. The authors obtained promising results, showing that their methods can scale to more complex scenes and they can be used in future applications in real-world scenarios where mobile robot are needed in areas such as service robots or industry in general. Research limitations/implications The experiments were carried out using a mobile KUKA youBot. Scalability and more robust algorithms will improve the present proposal. In the first stage, the authors carried out an experimental validation. Practical implications The current proposal describes a scalable architecture, where more agents can be added or reprogrammed to handle more complicated tasks. Originality/value The main contribution of this study resides in the dynamic categorization scheme for fast detection of 3D objects, and the issues and experiments carried out to test the viability of the methods. Usually, state-of-the-art treats categories as rigid and make static queries to datasets. In the present approach, there are no fixed categories and they are created and combined on the fly to speed up detection.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
More From: Industrial Robot: the international journal of robotics research and application
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.