Modeling and Learning of Object Placing Tasks from Human Demonstrations in Smart Manufacturing

2019-01-0700

04/02/2019

Event
WCX SAE World Congress Experience
Authors Abstract
Content
In this paper, we present a framework for the robot to learn how to place objects to a workpiece by learning from humans in smart manufacturing. In the proposed framework, the rational scene dictionary (RSD) corresponding to the keyframes of task (KFT) are used to identify the general object-action-location relationships. The Generalized Voronoi Diagrams (GVD) based contour is used to determine the relative position and orientation between the object and the corresponding workpiece at the final state. In the learning phase, we keep tracking the image segments in the human demonstration. For the moment when a spatial relation of some segments are changed in a discontinuous way, the state changes are recorded by the RSD. KFT is abstracted after traversing and searching in RSD, while the relative position and orientation of the object and the corresponding mount are presented by GVD-based contours for the keyframes. When the object or the relative position and orientation between the object and the workpiece are changed, the GVD, as well as the shape of contours extracted from the GVD, are also different. The Fourier Descriptor (FD) is applied to describe these differences on the shape of contours in the GVD. The proposed framework is validated through experimental results.
Meta TagsDetails
DOI
https://doi.org/10.4271/2019-01-0700
Pages
8
Citation
Chen, Y., Wang, W., Zhang, Z., Krovi, V. et al., "Modeling and Learning of Object Placing Tasks from Human Demonstrations in Smart Manufacturing," SAE Technical Paper 2019-01-0700, 2019, https://doi.org/10.4271/2019-01-0700.
Additional Details
Publisher
Published
Apr 2, 2019
Product Code
2019-01-0700
Content Type
Technical Paper
Language
English