One challenge of motion generation using robot learning from demonstration techniques is that human demonstrations follow a distribution with multiple modes for one task query. Previous approaches fail to capture all modes or tend to average modes of the demonstrations and thus generate invalid trajectories. The other difficulty is the small number of demonstrations that cannot cover the entire working space. To overcome this problem, a motion generation model with extrapolation ability is needed. Previous works restrict task queries as local frames and learn representations in local frames. We propose a model to solve both problems. For multiple modes, we suggest to learn local latent representations of motion trajectories with a density estimation method based on real-valued non-volume preserving (RealNVP) transformations that provides a set of powerful, stably invertible, and learnable transformations. To improve the extrapolation ability, we propose to shift the attention of the robot from one local frame to another during the task execution. In experiments, we consider the docking problem used also in previous works where a trajectory has to be generated to connect two dockers without collision. We increase complexity of the task and show that the proposed method outperforms other approaches. In addition, we evaluate the approach in real robot experiments.
翻译:使用机器人从演示技术中学习的机器人产生运动能力的一个挑战是,人类演示遵循多种模式的分布方式,用于一个任务查询。以前的方法未能捕捉所有模式,或倾向于演示的平均模式,从而产生无效的轨迹。另一个困难是无法覆盖整个工作空间的少量演示。为了克服这一问题,需要一个具有外推能力的动作生成模型。以前的工作限制任务查询,将其作为本地框架,并在当地框架中学习演示。我们提出了一个解决这两个问题的模型。对于多种模式,我们建议学习运动轨迹的局部潜在潜在表现,以基于实际估价的非数量保护(RealNVP)变形的密度估计方法,提供一套强力、可刺穿、不可翻转和可学习的变形。为了提高外推法能力,我们提议在任务执行期间将机器人的注意力从一个本地框架转移到另一个框架。在试验中,我们考虑到在先前的工程中使用的对接问题,即必须产生一条轨迹将两个码头连接在一起而没有碰撞。我们增加了任务的复杂性,并显示拟议的方法超越了其他方法。此外,我们评估了实际的实验方法。