Literature DB >> 29872389

Multimodal Hierarchical Dirichlet Process-Based Active Perception by a Robot.

Tadahiro Taniguchi1, Ryo Yoshino1, Toshiaki Takano2.   

Abstract

In this paper, we propose an active perception method for recognizing object categories based on the multimodal hierarchical Dirichlet process (MHDP). The MHDP enables a robot to form object categories using multimodal information, e.g., visual, auditory, and haptic information, which can be observed by performing actions on an object. However, performing many actions on a target object requires a long time. In a real-time scenario, i.e., when the time is limited, the robot has to determine the set of actions that is most effective for recognizing a target object. We propose an active perception for MHDP method that uses the information gain (IG) maximization criterion and lazy greedy algorithm. We show that the IG maximization criterion is optimal in the sense that the criterion is equivalent to a minimization of the expected Kullback-Leibler divergence between a final recognition state and the recognition state after the next set of actions. However, a straightforward calculation of IG is practically impossible. Therefore, we derive a Monte Carlo approximation method for IG by making use of a property of the MHDP. We also show that the IG has submodular and non-decreasing properties as a set function because of the structure of the graphical model of the MHDP. Therefore, the IG maximization problem is reduced to a submodular maximization problem. This means that greedy and lazy greedy algorithms are effective and have a theoretical justification for their performance. We conducted an experiment using an upper-torso humanoid robot and a second one using synthetic data. The experimental results show that the method enables the robot to select a set of actions that allow it to recognize target objects quickly and accurately. The numerical experiment using the synthetic data shows that the proposed method can work appropriately even when the number of actions is large and a set of target objects involves objects categorized into multiple classes. The results support our theoretical outcomes.

Entities:  

Keywords:  active perception; cognitive robotics; multimodal machine learning; submodular maximization; topic model

Year:  2018        PMID: 29872389      PMCID: PMC5972223          DOI: 10.3389/fnbot.2018.00022

Source DB:  PubMed          Journal:  Front Neurorobot        ISSN: 1662-5218            Impact factor:   2.650


  4 in total

1.  A Bayesian framework for active artificial perception.

Authors:  João Filipe Ferreira; Jorge Lobo; Pierre Bessière; Miguel Castelo-Branco; Jorge Dias
Journal:  IEEE Trans Cybern       Date:  2013-03-07       Impact factor: 11.448

2.  Category and feature identification.

Authors:  Charles Kemp; Kai-min K Chang; Luigi Lombardi
Journal:  Acta Psychol (Amst)       Date:  2010-01-18

3.  Bayesian exploration for intelligent identification of textures.

Authors:  Jeremy A Fishel; Gerald E Loeb
Journal:  Front Neurorobot       Date:  2012-06-18       Impact factor: 2.650

4.  Learning tactile skills through curious exploration.

Authors:  Leo Pape; Calogero M Oddo; Marco Controzzi; Christian Cipriani; Alexander Förster; Maria C Carrozza; Jürgen Schmidhuber
Journal:  Front Neurorobot       Date:  2012-07-23       Impact factor: 2.650

  4 in total
  2 in total

1.  Active Inference Through Energy Minimization in Multimodal Affective Human-Robot Interaction.

Authors:  Takato Horii; Yukie Nagai
Journal:  Front Robot AI       Date:  2021-11-26

2.  A Framework for Sensorimotor Cross-Perception and Cross-Behavior Knowledge Transfer for Object Categorization.

Authors:  Gyan Tatiya; Ramtin Hosseini; Michael C Hughes; Jivko Sinapov
Journal:  Front Robot AI       Date:  2020-10-09
  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.