| Literature DB >> 35840570 |
Ahalya Prabhakar1, Todd Murphey2.
Abstract
Intelligence involves processing sensory experiences into representations useful for prediction. Understanding sensory experiences and building these contextual representations without prior knowledge of sensor models and environment is a challenging unsupervised learning problem. Current machine learning methods process new sensory data using prior knowledge defined by either domain knowledge or datasets. When datasets are not available, data acquisition is needed, though automating exploration in support of learning is still an unsolved problem. Here we develop a method that enables agents to efficiently collect data for learning a predictive sensor model-without requiring domain knowledge, human input, or previously existing data-using ergodicity to specify the data acquisition process. This approach is based entirely on data-driven sensor characteristics rather than predefined knowledge of the sensor model and its physical characteristics. We learn higher quality models with lower energy expenditure during exploration for data acquisition compared to competing approaches, including both random sampling and information maximization. In addition to applications in autonomy, our approach provides a potential model of how animals use their motor control to develop high quality models of their sensors (sight, sound, touch) before having knowledge of their sensor capabilities or their surrounding environment.Entities:
Mesh:
Year: 2022 PMID: 35840570 PMCID: PMC9287329 DOI: 10.1038/s41467-022-31795-2
Source DB: PubMed Journal: Nat Commun ISSN: 2041-1723 Impact factor: 17.694
Fig. 1Model learning approach.
The learning approach consists of a learning algorithm for creating generative sensory models from data and an active exploration framework that collects data based on the conditional entropy of the learned model. Conditional autoencoders are used to learn a low-dimensional latent encoding of the high-dimensional sensor input. Sensor output at novel states—the generative aspect of the model—is then predicted using this low-dimensional encoding. We develop an active exploration approach for data acquisition based on ergodic coverage of the entropy of the latent space to improve model uncertainty. This framework demonstrates a feasible approach for real-time, generalizable sensor model learning without the need for predefined domain knowledge of the environment or sensor structure.
Fig. 2Electrosense model learning.
We compare the predicted measurement of the learned electrosensory model to the physics-based electrosensory model. The model is an approximation of the sensory system of a weakly electric fish (the black ghost knifefish[16]). Using a seed image, we show the reconstruction of the physics-based measurement field estimate near the object interest, using a conditional autoencoder with active learning using random sampling, and the proposed ergodic sampling strategy. The proposed approach explores the environment based on the evolution of the entropy of the latent space over time (shown evolving over time along with the resulting trajectory shown in blue). Despite no a priori information about the object's location (indicated with black markers) or measurement model, as time evolves the entropy of the latent space—representing where the information affecting the model estimate is highest—converges near the object location and the ergodic sampling spends significantly more time exploring that region without fixating on a particular state. As a result, the ergodic sampling approach results in a significantly better reconstruction of the predicted sensor observation with much lower energy expenditure needed during exploration.
Fig. 3RGB camera model learning.
We compare the predicted measurement generated from the learned 3-channel RGB camera model. Using a seed image, we show the reconstruction of the visual camera field estimate near the object interest, using a CVAE with active learning using random sampling, and the proposed ergodic sampling strategy. Despite no a priori information about the object's location or measurement model, as time evolves, ergodic sampling spends significantly more time exploring the region around the object in the environment. However, unlike the near-field electrosensory model, the camera characteristics generates exploration further away from the object, where the object appears in the periphery of the camera view, instead of remaining directly over the object. The ergodic sampling approach (shown in blue) requires much lower energy expenditure compared to random sampling (shown in green). Furthermore, the ergodic sampling approach results in a significantly better reconstruction of the predicted sensor observation, better capturing the object's characteristics (i.e, shape, color) and the characteristics of the sensor-environment interactions (i.e., camera lighting). More detail can be found in the Supplementary Information.