| Literature DB >> 33918839 |
Roberto J López-Sastre1, Marcos Baptista-Ríos2, Francisco Javier Acevedo-Rodríguez1, Soraya Pacheco-da-Costa3, Saturnino Maldonado-Bascón1, Sergio Lafuente-Arroyo1.
Abstract
In this paper, we present a new low-cost robotic platform that has been explicitly developed to increase children with neurodevelopmental disorders' involvement in the environment during everyday living activities. In order to support the children and youth with both the sequencing and learning of everyday living tasks, our robotic platform incorporates a sophisticated online action detection module that is capable of monitoring the acts performed by users. We explain all the technical details that allow many applications to be introduced to support individuals with functional diversity. We present this work as a proof of concept, which will enable an assessment of the impact that the developed technology may have on the collective of children and youth with neurodevelopmental disorders in the near future.Entities:
Keywords: action detection; artificial intelligence; assistive technology; robotics
Year: 2021 PMID: 33918839 PMCID: PMC8069234 DOI: 10.3390/ijerph18083974
Source DB: PubMed Journal: Int J Environ Res Public Health ISSN: 1660-4601 Impact factor: 3.390
Figure 1Pictures of our low-cost robotic platform. We show both a frontal picture and the internal structure, where it is possible to observe all the electronic and mechanical components of the platform.
A list with the prices of the main components needed to build our platform.
| Item | Estimated Price |
|---|---|
| Motor and encoders | 132 € |
| Arduino MEGA | 12 € |
| Battery | 24 € |
| Wheels | 66 € |
| Structure & Components | 100 € |
| Screen | 40 € |
| Jetson TX2 | 320 € |
| Camera | 20 € |
| LIDAR | 90 € |
| Total | 804 € |
Figure 2Robotic Operating System (ROS)-based complete software architecture.
Figure 3This figure shows the deep learning model OAD-3D-CNN, which is used for the monitoring of the daily-life activities of the users. Our model has 8 3D convolutions, 5 max-pooling layers and 3 fully connected layers, followed by a softmax output layer. All 3D convolution kernels are with stride = 1 in both spatial and temporal dimensions. Note that in each box we indicate the number of filters. The 3D pooling layers are denoted pool1 to pool5, where the pooling kernels are , except pool1, for which it is . The fully connected layers have 4096 output units.
A list detailing the subset of the action categories from the UCF-101 dataset that we have implemented in the online action detection (OAD) software module. We also report the number of videos available for training.
| Category | UCF-101 Class Identifier | Number of Videos |
|---|---|---|
| Apply Eye Makeup | 1 | >130 |
| Apply Lipstick | 2 | >100 |
| Blow Dry Hair | 13 | >120 |
| Brushing Teeth | 20 | >120 |
| Cutting In Kitchen | 25 | >100 |
| Haircut | 28 | >120 |
| Mixing Batter | 35 | >130 |
| Mopping Floor | 55 | >100 |
| Shaving Beard | 56 | >150 |
| Typing | 95 | >120 |
| Walking With Dog | 96 | >120 |
| Writing On Board | 100 | >150 |
Figure 4Visual examples of the 12 different action categories from the UCF-101 dataset used in our online action detection module.
Figure 5Qualitative results of the online action detection module. In this figure we show four images with the results of the action detection, which appear in the upper left margins, where we indicate the category recognized and the confidence of the OAD system for the prediction.