Literature DB >> 33498829

Activity Recognition for Ambient Assisted Living with Videos, Inertial Units and Ambient Sensors.

Caetano Mazzoni Ranieri1, Scott MacLeod2, Mauro Dragone2, Patricia Amancio Vargas2, Roseli Aparecida Francelin Romero1.   

Abstract

Worldwide demographic projections point to a progressively older population. This fact has fostered research on Ambient Assisted Living, which includes developments on smart homes and social robots. To endow such environments with truly autonomous behaviours, algorithms must extract semantically meaningful information from whichever sensor data is available. Human activity recognition is one of the most active fields of research within this context. Proposed approaches vary according to the input modality and the environments considered. Different from others, this paper addresses the problem of recognising heterogeneous activities of daily living centred in home environments considering simultaneously data from videos, wearable IMUs and ambient sensors. For this, two contributions are presented. The first is the creation of the Heriot-Watt University/University of Sao Paulo (HWU-USP) activities dataset, which was recorded at the Robotic Assisted Living Testbed at Heriot-Watt University. This dataset differs from other multimodal datasets due to the fact that it consists of daily living activities with either periodical patterns or long-term dependencies, which are captured in a very rich and heterogeneous sensing environment. In particular, this dataset combines data from a humanoid robot's RGBD (RGB + depth) camera, with inertial sensors from wearable devices, and ambient sensors from a smart home. The second contribution is the proposal of a Deep Learning (DL) framework, which provides multimodal activity recognition based on videos, inertial sensors and ambient sensors from the smart home, on their own or fused to each other. The classification DL framework has also validated on our dataset and on the University of Texas at Dallas Multimodal Human Activities Dataset (UTD-MHAD), a widely used benchmark for activity recognition based on videos and inertial sensors, providing a comparative analysis between the results on the two datasets considered. Results demonstrate that the introduction of data from ambient sensors expressively improved the accuracy results.

Entities:  

Keywords:  deep learning; human activity recognition; human–robot interaction; inertial sensors; multimodal datasets; video classification

Year:  2021        PMID: 33498829      PMCID: PMC7865705          DOI: 10.3390/s21030768

Source DB:  PubMed          Journal:  Sensors (Basel)        ISSN: 1424-8220            Impact factor:   3.576


  16 in total

1.  NTU RGB+D 120: A Large-Scale Benchmark for 3D Human Activity Understanding.

Authors:  Jun Liu; Amir Shahroudy; Mauricio Perez; Gang Wang; Ling-Yu Duan; Alex C Kot
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2019-05-14       Impact factor: 6.226

2.  Unobtrusive Activity Recognition of Elderly People Living Alone Using Anonymous Binary Sensors and DCNN.

Authors:  Munkhjargal Gochoo; Tan-Hsu Tan; Shing-Hong Liu; Fu-Rong Jean; Fady S Alnajjar; Shih-Chia Huang
Journal:  IEEE J Biomed Health Inform       Date:  2018-05-07       Impact factor: 5.772

3.  C-MHAD: Continuous Multimodal Human Action Dataset of Simultaneous Video and Inertial Sensing.

Authors:  Haoran Wei; Pranav Chopada; Nasser Kehtarnavaz
Journal:  Sensors (Basel)       Date:  2020-05-20       Impact factor: 3.576

Review 4.  A Comprehensive Survey of Vision-Based Human Action Recognition Methods.

Authors:  Hong-Bo Zhang; Yi-Xiang Zhang; Bineng Zhong; Qing Lei; Lijie Yang; Ji-Xiang Du; Duan-Sheng Chen
Journal:  Sensors (Basel)       Date:  2019-02-27       Impact factor: 3.576

5.  Human Activity Recognition Using Inertial Sensors in a Smartphone: An Overview.

Authors:  Wesllen Sousa Lima; Eduardo Souto; Khalil El-Khatib; Roozbeh Jalali; Joao Gama
Journal:  Sensors (Basel)       Date:  2019-07-21       Impact factor: 3.576

6.  Fusion of Video and Inertial Sensing for Deep Learning-Based Human Action Recognition.

Authors:  Haoran Wei; Roozbeh Jafari; Nasser Kehtarnavaz
Journal:  Sensors (Basel)       Date:  2019-08-24       Impact factor: 3.576

7.  A Novel Human Activity Recognition and Prediction in Smart Home Based on Interaction.

Authors:  Yegang Du; Yuto Lim; Yasuo Tan
Journal:  Sensors (Basel)       Date:  2019-10-15       Impact factor: 3.576

8.  A depth video sensor-based life-logging human activity recognition system for elderly care in smart indoor environments.

Authors:  Ahmad Jalal; Shaharyar Kamal; Daijin Kim
Journal:  Sensors (Basel)       Date:  2014-07-02       Impact factor: 3.576

9.  Deep Convolutional and LSTM Recurrent Neural Networks for Multimodal Wearable Activity Recognition.

Authors:  Francisco Javier Ordóñez; Daniel Roggen
Journal:  Sensors (Basel)       Date:  2016-01-18       Impact factor: 3.576

10.  Using Rough Sets to Improve Activity Recognition Based on Sensor Data.

Authors:  Hans W Guesgen
Journal:  Sensors (Basel)       Date:  2020-03-23       Impact factor: 3.576

View more
  5 in total

1.  Intention Recognition With ProbLog.

Authors:  Gary B Smith; Vaishak Belle; Ronald P A Petrick
Journal:  Front Artif Intell       Date:  2022-04-26

2.  Mathematical Criteria for a Priori Performance Estimation of Activities of Daily Living Recognition.

Authors:  Florentin Delaine; Gregory Faraut
Journal:  Sensors (Basel)       Date:  2022-03-22       Impact factor: 3.576

3.  Preparing Wi-Fi 7 for Healthcare Internet-of-Things.

Authors:  Yazdan Ahmad Qadri; Ali Nauman; Arslan Musaddiq; Eduard Garcia-Villegas; Sung Won Kim
Journal:  Sensors (Basel)       Date:  2022-08-18       Impact factor: 3.847

Review 4.  Characterizing Smart Environments as Interactive and Collective Platforms: A Review of the Key Behaviors of Responsive Architecture.

Authors:  Ju Hyun Lee; Michael J Ostwald; Mi Jeong Kim
Journal:  Sensors (Basel)       Date:  2021-05-14       Impact factor: 3.576

5.  Coherence between Decomposed Components of Wrist and Finger PPG Signals by Imputing Missing Features and Resolving Ambiguous Features.

Authors:  Pei-Yun Tsai; Chiu-Hua Huang; Jia-Wei Guo; Yu-Chuan Li; An-Yeu Andy Wu; Hung-Ju Lin; Tzung-Dau Wang
Journal:  Sensors (Basel)       Date:  2021-06-24       Impact factor: 3.576

  5 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.