Literature DB >> 31403450

Automation of the Timed-Up-and-Go Test Using a Conventional Video Camera.

Patrick Savoie, James A D Cameron, Mary E Kaye, Erik J Scheme.   

Abstract

The Timed-Up-and-Go (TUG) test is a simple clinical tool commonly used to quickly assess the mobility of patients. Researchers have endeavored to automate the test using sensors or motion tracking systems to improve its accuracy and to extract more resolved information about its sub-phases. While some approaches have shown promise, they often require the donning of sensors or the use of specialized hardware, such as the now discontinued Microsoft Kinect, which combines video information with depth sensors (RGBD). In this work, we leverage recent advances in computer vision to automate the TUG test using a regular RGB video camera without the need for custom hardware or additional depth sensors. Thirty healthy participants were recorded using a Kinect V2 and a standard video feed while performing multiple trials of 3 and 1.5 meter versions of the TUG test. A Mask Regional Convolutional Neural Net (R-CNN) algorithm and a Deep Multitask Architecture for Human Sensing (DMHS) were then used together to extract global 3D poses of the participants. The timing of transitions between the six key movement phases of the TUG test were then extracted using heuristic features extracted from the time series of these 3D poses. The proposed video-based vTUG system yielded the same error as the standard Kinect-based system for all six key transitions points, and average errors of less than 0.15 seconds from a multi-observer hand labeled ground truth. This work describes a novel method of video-based automation of the TUG test using a single standard camera, removing the need for specialized equipment and facilitating the extraction of additional meaningful information for clinical use.

Entities:  

Mesh:

Year:  2019        PMID: 31403450     DOI: 10.1109/JBHI.2019.2934342

Source DB:  PubMed          Journal:  IEEE J Biomed Health Inform        ISSN: 2168-2194            Impact factor:   5.772


  3 in total

1.  Automatic Functional Shoulder Task Identification and Sub-task Segmentation Using Wearable Inertial Measurement Units for Frozen Shoulder Assessment.

Authors:  Chih-Ya Chang; Chia-Yeh Hsieh; Hsiang-Yun Huang; Yung-Tsan Wu; Liang-Cheng Chen; Chia-Tai Chan; Kai-Chun Liu
Journal:  Sensors (Basel)       Date:  2020-12-26       Impact factor: 3.576

2.  Effect of baduanjin on the fall and balance function in middle-aged and elderly people: A protocol for systematic review and meta-analysis.

Authors:  Yao Xiao; Qin Luo; Yunyang Yu; Biwei Cao; Miao Wu; Yujun Luo; Yan Zhao; Jing Zhou
Journal:  Medicine (Baltimore)       Date:  2021-09-17       Impact factor: 1.889

3.  Deep Learning-Based Subtask Segmentation of Timed Up-and-Go Test Using RGB-D Cameras.

Authors:  Yoonjeong Choi; Yoosung Bae; Baekdong Cha; Jeha Ryu
Journal:  Sensors (Basel)       Date:  2022-08-23       Impact factor: 3.847

  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.