| Literature DB >> 31480384 |
Xiangbo Kong1, Lehan Chen1, Zhichen Wang1, Yuxi Chen2, Lin Meng3, Hiroyuki Tomiyama4.
Abstract
Vision-based fall-detection methods have been previously studied but many have limitations in terms of practicality. Due to differences in rooms, users do not set the camera or sensors at the same height. However, few studies have taken this into consideration. Moreover, some fall-detection methods are lacking in terms of practicality because only standing, sitting and falling are taken into account. Hence, this study constructs a data set consisting of various daily activities and fall events and studies the effect of camera/sensor height on fall-detection accuracy. Each activity in the data set is carried out by eight participants in eight directions and taken with the depth camera at five different heights. Many related studies heavily depended on human segmentation by using Kinect SDK but this is not reliable enough. To address this issue, this study proposes Enhanced Tracking and Denoising Alex-Net (ETDA-Net) to improve tracking and denoising performance and classify fall and non-fall events. Experimental results indicate that fall-detection accuracy is affected by camera height, against which ETDA-Net is robust, outperforming traditional deep learning based fall-detection methods.Entities:
Keywords: camera height; fall detection; practical; self-adaptation
Year: 2019 PMID: 31480384 PMCID: PMC6749320 DOI: 10.3390/s19173768
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Evaluation scenarios in data set.
| Fall Events | Non-fall Events |
|---|---|
| Facing left, with curled up legs | walking |
| Facing left, without curled up legs | running |
| Facing right, with curled up legs | kicking |
| Facing right, without curled up legs | bowing |
| Facing floor, with curled up legs | bending |
| Facing floor, without curled up legs | walking with stoop |
| Facing ceiling, with curled up legs | clipping |
| Facing ceiling, without curled up legs | raising hand |
| Facing left, with left hand moving | waving hand |
| Facing left, with right hand moving | looking at watch |
| Facing right, with left hand moving | using smartphone |
| Facing right, with right hand moving | throwing |
| Facing floor, with left hand moving | drinking water |
| Facing floor, with right hand moving | collecting |
| Facing ceiling, with left hand moving | walking with stick |
| Facing ceiling, with right hand moving | sitting on chair/floor |
Figure 1Diagram of our data set.
Information of 8 participants.
| Participant ID | Gender | Age | Weight [kg] | Height [cm] |
|---|---|---|---|---|
| A | Male | 24 | 89 | 172 |
| B | Female | 24 | 52 | 161 |
| C | Male | 22 | 93 | 173 |
| D | Male | 29 | 72 | 178 |
| E | Female | 23 | 48 | 158 |
| F | Male | 25 | 60 | 167 |
| G | Male | 24 | 63 | 171 |
| H | Female | 24 | 65 | 160 |
Figure 2Procedure of Enhanced Tracking and Denoising Alex-Net (ETDA-Net).
Figure 3Human segmentation using Kinect SDK. Five lines are human segmentation results from Kinect set at 1.7, 1.9, 2.1, 2.3 and 2.5 m.
Figure 4ETDA-Net.
Experimental results from camera-height detection.
|
| 1.70 | 1.90 | 2.10 | 2.30 | 2.50 |
|
| 1.62 | 1.83 | 2.03 | 2.23 | 2.43 |
|
| 1.71 | 1.92 | 2.12 | 2.32 | 2.52 |
Figure 5Experimental results of human segmentation. First row is depth images, second row is human segmentation using Kinect SDK and the third row is human segmentation using ETDA-Net.
Experimental results from sensitivity comparison.
| Training Data | Test Data | HOG+SVM | LeNet | AlexNet | GoogLeNet | ETDA-Net |
|---|---|---|---|---|---|---|
| 1.7 m | 1.7 m |
|
|
|
|
|
| 1.9 m | 88.2% | 80.6% | 87.0% |
| 89.4% | |
| 2.1 m | 92.0% | 99.0% | 99.0% |
| 99.6% | |
| 2.3 m | 43.6% | 23.0% | 66.0% |
| 61.2% | |
| 2.5 m | 47.2% | 15.8% | 69.0% |
| 76.6% | |
| 1.9 m | 1.7 m | 97.0% | 95.0% | 99.6% |
|
|
| 1.9 m | 99.8% |
|
|
|
| |
| 2.1 m |
|
|
|
|
| |
| 2.3 m | 74.4% | 85.8% | 91.4% |
| 96.4% | |
| 2.5 m | 81.2% | 81.2% | 92.0% |
| 95.6% | |
| 2.1 m | 1.7 m | 88.0% | 94.8% | 77.4% |
| 92.6% |
| 1.9m |
| 91.4% | 85.6% | 92.8% | 90.8% | |
| 2.1 m | 99.8% |
|
|
|
| |
| 2.3 m |
| 70.4% | 68.8% | 82.0% | 72.8% | |
| 2.5 m |
| 30.0% | 42.8% | 57.6% | 72.8% | |
| 2.3 m | 1.7 m | 47.4% | 37.2% | 93.6% | 95.2% |
|
| 1.9 m | 83.2% | 90.8% | 99.4% | 98.8% |
| |
| 2.1 m | 89.2% | 100.0% | 100.0% | 98.6% |
| |
| 2.3m | 99.8% | 99.8% |
| 100.0% |
| |
| 2.5 m | 97.0% | 69.8% |
| 96.6% | 99.8% | |
| 2.5 m | 1.7 m | 35.8% | 0.0% | 51.6% | 40.6% |
|
| 1.9 m | 80.4% | 46.4% |
| 55.6% | 88.4% | |
| 2.1 m | 91.4% | 54.2% | 67.8% | 50.4% |
| |
| 2.3 m | 98.0% | 98.6% | 97.4% | 79.6% |
| |
| 2.5 m | 100.0% | 99.8% |
|
|
| |
| All | 1.7 m | - | 93.7% | 98.4% | 98.9% | 99.2% |
| 1.9 m | - | 99.8% | 99.9% | 99.8% |
| |
| 2.1 m | - | 96.2% |
| 99.7% | 99.4% | |
| 2.3 m | - | 99.7% | 99.8% |
| 99.8% | |
| 2.5 m | - | 99.6% | 99.5% |
| 99.6% |
Experimental results from specificity comparison.
| Training Data | Test Data | HOG+SVM | LeNet | AlexNet | GoogLeNet | ETDA-Net |
|---|---|---|---|---|---|---|
| 1.7 m | 1.7 m |
|
|
|
|
|
| 1.9 m | 99.0% | 99.4% |
|
|
| |
| 2.1 m | 99.6% | 99.4% | 99.8% | 99.4% |
| |
| 2.3 m | 95.4% | 94.8% | 98.2% | 97.2% |
| |
| 2.5 m | 76.0% | 83.2% |
| 84.8% | 92.0% | |
| 1.9 m | 1.7 m | 99.6% |
|
| 99.8% |
|
| 1.9 m |
| 99.8% |
|
|
| |
| 2.1 m | 99.8% | 98.6% |
|
|
| |
| 2.3m | 96.8% | 97.6% | 99.2% | 98.8% |
| |
| 2.5 m | 83.6% | 94.4% | 97.2% | 95.2% |
| |
| 2.1 m | 1.7 m | 99.4% |
|
|
|
|
| 1.9 m | 97.2% |
|
|
|
| |
| 2.1 m |
|
|
|
|
| |
| 2.3 m | 94.4% | 99.4% | 99.8% |
|
| |
| 2.5 m | 76.2% | 98.6% | 99.4% | 99.4% |
| |
| 2.3 m | 1.7 m |
|
|
|
|
|
| 1.9 m | 96.4% | 97.0% | 97.4% | 99.0% |
| |
| 2.1 m | 96.4% | 98.0% | 99.0% | 99.8% |
| |
| 2.3 m | 99.4% | 99.6% |
| 99.8% | 99.8% | |
| 2.5 m | 99.4% |
| 99.4% | 99.6% | 99.4% | |
| 2.5 m | 1.7 m |
| 98.6% | 97.2% | 99.0% |
|
| 1.9 m | 94.0% | 89.2% | 82.0% |
| 97.0% | |
| 2.1 m | 89.2% | 88.8% | 82.0% | 98.4% |
| |
| 2.3 m | 94.4% | 94.0% | 96.2% | 99.8% |
| |
| 2.5 m | 99.6% | 99.6% |
|
|
| |
| All | 1.7 m | - | 98.9% | 99.8% | 99.7% |
|
| 1.9 m | - | 99.7% | 99.8% | 99.8% |
| |
| 2.1 m | - | 98.9% | 99.1% | 99.7% |
| |
| 2.3 m | - | 97.3% | 98.3% | 98.4% |
| |
| 2.5 m | - | 94.5% | 94.5% |
| 96.3% |
Experimental results from accuracy comparison.
| Training Data | Test Data | HOG+SVM | LeNet | AlexNet | GoogLeNet | ETDA-Net |
|---|---|---|---|---|---|---|
| 1.7 m | 1.7 m |
|
|
|
|
|
| 1.9 m | 93.6% | 90.0% | 93.5% |
| 94.7% | |
| 2.1 m | 95.8% | 99.2% | 99.4% | 99.7% |
| |
| 2.3 m | 69.5% | 58.9% | 82.1% |
| 80.1% | |
| 2.5 m | 61.6% | 49.5% | 80.7% |
| 84.3% | |
| 1.9 m | 1.7 m | 98.3% | 97.5% | 99.8% | 99.9% |
|
| 1.9 m | 99.9% | 99.9% |
|
|
| |
| 2.1 m | 99.9% | 99.3% |
|
|
| |
| 2.3 m | 85.6% | 91.7% | 95.3% |
| 98.0% | |
| 2.5 m | 82.4% | 87.8% | 94.6% | 95.8% |
| |
| 2.1 m | 1.7 m | 93.7% | 97.4% | 88.7% |
| 96.3% |
| 1.9 m |
| 95.7% | 92.8% | 96.4% | 95.4% | |
| 2.1 m | 99.9% |
|
|
|
| |
| 2.3 m | 87.2% | 84.9% | 84.3% |
| 86.4% | |
| 2.5 m | 79.7% | 64.3% | 71.1% | 78.5% |
| |
| 2.3 m | 1.7 m | 73.7% | 68.6% | 96.8% | 97.6% |
|
| 1.9 m | 89.8% | 93.9% | 98.4% | 98.9% |
| |
| 2.1 m | 92.8% | 99.0% | 99.5% | 99.2% |
| |
| 2.3 m | 99.6% | 99.7% |
| 99.9% | 99.9% | |
| 2.5 m | 98.2% | 84.7% |
| 98.1% | 99.6% | |
| 2.5 m | 1.7 m | 67.9% | 49.3% | 74.4% | 69.8% |
|
| 1.9 m | 87.2% | 67.8% | 86.1% | 77.4% |
| |
| 2.1 m | 90.3% | 71.5% | 74.9% | 74.4% |
| |
| 2.3 m | 96.2% | 96.3% | 96.8% | 89.7% |
| |
| 2.5 m | 99.8% | 99.7% | 99.3% | 99.0% |
| |
| All | 1.7 m | - | 96.3% | 99.1% | 99.3% |
|
| 1.9 m | - | 99.7% | 99.8% | 99.8% |
| |
| 2.1 m | - | 97.6% | 99.5% | 99.7% |
| |
| 2.3 m | - | 98.5% | 99.0% | 99.2% |
| |
| 2.5 m | - | 95.5% | 97.0% | 98.3% |
|