| Literature DB >> 35009677 |
Friedrich Niemann1, Stefan Lüdtke2, Christian Bartelt2, Michael Ten Hompel1.
Abstract
The automatic, sensor-based assessment of human activities is highly relevant for production and logistics, to optimise the economics and ergonomics of these processes. One challenge for accurate activity recognition in these domains is the context-dependence of activities: Similar movements can correspond to different activities, depending on, e.g., the object handled or the location of the subject. In this paper, we propose to explicitly make use of such context information in an activity recognition model. Our first contribution is a publicly available, semantically annotated motion capturing dataset of subjects performing order picking and packaging activities, where context information is recorded explicitly. The second contribution is an activity recognition model that integrates movement data and context information. We empirically show that by using context information, activity recognition performance increases substantially. Additionally, we analyse which of the pieces of context information is most relevant for activity recognition. The insights provided by this paper can help others to design appropriate sensor set-ups in real warehouses for time management.Entities:
Keywords: context awareness; context model; human activity recognition; industrial processes; logistics; motion capture; warehousing
Mesh:
Year: 2021 PMID: 35009677 PMCID: PMC8749739 DOI: 10.3390/s22010134
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1Data processing pipeline of the oMoCap system based on a person to goods order picking process: (a) Reconstruction of a point cloud consisting of markers. (b) Marker labelling based on patterns of markers from the subject and the picking cart. (c) Overlay of the oMoCap visualisation with the video recording.
Figure 2Extracts from the physical laboratory set-up of all three logistics scenarios.
Subjects—specifications and scenario assignment.
| ID | Sex | Age | Weight | Height | Handedness | No. of Two-Minute Recordings | ||
|---|---|---|---|---|---|---|---|---|
| [F/M] | [years] | [kg] | [cm] | [L/R] | L01 | L02 | L03 | |
| S17 | M | 30 | 85 | 176 | R | 5 | 15 | 15 |
| S18 | F | 26 | 62 | 177 | R | 5 | 15 | 15 |
Figure 3Entire laboratory set-up as an oMoCap visualisation.
Objects—specifications and scenario assignment.
| Name in | No. of | Logistics Scenario | ||
|---|---|---|---|---|
| Dataset Files | Marker | L01 | L02 | L03 |
| Base | 5 | x | ||
| Cart_large | 15 | x | ||
| Cart_small | 4 | x | ||
| Entrance_BreakRoom | 5 | x | x | x |
| Entrance_CartRoom | 4 | x | x | |
| Entrance_Office | 5 | x | x | |
| FlowThroughRack | 19 | x | x | |
| PackagingTable | 11 | x | ||
| Rack01 | 5 | x | x | |
| Rack02 | 5 | x | x | |
| Rack03 | 5 | x | x | |
| RackComplex | 11 | x | x | |
Figure 4Marker designations of the packaging table.
Annotation results divided by activity classes.
| Activity Class | Frames | Windows | |||||
|---|---|---|---|---|---|---|---|
| No. | % | No. | % | Min. Length [Frames] | Max. Length [Frames] | ||
|
| Standing | 112,228 | 6.68 | 432 | 7.51 | 32 | 2348 |
|
| Walking | 181,596 | 10.81 | 326 | 5.67 | 61 | 8800 |
|
| Cart | 207,774 | 12.37 | 315 | 5.48 | 60 | 4400 |
|
| Handling (upwards) | 137,911 | 8.21 | 589 | 10.24 | 26 | 1400 |
|
| Handling (centred) | 954,959 | 56.84 | 3732 | 64.90 | 26 | 4607 |
|
| Handling (downwards) | 72,368 | 4.31 | 339 | 5.90 | 29 | 1384 |
|
| None | 13,164 | 0.78 | 17 | 0.30 | 200 | 2600 |
| 1,680,000 | 100 | 5750 | 100 | ||||
Figure 5Example of an annotation sequence of six windows.
Figure 6Activity recognition architecture. The pre-trained neural network predicts high-level movement descriptors (attributes). Together with the context data, they are used by a shallow classifier to predict activity classes. Figure adapted from [33].
Macro F1 scores of the different classifiers and feature subsets.
| Classifier | Att. (Base Model) | Dists | Att. + Dists | Raw | Att. + Raw |
|---|---|---|---|---|---|
| QDA | 0.670 | 0.515 | 0.619 | – | – |
| XGBoost | 0.717 | 0.730 | 0.643 | 0.691 | 0.727 |
| RF | 0.716 | 0.646 | 0.733 | 0.745 | 0.820 |
Class-wise F1 scores of RF model.
| Standing | Walking | Cart | Handling | Handling | Handling | |
|---|---|---|---|---|---|---|
| Upwards | Centred | Downwards | ||||
| Att. (Base model) | 0.210 | 0.743 | 0.809 | 0.880 | 0.761 | 0.815 |
| Dists | 0.063 | 0.716 | 0.819 | 0.858 | 0.508 | 0.593 |
| Att. + Dists | 0.156 | 0.842 | 0.880 | 0.886 | 0.743 | 0.822 |
| Raw | 0.214 | 0.823 | 0.803 | 0.808 | 0.687 | 0.804 |
| Att. + Raw | 0.150 | 0.833 | 0.880 | 0.825 | 0.753 | 0.826 |
Greedy feature selection results (QDA, distance features).
| Marker 1 | Marker 2 | F1 Score |
|---|---|---|
| Attributes (Base model) | 0.680 | |
| + Cart_large:C_large_PtL_3_l | Subject:RFIN | 0.698 |
| + PackagingTable:barcodes | Subject:RFIN | 0.705 |
| + Cart_small:C_small_handle | Subject:RFIN | 0.711 |
| + Cart_small:C_small_ba_t | Subject:RFIN | 0.718 |
| + Cart_small:C_small_ba_t | Subject:LFIN | 0.719 |
| + PackagingTable:bubblewrap | Subject:LFIN | 0.719 |
Greedy feature selection results (RF, distance features).
| Marker 1 | Marker 2 | F1 Score |
|---|---|---|
| Attributes (Base model) | 0.717 | |
| + PackagingTable:barcodes | Subject:LFIN | 0.740 |
Greedy feature selection results (RF, raw features).
| Marker | Axis | F1 Score |
|---|---|---|
| Attributes (Base model) | 0.717 | |
| + PackagingTable:Pac_t_r | z | 0.748 |
| + Cart_small:C_small_handle | y | 0.761 |