| Literature DB >> 33585169 |
Jie Su1, Xiaohai He1, Linbo Qing1, Tong Niu1, Yongqiang Cheng2, Yonghong Peng3.
Abstract
Social distancing in public spaces plays a crucial role in controlling or slowing down the spread of coronavirus during the COVID-19 pandemic. Visual Social Distancing (VSD) offers an opportunity for real-time measuring and analysing the physical distance between pedestrians using surveillance videos in public spaces. It potentially provides new evidence for implementing effective prevention measures of the pandemic. The existing VSD methods developed in the literature are primarily based on frame-by-frame pedestrian detection, addressing the VSD problem from a static and local perspective. In this paper, we propose a new online multi-pedestrian tracking approach for spatio-temporal trajectory and its application to multi-scale social distancing measuring and analysis. Firstly, an online multi-pedestrian tracking method is proposed to obtain the trajectories of pedestrians in public spaces, based on hierarchical data association. Then, a new VSD method based on spatio-temporal trajectories is proposed. The proposed method not only considers the Euclidean distance between tracking objects frame-by-frame but also takes into account the discrete Fréchet distance between trajectories, hence forms a comprehensive solution from both static and dynamic, local and holistic perspectives. We evaluated the performance of the proposed tracking method using the public dataset MOT16 benchmark. We also collected our own pedestrian dataset "SCU-VSD" and designed a multi-scale VSD analysis scheme for benchmarking the performance of the social distancing monitoring in the crowd. Experiments have demonstrated that the proposed method achieved outstanding performance on the analysis of social distancing.Entities:
Keywords: Crowd gathering; Discrete Fréchet distance; Hierarchical data association; Multi-pedestrian tracking; Spatio-temporal trajectory; Visual social distancing
Year: 2021 PMID: 33585169 PMCID: PMC7865092 DOI: 10.1016/j.scs.2021.102765
Source DB: PubMed Journal: Sustain Cities Soc ISSN: 2210-6707 Impact factor: 7.587
Fig. 1The distinction between the detection-based VSD and the trajectory-based VSD.
Fig. 2The transition mechanism of the four states of the tracklets.
Fig. 3The architecture of the AAM-Softmax appearance feature descriptor network.
Fig. 4The flow chart of the proposed hierarchical data association.
Fig. 5The entire process of online multi-object tracking based on hierarchical data association.
Fig. 6The gathering group and the corresponding gathering degree.
Comparisons of different online algorithms on MOT16 benchmark (with private detectors).
| Method | MOTA( | MOTP( | IDS | FM | MT( | ML ( |
|---|---|---|---|---|---|---|
| Config-MOT ( | 43.9 | 76.0 | 1030 | 17.4 | 30.2 | |
| MOTDT ( | 47.6 | 50.9 | 792 | – | 15.2 | |
| STRN ( | 48.5 | 73.7 | – | 17.0 | 34.9 | |
| Deep Sort ( | 781 | 2008 | ||||
| EAMTT ( | 910 | – | 19.0 | 34.9 | ||
| 19.9 |
The values in red and blue represent the optimal and the suboptimal results respectively.
The information of the selected rectangular reference areas for SCU-VSD Dataset.
| SCU-VSD | 01 | 02 | 03 | 04 | 05 | 06 | 07 | 08 | |
|---|---|---|---|---|---|---|---|---|---|
| In the real world | Width (m) | 6 | 6 | 7.8 | 6 | 6 | 6 | 6 | 7.2 |
| Length (m) | 14.7 | 12 | 11 | 24.5 | 18 | 12 | 15 | 12.6 | |
| In the bird’s eye view | Width (pixel) | 60 | 60 | 78 | 60 | 60 | 60 | 60 | 72 |
| Length (pixel) | 147 | 120 | 110 | 245 | 180 | 120 | 150 | 126 | |
Fig. 7The comparisons between the original perspective view and the calibrated bird’s eye view for SCU-VSD.
The coordinates of the four vertex pairs and the perspective transformation matrix for SCU-VSD.
| Dataset | In the original video | In the bird’s eye view | ||
|---|---|---|---|---|
| SCU-VSD-01 | ||||
| SCU-VSD-02 | ||||
| SCU-VSD-03 | ||||
| SCU-VSD-04 | ||||
| SCU-VSD-05 | ||||
| SCU-VSD-06 | ||||
| SCU-VSD-07 | ||||
| SCU-VSD-08 |
Fig. 8The real-time social distancing measurement and analysis for SCU-VSD dataset.
Fig. 9The colourmaps of ARP-USD and AGD for every 10 s.
The four metrics ARP-USD, NTP-USD, NPPC-USD and AGD for each entire video.
| Datasets | ARP-USD ( | NTP-USD | NPP-CUSD | AGD |
|---|---|---|---|---|
| SCU-VSD-01 | 67.01 | 32 | 11 | 1.2 |
| SCU-VSD-02 | 75.90 | 10 | 9 | 1.0 |
| SCU-VSD-03 | 47.62 | 3 | 2 | 0.57 |
| SCU-VSD-04 | 63.84 | 22 | 11 | 1.03 |
| SCU-VSD-05 | 64.48 | 14 | 10 | 0.98 |
| SCU-VSD-06 | 49.59 | 5 | 4 | 0.76 |
| SCU-VSD-07 | 67.64 | 19 | 6 | 0.99 |
| SCU-VSD-08 | 61.60 | 6 | 6 | 0.92 |