| Literature DB >> 35545645 |
Yujie Lei1,2, Pengmei Dong3, Yan Guan1, Ying Xiang1, Meng Xie4, Jiong Mu5,6, Yongzhao Wang1, Qingyong Ni7.
Abstract
The precise identification of postural behavior plays a crucial role in evaluation of animal welfare and captive management. Deep learning technology has been widely used in automatic behavior recognition of wild and domestic fauna species. The Asian slow loris is a group of small, nocturnal primates with a distinctive locomotion mode, and a large number of individuals were confiscated into captive settings due to illegal trade, making the species an ideal as a model for postural behavior monitoring. Captive animals may suffer from being housed in an inappropriate environment and may display abnormal behavior patterns. Traditional data collection methods are time-consuming and laborious, impeding efforts to improve lorises' captive welfare and to develop effective reintroduction strategies. This study established the first human-labeled postural behavior dataset of slow lorises and used deep learning technology to recognize postural behavior based on object detection and semantic segmentation. The precision of the classification based on YOLOv5 reached 95.1%. The Dilated Residual Networks (DRN) feature extraction network showed the best performance in semantic segmentation, and the classification accuracy reached 95.2%. The results imply that computer automatic identification of postural behavior may offer advantages in assessing animal activity and can be applied to other nocturnal taxa.Entities:
Mesh:
Year: 2022 PMID: 35545645 PMCID: PMC9095646 DOI: 10.1038/s41598-022-11842-0
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.996
Enclosure characteristics of each captive site for video data collection.
| Captive site | Dehong | Xishuangbanna | Puer |
|---|---|---|---|
| Coordinate | 24.38287°N, 98.45872°E | 22.39276°N, 100.89636°E | 22.62198°N, 101.08916°E |
| Altitude (m) | 850 | 1060 | 1600 |
| Annual mean temperature (℃) | 19.6 | 17.5 | 17.5 |
| No. of individuals | 4 | 9 | 9 |
| No. of enclosures | 1 | 1 | 1 |
| Enclosure size (L × W × H) (m) | 3.5 × 3.4 × 3.8 | 5.7 × 4.2 × 3.5 | 3.5 × 2.1 × 2.0 |
| No. of cameras | 2 | 3 | 2 |
Figure 1Processing of target detection, semantic segmentation and classification.
Definition of the parameters.
| Actual result | |||
|---|---|---|---|
| Positive | Negative | ||
| Expected result | Positive | TP | FN |
| Negative | FP | TN | |
Figure 2The images of postural behavior extracted by YOLOv5.
Figure 3DRN network structure diagram.
Figure 4DeepLabv3 + network structure diagram.
Figure 5Semantic segmentation image of loris postural behavior (red indicate moving behavior and green indicate resting).
The effect of the YOLOv5 network.
| Postural behavior | Precision | Recall | mAP |
|---|---|---|---|
| All | 0.951 | 0.938 | 0.949 |
| Feeding | 0.951 | 0.932 | 0.951 |
| Move-rest | 0.941 | 0.948 | 0.953 |
| Socializing | 0.961 | 0.940 | 0.942 |
Figure 6Precision, recall and mAP of object detection classification.
Training effects of four different networks.
| Network | Training time | Category | Eopch10 | Eopch20 | Eopch30 | Eopch40 | Epoch50 |
|---|---|---|---|---|---|---|---|
| DRN | 2h20m | Acc | 0.926 | 0.955 | 0.961 | 0.964 | 0.968 |
| AccClass | 0.913 | 0.940 | 0.940 | 0.943 | 0.952 | ||
| MobilNet | 38 m | Acc | 0.908 | 0.933 | 0.939 | 0.955 | 0.959 |
| AccClass | 0.902 | 0.916 | 0.894 | 0.939 | 0.943 | ||
| ResNet | 1h23m | Acc | 0.895 | 0.928 | 0.933 | 0.938 | 0.944 |
| AccClass | 0.878 | 0.934 | 0.938 | 0.938 | 0.944 | ||
| Xception | 1h34m | Acc | 0.764 | 0.782 | 0.815 | 0.847 | 0.860 |
| AccClass | 0.582 | 0.677 | 0.706 | 0.787 | 0.801 |
Figure 7Evaluation index of the second step of joint training: (a) Semantic segmentation accuracy, (b) Classification accuracy, (c) MIoU, (d) FWIoU.