| Literature DB >> 32599604 |
Lukas von Ziegler1,2, Oliver Sturman1,2, Johannes Bohacek3,4.
Abstract
The assessment of rodent behavior forms a cornerstone of preclinical assessment in neuroscience research. Nonetheless, the true and almost limitless potential of behavioral analysis has been inaccessible to scientists until very recently. Now, in the age of machine vision and deep learning, it is possible to extract and quantify almost infinite numbers of behavioral variables, to break behaviors down into subcategories and even into small behavioral units, syllables or motifs. However, the rapidly growing field of behavioral neuroethology is experiencing birthing pains. The community has not yet consolidated its methods, and new algorithms transfer poorly between labs. Benchmarking experiments as well as the large, well-annotated behavior datasets required are missing. Meanwhile, big data problems have started arising and we currently lack platforms for sharing large datasets-akin to sequencing repositories in genomics. Additionally, the average behavioral research lab does not have access to the latest tools to extract and analyze behavior, as their implementation requires advanced computational skills. Even so, the field is brimming with excitement and boundless opportunity. This review aims to highlight the potential of recent developments in the field of behavioral analysis, whilst trying to guide a consensus on practical issues concerning data collection and data sharing.Entities:
Mesh:
Year: 2020 PMID: 32599604 PMCID: PMC7688651 DOI: 10.1038/s41386-020-0751-7
Source DB: PubMed Journal: Neuropsychopharmacology ISSN: 0893-133X Impact factor: 8.294
Fig. 1Families of machine learning approaches used in behavioral research.
a Supervised machine learning methods are used to first train a classifier on manually defined behaviors that then recognizes these based on feature data in new videos. b Unsupervised machine learning methods are used to find clusters of similar behavioral syllables without human interaction directly from video data. c Pose estimation algorithms track animal body points in videos.
Studies that implemented automated behavior recognition solutions using supervised machine learning approaches.
| Publication | Features | Classifier | Reported Performance | Issues | Available dataset |
|---|---|---|---|---|---|
| Rousseau et al. (2000) [ | 3 points | Feed-forward neural network | Sub-human | Overfitting | No |
| Jhuang et al. (2010) [ | Location, movement | SVM HMM | Human | Position is cage dependent | Yes |
| Burgos-Artizzu (2012) [ | Trajectory, spatio-temporal | Adaboost | Sub-human | Performance | Yes |
| Kabra et al. (2013) [ | Appearance, location | Gentleboost | Human | Includes arena dependent features | Yes |
| Giancardo et al. (2013) [ | Location, appearance, movement | Temporal random forest | Human | More data needed for some behaviors | Yes |
| Van Dam et al. (2013) [ | Appearance, location, movement | GMM | Sub-human | Performance; Company selling the system | No |
| Hong et al (2015) [ | Location, appearance, movement | Random decision forest | No human comparison | Relies on different colored animals | No |
| Lorbach et al. (2018) [ | Location, movement | GMM | No human comparison | Company selling the system | No |
| Le et al. (2019) [ | Spatio-temporal features from 3dCNN (convolutional neural network) 8 × 128 × 128 pixel segments | LSTM | Worse than [ | Computationally Intensive approach with no increase ìn performance | No |
| Chaumont et al. (2019) [ | Intensity histogram, depth histogram (3d camera) | Random decision forest | No human comparison | Requires implanted trackers and 3d cameras | Yes |
| Nguyen et al. (2019) [ | Video segments, 224×224 pixels | I3D and R(2 + 1)D | Better than [ | No transferability assessed | No |
| Van Dam et al. (2020) [ | Video segments, 225×225 pixels | 3dCNN with multifiber blocks | Better than [ | Changes in environment/treatment render it useless | No |
| Sturman et al. (2020) [ | Temporally resolved skeleton features from 2D pose estimate | Feed forward neural network | Human, better than commercial solutions | Only few behaviors analysed, transferability not assessed, preprint | Yes |
| Nilsson et al. (2020) [ | Temporally resolved skeleton features from 2D pose estimate | Random decision forest | No human comparison | preprint | No, reannotation of [ |
SVM support vector machine, HMM hidden markov model, GMM gaussian mixture model, LSTM long short-term memory, 3dCNN 3 dimensional convolutional neural network, I3D two-stream inflated 3dCNN.
Fig. 2Proposed workflow for high-fidelity, transferable behavior recording.
a A high-fidelity point-set is selected that retains most of the animal information for the least storage space required. b Behavioral tests are recorded from multiple perspectives with synchronized cameras. Pose estimation algorithms such as DeepLabCut are used to track the defined points. Undistortion is applied either to the videos directly or to the data. c Tracked point data is used to create a behavior tracking data object that contains all essential information about the behavioral test that can be used for any post-hoc analysis. This object is used for long-term storage in online repositories. d Behavior tracking data objects can be used to create a feature data object that contains all features that are important to recognize a selected behavior. Setup-specific normalization factors are contained within the feature object to allow easy transferability. e Feature objects are used in combination with existing classifiers to automatically track behaviors, or a new classifier can be trained in combination with manually annotated training data. f Example data comparing the proposed workflow to commercial solutions (Ethovision XT 14, TSE Systems) and humans. Supported rearing behavior is recognized with human accuracy when using features generated from 2D (top view) point-tracking data (adapted from ref. [26]. g Correlation between three human raters, the machine learning classifiers, and the commercial systems from the same study.