| Literature DB >> 32328484 |
Junjie Wang1, Xiaohong Su1, Lingling Zhao1, Jun Zhang2.
Abstract
Accurate target detection and association are vital for the development of reliable target tracking, especially for cell tracking based on microscopy images due to the similarity of cells. We propose a deep reinforcement learning method to associate the detected targets between frames. According to the dynamic model of each target, the cost matrix is produced by conjointly considering various features of targets and then used as the input of a neural network. The proposed neural network is trained using reinforcement learning to predict a distribution over the association solution. Furthermore, we design a residual convolutional neural network that results in more efficient learning. We validate our method on two applications: the multiple target tracking simulation and the ISBI cell tracking. The results demonstrate that our approach based on reinforcement learning techniques could effectively track targets following different motion patterns and show competitive results.Entities:
Keywords: cell tracking; data association; deep learning; deep reinforcement learning; linear assignment problem; residual CNN
Year: 2020 PMID: 32328484 PMCID: PMC7161216 DOI: 10.3389/fbioe.2020.00298
Source DB: PubMed Journal: Front Bioeng Biotechnol ISSN: 2296-4185
Figure 1Illustration of the proposed association matrix. is the prediction by the Kalman filter. are the measurements.
Figure 2The architecture of the proposed ResCNN network.
Median optimality ratios on the MWM test set.
| AC+Matching | 0.935 | 0.897 | 0.725 |
| SPG+Matching | 0.904 | 0.895 | 0.889 |
| Ours | 0.977 | 0.968 | 0.965 |
Figure 3Comparison of the track maintenance performance of different algorithms: (A) Ground-truth trajectories of the five targets, (B) the measurements of the five targets, (C) the JPDA filter, (D) our proposed method. Each color corresponds to a particular target. Note that our method correctly resolves this crossing case, whereas the JPDA filter switches the two trajectories after the targets cross.
Average OSPA-T distance and IDSW for different methods over 100 random runs.
| JPDA | 0.19(0.05) | 0.90(0.88) | 0.34(0.11) | 0.70(0.82) | 0.41(0.12) | 0.40(0.70) |
| JPDA10 | 0.23(0.10) | 0.70(0.67) | 0.37(0.11) | 0.90(0.88) | 0.43(0.09) | 1.10(0.99) |
| JPDA-HA | 0.28(0.06) | 0.60(0.84) | 0.37(0.10) | 0.70(0.95) | 0.46(0.14) | 1.30(0.82) |
| JPDA-RL | 0.28(0.06) | 0.60(0.84) | 0.36(0.08) | 0.60(0.70) | 0.45(0.13) | 1.10(0.99) |
| LSTM | 0.11(0.01) | 1.07(0.84) | 0.21(0.01) | 1.00(0.74) | 0.37(0.11) | 0.60(0.89) |
The standard deviations are given in parentheses.
TRA, SEG and OPT performance for our method, CPN, KTH (Magnusson and Jaldén, 2012), BLOB (Akram et al., 2016), U-Net (Ronneberger et al., 2015), U-Net-S (Gupta et al., 2019), and GC-ME (Bensch and Ronneberger, 2015).
| Fluo-N2DH-GOWT1-01 | CPN | 0.9864 | 0.8506 | 0.9185 |
| BLOB | 0.9733 | 0.7415 | 0.8574 | |
| KTH | 0.9462 | 0.6849 | 0.8155 | |
| Ours | ||||
| Fluo-N2DH-GOWT1-02 | CPN | 0.8725 | 0.9222 | |
| BLOB | 0.9628 | 0.9046 | 0.9337 | |
| KTH | 0.9452 | 0.8942 | 0.9197 | |
| Ours | 0.9575 | |||
| PhC-C2DH-U373-01 | CPN | 0.9594 | 0.7336 | 0.8456 |
| U-Net | 0.9869 | |||
| GC-ME | 0.9779 | 0.8748 | 0.9264 | |
| Ours | 0.8527 | 0.9223 | ||
| PhC-C2DH-U373-02 | CPN | 0.9346 | 0.7376 | 0.8361 |
| U-Net | 0.8925 | |||
| GC-ME | 0.9040 | 0.7567 | 0.8304 | |
| Ours | 0.9318 | 0.7735 | 0.8527 | |
| Fluo-N2DH-SIM+-01 | U-Net-S | |||
| Ours | 0.9841 | 0.8854 | 0.9348 | |
| Fluo-N2DH-SIM+-02 | U-Net-S | 0.9597 | 0.7381 | 0.8489 |
| Ours |
The best TRA and SEG values for each sequence are highlighted.
Training Procedure
| 1: Training set |
| 2: Initialize the neural net params θ. |
| 3: Initialize baseline value. |
| 4: |
| 5: Select a batch of samples |
| 6: Sample solution π |
| 7: Let |
| 8: Update θ = |
| 9: Update baseline |
| 10: |
| 11: return neural net parameters θ. |