| Literature DB >> 26134417 |
V De Luca1, T Benz, S Kondo, L König, D Lübke, S Rothlübbers, O Somphone, S Allaire, M A Lediju Bell, D Y F Chung, A Cifor, C Grozea, M Günther, J Jenne, T Kipshagen, M Kowarschik, N Navab, J Rühaak, J Schwaab, C Tanner.
Abstract
The Challenge on Liver Ultrasound Tracking (CLUST) was held in conjunction with the MICCAI 2014 conference to enable direct comparison of tracking methods for this application. This paper reports the outcome of this challenge, including setup, methods, results and experiences. The database included 54 2D and 3D sequences of the liver of healthy volunteers and tumor patients under free breathing. Participants had to provide the tracking results of 90% of the data (test set) for pre-defined point-landmarks (healthy volunteers) or for tumor segmentations (patient data). In this paper we compare the best six methods which participated in the challenge. Quantitative evaluation was performed by the organizers with respect to manual annotations. Results of all methods showed a mean tracking error ranging between 1.4 mm and 2.1 mm for 2D points, and between 2.6 mm and 4.6 mm for 3D points. Fusing all automatic results by considering the median tracking results, improved the mean error to 1.2 mm (2D) and 2.5 mm (3D). For all methods, the performance is still not comparable to human inter-rater variability, with a mean tracking error of 0.5-0.6 mm (2D) and 1.2-1.8 mm (3D). The segmentation task was fulfilled only by one participant, resulting in a Dice coefficient ranging from 76.7% to 92.3%. The CLUST database continues to be available and the online leader-board will be updated as an ongoing challenge.Entities:
Mesh:
Year: 2015 PMID: 26134417 PMCID: PMC5454593 DOI: 10.1088/0031-9155/60/14/5571
Source DB: PubMed Journal: Phys Med Biol ISSN: 0031-9155 Impact factor: 3.609
Summary of the challenge data with annotation of point-landmarks (2D sequences: ETH, MED1 and MED2; and 3D sequences: EMC, ICR and SMT) and segmentations of tumor areas (2D sequences: OX).
| Sequence | Sequence info | No.ann. frames | Acquisition info | ||||||
|---|---|---|---|---|---|---|---|---|---|
| Im.size [pix] | Im.res. [mm] | Im.rate [Hz] | Scanner | Probe | Center Freq. [MHz] | ||||
| ETH-01 | 0.71 | 14516 | 25 | 1 | 1453 | Siemens Antares | CH4-1 | 2.22 | |
| ETH-02 | 0.40 | 5244 | 16 | 1 | 525 | Siemens Antares | CH4-1 | 2.00 | |
| ETH-03 | 0.36 | 5578 | 17 | 3 | 559 | Siemens Antares | CH4-1 | 1.82 | |
| ETH-04 | 0.42 | 2620 | 15 | 1 | 263 | Siemens Antares | CH4-1 | 2.22 | |
| ETH-05 | 0.40 | 3652 | 15 | 2 | 366 | Siemens Antares | CH4-1 | 2.22 | |
| ETH-06 | 0.37 | 5586 | 17 | 2 | 560 | Siemens Antares | CH4-1 | 1.82 | |
| ETH-07 | 0.28 | 4588 | 14 | 1 | 460 | Siemens Antares | CH4-1 | 2.22 | |
| ETH-08 | 0.36 | 5574 | 17 | 2 | 558 | Siemens Antares | CH4-1 | 1.82 | |
| ETH-09 | 0.40 | 5247 | 16 | 2 | 525 | Siemens Antares | CH4-1 | 1.82 | |
| ETH-10 | 0.40 | 4587 | 15 | 4 | 460 | Siemens Antares | CH4-1 | 1.82 | |
| ETH-11 | 0.42 | 4615 | 15 | 2 | 463 | Siemens Antares | CH4-1 | 1.82 | |
| ETH-12 | 0.45 | 4284 | 14 | 2 | 429 | Siemens Antares | CH4-1 | 2.22 | |
| MED-01 | 0.41 | 2470 | 20 | 3 | 248 | DiPhAs Fraunhofer | VermonCLA | 5.5 | |
| MED-02 | 0.41 | 2478 | 20 | 3 | 248 | DiPhAs Fraunhofer | VermonCLA | 5.5 | |
| MED-03 | 0.41 | 2456 | 20 | 4 | 246 | DiPhAs Fraunhofer | VermonCLA | 5.5 | |
| MED-04 | 0.41 | 2455 | 20 | 3 | 246 | DiPhAs Fraunhofer | VermonCLA | 5.5 | |
| MED-05 | 0.41 | 2458 | 20 | 3 | 246 | DiPhAs Fraunhofer | VermonCLA | 5.5 | |
| MED-06 | 0.41 | 2443 | 20 | 3 | 245 | DiPhAs Fraunhofer | VermonCLA | 5.5 | |
| MED-07 | 0.41 | 2450 | 20 | 3 | 246 | DiPhAs Fraunhofer | VermonCLA | 5.5 | |
| MED-08 | 0.41 | 2442 | 20 | 2 | 245 | DiPhAs Fraunhofer | VermonCLA | 5.5 | |
| MED-09 | 0.41 | 2436 | 20 | 5 | 244 | DiPhAs Fraunhofer | VermonCLA | 5.5 | |
| MED-10 | 0.41 | 2427 | 20 | 4 | 243 | DiPhAs Fraunhofer | VermonCLA | 5.5 | |
| MED-11 | 0.41 | 2424 | 20 | 3 | 243 | DiPhAs Fraunhofer | VermonCLA | 5.5 | |
| MED-12 | 0.41 | 2450 | 20 | 3 | 246 | DiPhAs Fraunhofer | VermonCLA | 5.5 | |
| MED-13 | 0.35 | 3304 | 11 | 3 | 331 | Zonare z.one | C4-1 | 4.0 | |
| MED-14 | 0.35 | 3304 | 11 | 3 | 331 | Zonare z.one | C4-1 | 4.0 | |
| MED-15 | 0.35 | 3304 | 11 | 1 | 331 | Zonare z.one | C4-1 | 4.0 | |
| MED-16 | 0.35 | 3304 | 11 | 2 | 331 | Zonare z.one | C4-1 | 4.0 | |
| OX-01 | 0.30 | 71 | 12 | 1 | 71 | Zonare z.one | P4-1 | 3.6 | |
| OX-02 | 0.40 | 82 | 12 | 1 | 82 | Zonare z.one | P4-1 | 3.6 | |
| OX-03 | 0.38 | 82 | 12 | 1 | 82 | Zonare z.one | P4-1 | 3.4 | |
| OX-04 | 0.36 | 51 | 14.5 | 1 | 51 | Zonare z.one | C6-2 | 4.4 | |
| OX-05 | 0.46 | 101 | 11.7 | 1 | 101 | Zonare z.one | C6-2 | 3.8 | |
| OX-06 | 0.55 | 76 | 11 | 1 | 76 | Zonare z.one | C6-2 | 3.8 | |
| OX-07 | 0.50 | 63 | 10 | 2 | 63 | Zonare z.one | C6-2 | 3.8 | |
| OX-08 | 0.46 | 105 | 11 | 1 | 105 | Zonare z.one | C6-2 | 3.8 | |
| OX-09 | 0.50 | 98 | 10 | 2 | 98 | Zonare z.one | C6-2 | 3.8 | |
| OX-10 | 0.55 | 92 | 11 | 1 | 92 | Zonare z.one | C6-2 | 3.8 | |
| EMC-01 | 79 | 6 | 1 | 8 | iU22 | X6-1 | 3.2 | ||
| EMC-02 | 54 | 6 | 4 | 6 | iU22 | X6-1 | 3.2 | ||
| EMC-03 | 159 | 6 | 1 | 16 | iU22 | X6-1 | 3.2 | ||
| EMC-04 | 140 | 6 | 1 | 15 | iU22 | X6-1 | 3.2 | ||
| EMC-05 | 147 | 6 | 1 | 15 | iU22 | X6-1 | 3.2 | ||
| ICR-01 | 141 | 24 | 1 | 15 | Siemens SC2000 | 4Z1c | 2.8 | ||
| ICR-02 | 141 | 24 | 1 | 20 | Siemens SC2000 | 4Z1c | 2.8 | ||
| SMT-01 | 0.70 | 97 | 8 | 3 | 96 | GE E9 | 4V-D | 2.5 | |
| SMT-02 | 0.70 | 96 | 8 | 3 | 92-93 | GE E9 | 4V-D | 2.5 | |
| SMT-03 | 0.70 | 96 | 8 | 2 | 45-96 | GE E9 | 4V-D | 2.5 | |
| SMT-04 | 0.70 | 97 | 8 | 1 | 96 | GE E9 | 4V-D | 2.5 | |
| SMT-05 | 0.70 | 96 | 8 | 2 | 64-96 | GE E9 | 4V-D | 2.5 | |
| SMT-06 | 0.70 | 97 | 8 | 3 | 49-96 | GE E9 | 4V-D | 2.5 | |
| SMT-07 | 0.70 | 97 | 8 | 2 | 95 | GE E9 | 4V-D | 2.5 | |
| SMT-08 | 0.70 | 97 | 8 | 3 | 96 | GE E9 | 4V-D | 2.5 | |
| SMT-09 | 0.70 | 97 | 8 | 3 | 96 | GE E9 | 4V-D | 2.5 | |
Note: The test set is listed in black font. The training sequences, for which all available annotations were provided, are highlighted in red.
Figure 1.Examples of first frame I(0) of the training data: (top row) 2D sequences (ETH, MED, OX) and (bottom row) 3D sequences (EMC, SMT). Point-landmarks P(0) and the contour of the tumor segmentation S(0) are depicted in yellow.
Summary of the main features of the evaluated tracking methods: the tracking Task namely 2D and 3D point-landmark (2D p. and 3D p. respectively) and 2D tumor segmentation (2D s.); the key components of the tracking algorithm (Keywords).
| Participant | Task | Keywords | Real-time | ||
|---|---|---|---|---|---|
| 2D p. | 2D s. | 3D p. | |||
| KM | ✓ | ✗ | ✗ | block matching, NCC, local translation, exhaustive search | ✗ |
| MEVIS | ✓ | ✗ | ✗ | variational, large moving ROI, SSD & NGF, curvature regularizer | ✗ |
| MEVIS + FOKUS | ✓ | ✗ | ✓ | optical flow, histogram equalization, 30% downsampling, polynomial expansion, bilateral filtering, outlier detection, 2 orthogonal slices | ✓ |
| MEVIS + MED | ✓ | ✗ | ✓ | Bayesian approach, particle filter, intensity difference, local translation | ✓ |
| PhR | ✓ | ✓ | ✓ | sparse Demons, ROI, SSD, fluid regularizer, gradient descent, drift prevention strategy | ✓ |
| TUM | ✓ | ✗ | ✗ | kernel-based, intensity distribution similarity, adaptive ellipsoidal target descriptor, local affine, failure recovery strategy | ✓ |
Note: Real-time capability was assessed w.r.t. the average frame rate of 20 Hz (50 ms) for 2D sequences and 8 Hz (125 ms) for 3D sequences.
Figure 2.Tracking scheme. First W(t − 1, I (t)) is registered to W(0, I(0)) providing . If this registration fails, W(t − 1, I(t)) is registered to corresponding window of previous frame ().
Figure 3.Initialization: (Left) Within radius R0 of a given position, points on a local triangular grid with grid constant R1 are chosen. (Right) Example of point weights (, see text) in a first frame: area indicates value and color encodes sign (red: negative, green: positive).
Results of 2D point-landmark tracking.
| Method | Tracking error | Mean error range of sequences | ||||
|---|---|---|---|---|---|---|
| MTE2D | STD | 95thTE | MTE | MTE | MTE | |
| Fusion | 1.23 | 1.52 | 3.26 | [0.36, 2.04] | [0.81, 7.83] | [0.89, 2.97] |
| MEVIS | 1.44 | 2.04 | 3.86 | [0.31, 2.61] | [0.90, 8.75] | [1.02, 3.47] |
| MEVIS + MED | 1.53 | 2.45 | 3.95 | [0.32, 3.02] | [0.94, 5.12] | [1.20, 12.71] |
| TUM | 1.64 | 1.84 | 4.68 | [0.43, 7.48] | [0.56, 4.24] | [0.95, 2.83] |
| KM | 1.83 | 3.16 | 4.82 | [0.37, 1.73] | [0.93, 13.22] | [1.62, 3.63] |
| PhR | 2.00 | 2.87 | 5.59 | [0.51, 3.47] | [0.79, 12.72] | [0.88, 3.54] |
| MEVIS + FOKUS | 2.09 | 2.87 | 6.22 | [0.52, 10.05] | [0.59, 11.27] | [0.88, 3.36] |
| Motion | 6.64 | 4.81 | 15.53 | [2.90, 13.56] | [3.78, 12.48] | [4.33, 12.31] |
Note: The results are in millimeters and ranked (top to bottom) according to increasing MTE2D.
Figure 6.Percentage of failure cases: ratio of annotated 2D landmarks whose mm (orange) or mm (red) shown for all methods. Results are shown (left to right) according to decreasing MTE2D (see table 3). TE is evaluated with respect to one observer.
Figure 4.Illustration of tracking performance for landmark P1 from sequence MED-07. Tracking errors MTE1∈ were 13.22 (KM), 3.84 (MEVIS), 7.46 (MEVIS + FOKUS), 2.88 (MEVIS + MED), 12.72 (PhR) and 1.93 mm (TUM). The mean motion for the landmark was 11.23 mm. Frames at t, t and t correspond to end-inhalations (with a deep inhale happening at t), while t, t and t correspond to end-exhalations. In ROI(t) the manual annotation is shown as a yellow circle.
Summary of the results of point-landmark tracking w.r.t. mean manual annotation of three observers.
| Method | 2D tracking error | 3D tracking error | |||||||
|---|---|---|---|---|---|---|---|---|---|
| MTE2D | STD | MedianE | 95thTE | MTE | MTE3D | STD | MedianE | 95thTE | |
| Fusion | 1.08 | 1.42 | 0.75 | 2.85 | 1.32 | 2.43 | 2.76 | 1.49 | 7.61 |
| KM | 1.75 | 3.05 | 1.03 | 4.76 | ✗ | ✗ | ✗ | ✗ | ✗ |
| MEVIS | 1.33 | 1.94 | 0.88 | 3.56 | ✗ | ✗ | ✗ | ✗ | ✗ |
| MEVIS + FOKUS | 1.90 | 2.75 | 1.11 | 6.02 | 2.33 | 4.79 | 4.72 | 2.99 | 13.48 |
| MEVIS + MED | 1.45 | 2.48 | 0.95 | 3.49 | 1.78 | 2.76 | 4.10 | 1.52 | 8.80 |
| PhR | 1.94 | 2.93 | 1.12 | 5.53 | 2.38 | 2.83 | 2.97 | 1.46 | 9.67 |
| TUM | 1.41 | 1.89 | 0.78 | 4.70 | ✗ | ✗ | ✗ | ✗ | ✗ |
| Motion | 6.69 | 4.78 | 5.65 | 15.52 | 8.19 | 5.83 | 4.21 | 4.52 | 14.80 |
| Obs1 | 0.58 | 0.42 | 0.48 | 1.40 | 0.71 | 1.21 | 1.21 | 0.79 | 4.43 |
| Obs2 | 0.48 | 0.34 | 0.41 | 1.06 | 0.59 | 1.73 | 2.66 | 0.78 | 8.74 |
| Obs3 | 0.50 | 0.41 | 0.40 | 1.23 | 0.61 | 1.81 | 2.61 | 0.95 | 5.61 |
Note: The results are in millimeters and ranked according to alphabetical order of the methods.
Figure 5.Box-plot summarizing the 2D tracking error (in mm) w.r.t. mean manual annotation of three observers (Obs). Results are ranked (left to right) according to decreasing MTE2D (in green). On each box, the central red line is the median and the edges of the box are given by q1 = 25th and q3 = 75th percentiles of the error. Outliers are drawn as red crosses if larger than q3 + w(q3 − q1), where w = 1.5 corresponds to approximately ±2.7 STD of the data.
Results of 3D point-landmark tracking.
| Method | Tracking error | Mean error range of sequences | ||||
|---|---|---|---|---|---|---|
| MTE3D | STD | 95thTE | MTE | MTE | MTE | |
| Fusion | 2.48 | 2.46 | 6.91 | [1.19, 9.84] | [2.53, 2.59] | [0.94, 8.12] |
| PhR | 2.55 | 2.46 | 7.98 | [1.03, 9.63] | [2.54, 3.89] | [0.99, 11.57] |
| MEVIS + MED | 2.71 | 3.01 | 7.58 | [1.36, 10.40] | [1.59, 2.76] | [1.00, 6.59] |
| MEVIS + FOKUS | 4.63 | 4.03 | 12.44 | [2.41, 11.26] | [4.28, 5.88] | [1.23, 10.10] |
| Motion | 6.19 | 4.64 | 14.83 | [3.59, 13.16] | [4.47, 5.72] | [2.46, 12.89] |
Note: The results are in millimeters and ranked according to increasing MTE3D.
Figure 8.Percentage of failure cases: ratio of annotated 3D landmarks whose mm (orange) or mm (red) shown for all methods. Results are shown (left to right) according to decreasing MTE3D (see table 4). TE is evaluated with respect to one observer.
Figure 7.Box-plot summarizing the 3D tracking error (in mm) w.r.t. mean manual annotation of three observers (Obs). Results are ranked (left to right) according to decreasing MTE3D (in green). On each box, the central red line is the median and the edges of the box are given by q1 = 25th and q3 = 75th percentiles of the error. Outliers are drawn as red crosses if larger than q3 + w(q3 − q1), where w = 1.5 corresponds to approximately ±2.7 STD of the data.
Figure 9.Illustration of tracking performance for landmark P1 from sequence EMC-03. Tracking errors MTE1∈EMC−03 with respect to the mean of 3 observers were 7.77 mm (MEVIS + FOKUS), 5.63 mm (MEVIS + MED) and 9.93 mm (PhR). The mean motion for the landmark was 11.71 mm. Inter-observer errors were 4.06 mm (Obs1), 11.59 mm (Obs2) and 9.76 mm (Obs3). The tracking results and annotations are shown at time for the same ROI() with planes cut at the corresponding from each method.
Summary of the results of point-landmark tracking w.r.t. mean manual annotation of three observers.
| Method | 2D tracking error | 3D tracking error | |||||||
|---|---|---|---|---|---|---|---|---|---|
| MTE2D | STD | MedianE | 95thTE | MTE | MTE3D | STD | MedianE | 95thTE | |
| Fusion | 2.68 | 3.52 | 1.81 | 7.02 | 3.28 | 3.35 | 3.77 | 2.14 | 10.02 |
| KM | 4.33 | 7.54 | 2.41 | 11.86 | ✗ | ✗ | ✗ | ✗ | ✗ |
| MEVIS | 3.29 | 4.73 | 2.10 | 8.89 | ✗ | ✗ | ✗ | ✗ | ✗ |
| MEVIS + FOKUS | 4.72 | 7.00 | 2.75 | 15.85 | 5.78 | 6.34 | 6.13 | 3.94 | 16.78 |
| MEVIS + MED | 3.68 | 6.89 | 2.32 | 8.86 | 4.51 | 3.66 | 5.12 | 2.11 | 11.36 |
| PhR | 4.71 | 7.12 | 2.71 | 13.56 | 5.77 | 3.80 | 3.88 | 2.09 | 13.22 |
| TUM | 3.37 | 4.39 | 1.98 | 11.48 | ✗ | ✗ | ✗ | ✗ | ✗ |
| Motion | 16.91 | 12.49 | 14.13 | 40.39 | 20.71 | 8.00 | 5.82 | 6.03 | 21.00 |
| Obs1 | 1.45 | 1.09 | 1.12 | 3.66 | 1.78 | 1.71 | 1.56 | 1.14 | 4.56 |
| Obs2 | 1.19 | 0.87 | 0.99 | 2.70 | 1.46 | 1.60 | 1.55 | 1.11 | 5.11 |
| Obs3 | 1.25 | 1.05 | 0.98 | 3.11 | 1.53 | 1.82 | 1.56 | 1.36 | 4.69 |
Note: The results are in pixels/voxels and ranked according to alphabetical order of the methods.
Results of 2D tumor segmentation tracking (PhR) w.r.t. manual annotation of a clinical expert.
| Segmentation | Tracking overlap [%] | Initial Dice [%] | ||||
|---|---|---|---|---|---|---|
| Mean | STD | 5th TO | Mean | STD | 5th TO | |
| OX-1, | 86.6 | 5.4 | 78.4 | 49.9 | 25.3 | 20.6 |
| OX-2, | 85.5 | 4.8 | 76.6 | 73.7 | 10.7 | 57.1 |
| OX-4, | 92.3 | 1.9 | 90.0 | 74.9 | 17.6 | 46.3 |
| OX-5, | 79.8 | 6.6 | 68.4 | 87.3 | 5.5 | 77.3 |
| OX-6, | 76.7 | 9.2 | 58.6 | 72.6 | 12.7 | 51.9 |
| OX-7, | 89.6 | 4.3 | 78.8 | 80.5 | 6.2 | 71.5 |
| OX-7, | 77.2 | 4.8 | 70.0 | 58.3 | 15.5 | 33.2 |
| OX-8, | 88.7 | 2.7 | 84.9 | 92.1 | 3.5 | 84.9 |
| OX-9, | 92.2 | 2.2 | 88.8 | 91.8 | 5.0 | 81.9 |
| OX-9, | 77.6 | 5.7 | 66.7 | 80.2 | 8.2 | 65.7 |
| OX-10, | 81.8 | 3.4 | 76.7 | 75.5 | 6.8 | 63.2 |
| OX | 84.1 | 7.5 | 71.1 | 77.8 | 16.7 | 41.5 |
Note: The results are in % of the Dice coefficient.
Figure 10.Illustration of tracking performance for S1 from OX-6. The Dice coefficient ranged from 92.5 % (at t) to 53.6% (at t). ROI(t) and ROI(t) show the overlap of the manual (in yellow) and PhR (in light blue) segmentations.
Sparse Demons—Gradient Descent
| 1: |
| 2: |
| 3: |
| 4: |
| 5: |
| 6: |
| 7: |
| 8: |
| 9: |
| 10: |
| 11: |
| 12: |