| Literature DB >> 35873808 |
Youssef Beauferris1,2,3, Jonas Teuwen4,5,6, Dimitrios Karkalousos7, Nikita Moriakov4,5, Matthan Caan7, George Yiasemis5,6, Lívia Rodrigues8, Alexandre Lopes9, Helio Pedrini9, Letícia Rittner8, Maik Dannecker10, Viktor Studenyak10, Fabian Gröger10, Devendra Vyas10, Shahrooz Faghih-Roohi10, Amrit Kumar Jethi11, Jaya Chandra Raju11, Mohanasankar Sivaprakasam11,12, Mike Lasby1,3, Nikita Nogovitsyn13,14, Wallace Loos2,3,15,16, Richard Frayne2,3,15,16, Roberto Souza1,2,3.
Abstract
Deep-learning-based brain magnetic resonance imaging (MRI) reconstruction methods have the potential to accelerate the MRI acquisition process. Nevertheless, the scientific community lacks appropriate benchmarks to assess the MRI reconstruction quality of high-resolution brain images, and evaluate how these proposed algorithms will behave in the presence of small, but expected data distribution shifts. The multi-coil MRI (MC-MRI) reconstruction challenge provides a benchmark that aims at addressing these issues, using a large dataset of high-resolution, three-dimensional, T1-weighted MRI scans. The challenge has two primary goals: (1) to compare different MRI reconstruction models on this dataset and (2) to assess the generalizability of these models to data acquired with a different number of receiver coils. In this paper, we describe the challenge experimental design and summarize the results of a set of baseline and state-of-the-art brain MRI reconstruction models. We provide relevant comparative information on the current MRI reconstruction state-of-the-art and highlight the challenges of obtaining generalizable models that are required prior to broader clinical adoption. The MC-MRI benchmark data, evaluation code, and current challenge leaderboard are publicly available. They provide an objective performance assessment for future developments in the field of brain MRI reconstruction.Entities:
Keywords: benchmark; brain imaging; image reconstruction; inverse problems; machine learning; magnetic resonance imaging (MRI)
Year: 2022 PMID: 35873808 PMCID: PMC9298878 DOI: 10.3389/fnins.2022.919186
Source DB: PubMed Journal: Front Neurosci ISSN: 1662-453X Impact factor: 5.152
Summary of the raw MRI k-space datasets used in the first edition of the challenge.
|
|
|
|
|
|---|---|---|---|
| 12-channel | Train | 47 | 12,032 |
| Validation | 20 | 5,120 | |
| Test | 50 | 7, 800 | |
| 32-channel | Test | 50 | 7, 800 |
Reported are the number of slices in the test sets after removal of the initial 50 and last 50 slices (see text).
Summary of the submissions including processing domain, presence of coil sensitivity estimation (SE), presence of data consistency (DC), and basis of the training loss functions.
|
|
|
|
|
|
|---|---|---|---|---|
| ResoNNance 2.0 | Hybrid | Yes | Yes | MAE and SSIM |
| The Enchanted 2.0 | Hybrid | Yes | Yes | Cross entropy (pretext) and SSIM (main task) |
| ResoNNance 1.0 | Image | Yes | Yes | MAE and SSIM |
| The-Enchanted 1.0 | Image | Yes | Yes | MSE (first step) and SSIM (second step) |
| TUMRI | Hybrid | No | Yes | MS-SSIM and VIF |
| WW-Net | Hybrid | No | Yes | MSE |
| Hybrid-cascade | Hybrid | No | Yes | MSE |
| M-L UNICAMP | Hybrid | No | Yes | MSE |
| U-Net | Image | No | No | MSE |
| Zero-filled | N/A | No | N/A | N/A |
indicates a baseline model. Loss functions: Mean Absolute Error (MAE), Structural Similarity (SSIM), Mean Squared Error (MSE), Multi-Scale SSIM (MS-SSIM), and Visual Information Fidelity (VIF).
Summary of the Track 01 results for R = 5.
|
|
|
|
|
|---|---|---|---|
| ResoNNance 2.0 |
|
| 0.957 ± 0.034 |
| The Enchanted 2.0 | 0.937 ± 0.033 | 34.9 ± 2.4 | 0.973 ± 0.036 |
| ResoNNance 1.0 | 0.936 ± 0.031 | 35.3 ± 1.8 | 0.960 ± 0.035 |
| The-Enchanted 1.0 | 0.912 ± 0.034 | 30.3 ± 2.8 |
|
| TUMRI | 0.868 ± 0.044 | 32.5 ± 1.7 | 0.989 ± 0.045 |
| WW-net | 0.870 ± 0.043 | 32.5 ± 1.7 | 0.929 ± 0.049 |
| Hybrid-cascade | 0.860 ± 0.044 | 32.7 ± 1.6 | 0.954 ± 0.042 |
| M-L UNICAMP | 0.868 ± 0.044 | 32.4 ± 1.7 | 0.918 ± 0.053 |
| U-Net | 0.779 ± 0.039 | 26.8 ± 1.7 | 0.642 ± 0.068 |
| Zero-filled | 0.726 ± 0.045 | 25.2 ± 1.5 | 0.518 ± 0.066 |
The best value for each metric and acceleration is emboldened. Mean ± standard deviation are reported.
indicates a baseline model.
Figure 1Representative reconstructions of the different models submitted to Track 01 (i.e., 12-channel) of the challenge for R = 5. Note that the reconstructions from the top four methods, ResoNNance 1.0 and 2.0, and The Enchanted 1.0 and 2.0, try to match the noise pattern seen in the background of the reference image, while ML-UNICAMP, Hybrid-cascade, WW-net, and TUMRI seem to have partially filtered this background noise.
Figure 2Quality assessment comparing the fully sampled reference and the reconstruction obtained by team ResoNNance 2.0. (A) The top row shows the border of the left putamen, where the reconstructed image has a discrepancy in shape compared to the reference image (highlighted with red circles). The bottom row shows that changes in the shape of the structure are also visible in the next slice of the same subject (highlighted with red arrows). It is important to emphasize that these discrepancies are not restricted to the putamen, but a systematic evaluation of where these changes occur is out of scope for this work. (B) Illustration of a case where the expert observed rated that the deep-learning-based reconstruction improved image quality. In this figure, we can see smoothening of cortical white matter without loss of information as no changes appeared in the pattern of gyrification within cortical gray matter.
Summary of the Track 02 results for R = 5 using the 32-channel test set.
|
|
|
|
|
|---|---|---|---|
| ResoNNance 2.0 |
| 38.3 ± 2.2 | 0.955 ± 0.036 |
| The Enchanted 2.0 | 0.960 ± 0.037 |
|
|
| ResoNNance 1.0 | 0.947 ± 0.033 | 37.7 ± 2.9 | 0.992 ± 0.030 |
| The Enchanted 1.0 | 0.907 ± 0.046 | 30.1 ± 2.7 | 0.834 ± 0.236 |
| U-Net | 0.832 ± 0.058 | 31.5 ± 2.6 | 0.804 ± 0.045 |
| Zero-filled | 0.780 ± 0.041 | 26.4 ± 1.5 | 0.472 ± 0.064 |
The best value for each metric and acceleration is emboldened. Mean ± standard deviation are reported.
indicates a baseline model.
Figure 3Representative reconstructions of the different models submitted to Track 02 of the challenge for R = 5 using the 32-channel coil.
Figure 4Sample reconstruction illustrating artifacts (highlighted in red boxes) that seem to be present on images reconstructed by models that used coil sensitivity estimation as part of their method.
Figure 5Three sample reconstructions, one per row, for the top two models. The Enchanted 2.0 and ResoNNance 2.0 and the reference are illustrated. The arrows in the figure indicate regions of interest that indicate deviations between the deep-learning-based reconstructions and the fully sampled reference.