| Literature DB >> 36152237 |
Mariachiara Di Cosmo1, Maria Chiara Fiorentino2, Francesca Pia Villani3, Emanuele Frontoni4, Gianluca Smerilli5, Emilio Filippucci5, Sara Moccia6.
Abstract
Ultrasound (US) imaging is recognized as a useful support for Carpal Tunnel Syndrome (CTS) assessment through the evaluation of median nerve morphology. However, US is still far to be systematically adopted to evaluate this common entrapment neuropathy, due to US intrinsic challenges, such as its operator dependency and the lack of standard protocols. To support sonographers, the present study proposes a fully-automatic deep learning approach to median nerve segmentation from US images. We collected and annotated a dataset of 246 images acquired in clinical practice involving 103 rheumatic patients, regardless of anatomical variants (bifid nerve, closed vessels). We developed a Mask R-CNN with two additional transposed layers at segmentation head to accurately segment the median nerve directly on transverse US images. We calculated the cross-sectional area (CSA) of the predicted median nerve. Proposed model achieved good performances both in median nerve detection and segmentation: Precision (Prec), Recall (Rec), Mean Average Precision (mAP) and Dice Similarity Coefficient (DSC) values are 0.916 ± 0.245, 0.938 ± 0.233, 0.936 ± 0.235 and 0.868 ± 0.201, respectively. The CSA values measured on true positive predictions were comparable with the sonographer manual measurements with a mean absolute error (MAE) of 0.918 mm2. Experimental results showed the potential of proposed model, which identified and segmented the median nerve section in normal anatomy images, while still struggling when dealing with infrequent anatomical variants. Future research will expand the dataset including a wider spectrum of normal anatomy and pathology to support sonographers in daily practice.Entities:
Keywords: Carpal tunnel syndrome; Deep learning; Median nerve; Segmentation; Ultrasound imaging
Mesh:
Year: 2022 PMID: 36152237 PMCID: PMC9537213 DOI: 10.1007/s11517-022-02662-5
Source DB: PubMed Journal: Med Biol Eng Comput ISSN: 0140-0118 Impact factor: 3.079
Fig. 1US transverse scan sample acquired at the proximal carpal tunnel inlet. A red box includes the median nerve section; asterisks of different colors mark other relevant structures: pisiform bone profile in blue, semilunar bone profile in purple, ulnar artery in green, digital flexor tendons in orange
Overview of the US dataset characteristics in DL literature for median nerve segmentation, in terms of US acquisition site, dataset size (frames selection or frame sequences, total number of images) and patients involved in the study
| Acquisition site | Frame sequences | N. of US images | N. of patients | |
|---|---|---|---|---|
| Kakade and Dumbali [ | Brachial plexus forearm | No | 11508 | - |
| Wang et al. [ | Carpal tunnel | Yes (100) | 18000 | 50 |
| Horng et al. [ | Carpal tunnel | Yes (24) | 10080 | 6 |
| Festen et al. [ | Proximal carpal tunnel inlet | Yes (505) | 5560 | 99 |
| Wu et al. [ | Proximal carpal tunnel inlet | Yes (36) | 18625 | 36 |
| Proposed model | Proximal carpal tunnel inlet | No | 246 | 103 |
Fig. 2Schematic representation of model architecture, composed by a backbone, Region Proposal Network (RPN), and the three heads for classification, bounding box regression and segmentation, all fed from the ROIAlign with 100 ROI candidates. The segmentation head is represented more in details as it was provided with two additional transposed layers compared with original Mask-RCNN [13]
Performance evaluation metrics in terms of mean value and standard deviation. Mean average precision (mAP), Recall (Rec), Precision (Prec) and Dice Similarity Coefficient (DSC) are reported for the proposed model and the ablation studies conducted over it: Mask-R50 is the model trained using as backbone Resnet50 combined with FPN; NoAug is the model trained using no augmentations on the training data; Mask28 and Mask56 are variants of the model with a different output resolution from the segmentation head, including one and two transposed convolutional layers, respectively
| Mask-R50 | 0.889 ± 0.277 | 0.888 ± 0.271 | 0.862 ± 0.261 | 0.843 ± 0.208 |
| NoAug | 0.891 ± 0.241 | 0.902 ± 0.294 | 0.870 ± 0.308 | 0.838 ± 0.247 |
| Mask28 | 0.908 ± 0.364 | 0.923 ± 0.254 | 0.877 ± 0.285 | 0.821 ± 0.261 |
| Mask56 | 0.926 ± 0.235 | 0.895 ± 0.284 | 0.899 ± 0.270 | 0.843 ± 0.219 |
| Proposed Model | 0.936 ± 0.235 | 0.938 ± 0.233 | 0.916 ± 0.245 | 0.868 ± 0.201 |
Comparison of segmentation performance in terms of DSC of the proposed model and of the U-Net and Lightweight U-Net trained using two different losses, i.e., the BCE loss and the BCE − DSC loss
| DSC | |
|---|---|
| U-NET ( | 0.783 ± 0.229 |
| U-NET ( | 0.822 ± 0.205 |
| Lightweight U-NET ( | 0.780 ± 0.195 |
| Lightweight U-NET ( | 0.764 ± 0.216 |
| Proposed Model | 0.868 ± 0.201 |
Fig. 3Four visual samples of the median nerve section. From top to bottom row: original US image, ground truth mask, U-Net trained with BCE loss prediction, U-Net trained with BCE − DSC loss prediction, Lightweight U-Net trained with BCE loss prediction, Lightweight U-Net trained with BCE − DSC loss prediction, proposed model prediction. For displaying purpose, only the upper part of the US images, which contains the median nerve section, is shown