| Literature DB >> 23781276 |
Turgay Ayer1, Qiushi Chen, Elizabeth S Burnside.
Abstract
Screening mammography is the most effective means for early detection of breast cancer. Although general rules for discriminating malignant and benign lesions exist, radiologists are unable to perfectly detect and classify all lesions as malignant and benign, for many reasons which include, but are not limited to, overlap of features that distinguish malignancy, difficulty in estimating disease risk, and variability in recommended management. When predictive variables are numerous and interact, ad hoc decision making strategies based on experience and memory may lead to systematic errors and variability in practice. The integration of computer models to help radiologists increase the accuracy of mammography examinations in diagnostic decision making has gained increasing attention in the last two decades. In this study, we provide an overview of one of the most commonly used models, artificial neural networks (ANNs), in mammography interpretation and diagnostic decision making and discuss important features in mammography interpretation. We conclude by discussing several common limitations of existing research on ANN-based detection and diagnostic models and provide possible future research directions.Entities:
Mesh:
Year: 2013 PMID: 23781276 PMCID: PMC3677609 DOI: 10.1155/2013/832509
Source DB: PubMed Journal: Comput Math Methods Med ISSN: 1748-670X Impact factor: 2.238
Summary of ANN studies in mammography interpretation and diagnostic decision making.
| Study | Type | ANN structure | Input | Dataset and training/testing strategy | Results and findings |
|---|---|---|---|---|---|
| Stafford et al. (1993) [ | CADe | A committee of four three-layer BP-ANNs | Pixel information | 167 mammograms with pathologies and 89 without pathologies. | Test on 20 out of 128 mammograms covering microcalcification size-range of |
|
| |||||
| Zhang et al. (1994) [ | CADe | The Shift-Invariant ANN (SI-ANN) | Pixel information | 168 ROIs from 34 digitized mammograms. | ROC index: AZ = 0.91 ± 0.02, |
|
| |||||
| Chan et al. (1995) [ | CADe | The Convolution Neural Network (CNN) | Pixel information | 52 mammograms |
|
|
| |||||
| Nagel et al. (1998) [ | CADe | SI-ANN | Features extracted from image | 196 TPs and 1,252 FPs. |
|
|
| |||||
| Wu et al. (1992) [ | CADe | BP-ANN | Pixel information | 56 positive, 56 negative, and 56 FP ROIs, respectively. |
|
|
| |||||
| Jiang et al. (1996) [ | CADx | BP-ANN | Computer-extracted morphological features | 40 malignant and 67 benign cases from 100 images. | Identified 100% malignant and 82% of the benign cases. |
|
| |||||
| Jiang et al. (1999) [ | CADx | BP-ANN | Computer-extracted morphological features | 46 malignant and 58 benign cases. |
|
|
| |||||
| Huo et al. (1998) [ | CADx | BP-ANN | Morphological features characterizing margin and density | 38 benign and 57 malignant cases from 65 patients. |
|
|
| |||||
| Kallergi (2004) [ | CADx | BP-ANN | Morphological and distributional descriptors | 50 benign and 50 malignant cases. | AZ = 0.98 ± 0.01, |
|
| |||||
| Chan et al. (1997) [ | CADx | BP-ANN | Texture features SGLD matrices | 41 malignant and 45 benign cases from 54 patients. | With best subset of features: |
|
| |||||
| Baker et al. (1995) [ | CADx | BP-ANN | BI-RADS lesion descriptors and medical history variables | 133 benign and 73 malignant cases. |
|
|
| |||||
| Lo et al. (1999) [ | CADx | BP-ANN | BI-RADS lesion descriptors, age, and history variables | 326 benign and 174 malignant cases. |
|
|
| |||||
| Ayer et al. (2010) [ | CADx | BP-ANN | Demographic, mammographic features, and BI-RADS categories | 510 malignant and 61,709 benign cases. | AZ = 0.965 (ANN) versus 0.939 (radiologists), |
|
| |||||
| Jesneck et al. (2007) [ | CADx | BP-ANN | Mammographic features, sonographic features, and history features | 296 malignant and 507 benign cases. | Training and validation set: |
|
| |||||
| Tourassi et al. (2003) [ | CADx | CSNN | BI-RADS features, age and history | Training set: 174 malignant and 326 benign cases. | On training set: AZ = 0.84 ± 0.02 |
|
| |||||
| Orr (2001) [ | CADx and risk estimation | BP-ANN | Age and radiographic features | 185 malignant and 1,103 benign cases. | AZ = 0.89 (surgeons) versus 0.86 (ANN), |
CADe: computer-aided detection, CADx: computer-aided diagnosis, ANN: artificial neural network, BP-ANN: back-propagation artificial neural network, FP: false positive, TP: true positive, ROI: region of interest, SGLD: spatial grey level dependence, PPV: positive prediction value, BI-RADS: the breast imaging reporting and data system, CSNN: constraint satisfaction neural network, and SI-ANN: shift-invariant artificial neural network.
Figure 1Inputs to the network are lesion descriptors and family history of the patient. Nodes at each layer are connected to the nodes at the succeeding layer by weighted arcs. Each hidden node in the first hidden layer performs a nonlinear weighted sum of all input values. The outputs of the last hidden layer are then similarly combined to the output layer. The single output value shows the probability of the lesion being malignant.
Figure 2Neurons are organized in a non-hierarchical structure in constraint satisfaction neural network (CSNN). Each neuron is assigned a value (activation level). These values represent the network state. Inputs to each neuron include both the external input and the activation levels of other neurons connected by the bidirectional symmetric weights. The activation levels are updated by passing the weighted sum of input values through a transfer function. The training is terminated when the network achieves a globally stable state with all constraints satisfied.
Advantages and disadvantages of ANNs.
| Advantage | Disadvantage |
|---|---|
| (i) Easy model building with less formal statistical knowledge required. | (i) Clinical interpretation of model parameters is difficult (black boxes). |
| (ii) Capable of capturing interactions between predictors. | (ii) Sharing an existing ANN model is difficult. |
| (iii) Capable of capturing nonlinearities between predictors and outcomes. | (iii) Prone to overfitting due to the complexity of model structure. |
| (iv) Users can apply multiple different training algorithms | (iv) Confidence intervals of the predicted risks are difficult to obtain. |
| (v) The model development is empirical. Few guidelines exist to determine the best network structures and training algorithms. |