Literature DB >> 30444907

Automatic inference model construction for computer-aided diagnosis of lung nodule: Explanation adequacy, inference accuracy, and experts' knowledge.

Masami Kawagishi1, Takeshi Kubo2, Ryo Sakamoto2,3, Masahiro Yakami2,3, Koji Fujimoto4, Gakuto Aoyama1, Yutaka Emoto5, Hiroyuki Sekiguchi2, Koji Sakai6, Yoshio Iizuka1, Mizuho Nishio2,3, Hiroyuki Yamamoto1, Kaori Togashi2.   

Abstract

We aimed to describe the development of an inference model for computer-aided diagnosis of lung nodules that could provide valid reasoning for any inferences, thereby improving the interpretability and performance of the system. An automatic construction method was used that considered explanation adequacy and inference accuracy. In addition, we evaluated the usefulness of prior experts' (radiologists') knowledge while constructing the models. In total, 179 patients with lung nodules were included and divided into 79 and 100 cases for training and test data, respectively. F-measure and accuracy were used to assess explanation adequacy and inference accuracy, respectively. For F-measure, reasons were defined as proper subsets of Evidence that had a strong influence on the inference result. The inference models were automatically constructed using the Bayesian network and Markov chain Monte Carlo methods, selecting only those models that met the predefined criteria. During model constructions, we examined the effect of including radiologist's knowledge in the initial Bayesian network models. Performance of the best models in terms of F-measure, accuracy, and evaluation metric were as follows: 0.411, 72.0%, and 0.566, respectively, with prior knowledge, and 0.274, 65.0%, and 0.462, respectively, without prior knowledge. The best models with prior knowledge were then subjectively and independently evaluated by two radiologists using a 5-point scale, with 5, 3, and 1 representing beneficial, appropriate, and detrimental, respectively. The average scores by the two radiologists were 3.97 and 3.76 for the test data, indicating that the proposed computer-aided diagnosis system was acceptable to them. In conclusion, the proposed method incorporating radiologists' knowledge could help in eliminating radiologists' distrust of computer-aided diagnosis and improving its performance.

Entities:  

Mesh:

Year:  2018        PMID: 30444907      PMCID: PMC6239329          DOI: 10.1371/journal.pone.0207661

Source DB:  PubMed          Journal:  PLoS One        ISSN: 1932-6203            Impact factor:   3.240


Introduction

Advances in imaging modalities have made it possible to acquire large amounts of medical image data for radiologists to assess, increasing their workload. Computer-aided diagnosis (CAD) has been developed to decrease the workload and categorized into two types [1]: computer-aided detection (CADe), which supports lesion detection [2-4], and computer-aided diagnosis (CADx), which supports differential diagnosis [5-7]. CAD systems, particularly CADx systems, employ an inference model to present suggestions to radiologists based on the data input (e.g., imaging findings). Several studies have reported the usefulness of CAD for lung nodules [8-12]. Shiraishi et al. proposed a system that calculated the possibility of the presence of a malignant lung nodule from two clinical parameters and 75 imaging features, using linear distinct analysis in chest radiographs [8]; they showed that radiologists’ performance significantly improved with the use of CADx. In addition, Chen et al. proposed a CADx system that estimated nodule type based on 15 image features, using an ensemble model of artificial neural network with chest computed tomography (CT) [9]. Notably, their system showed performance comparable to that of senior radiologists while classifying the nodule type. Nevertheless, CADx systems are rarely used in clinical practices, possibly because radiologists distrust the suggestions of the CAD system because they do not provide explanations for the decisions. Supporting this, Kawamoto et al. suggested that CAD should at least provide adequate details about the reasoning for any inference results [13]. Accordingly, we believe that the barriers to its use could disappear, or at least diminish, if the CAD system could provide justifications for its suggestions. Remarkably, few systems offer reasons behind their suggestions. Green et al. proposed such a system based on sensitivity analysis with electrocardiogram interpretation [14], whereas Kawagishi et al. proposed a system that disclosed the reasoning based on the influence on the inference result in chest CT [15]. Although their inference models showed high accuracy of the inferences and high adequacy of the reasons, the reports did not describe the model construction. However, it can be challenging to manually construct an inference model with high accuracy and high adequacy because the number of possible models is vast. Although various automatic construction methods have been proposed [16-19], they have only considered inference accuracy as a performance metric and not explanation adequacy or subjective interpretability. In our study, we have proposed a method for automatically constructing inference models using a metric that considers explanation adequacy and inference accuracy. Moreover, we have evaluated the usefulness of radiologists’ knowledge while constructing these models.

Materials and methods

This retrospective study was approved by the Ethics Committee of Kyoto University Hospital (Kyoto, Japan), which waived the requirement of informed consent. The notations used in this paper are shown in Table 1.
Table 1

List of notations.

NotationDescriptionExample
D“Diagnosis” as the inference target node (random variable)NA
distate of random variable Dd1, primary lung cancer
Xjimaging findings and clinical data as the other nodes (random variable)shape, tumor marker
xjkstate of random variable Xjx31, irregular
EEvidence, as a set of xjk{x11, x21}
p(di|E)posterior probability of di when E is given to the inference modelNA
dfinference diagnosis with the highest posterior probability among p(di|E)d1, primary lung cancer
Rcreason candidate (a proper subset of E)If x11 and x21 are given as E, then Rc can be {x11, x21}, {x11}, {x21}.
|Rc|the number of elements of RcIf Rc is {x11, x21}, then |Rc| is 2.
RctRc with only one element{x11}
I(Rc)influence of Rc on the inference diagnosis dfNA
p(df)prior probability of the inference diagnosis dfNA
pd(Rc)difference between p(df|Rc) and p(df) for the inference diagnosis dfNA
V(S)the performance metric of inference model SNA
Vr(S)explanation (reasoning) adequacy of inference model SNA
Vi(S)inference accuracy of inference model SNA
RgReference reasons (1–7 imaging findings and/or clinical data chosen by radiologists)“shape is polygon,” “diameter is small and cavitation exists,” and “satellite lesion exists”
RdReasons derived by the inference system“shape is polygon” and “diameter is small and cavitation exists”

Abbreviation: NA, not available

Abbreviation: NA, not available

Dataset

We used thin-slice chest CT images and clinical information of 179 patients treated at Kyoto University Hospital. Each case had 1–5 pulmonary nodules, ranging in size from 10 to 30 mm, with the clinical diagnosis confirmed pathologically, clinically, or radiologically as primary lung cancer, lung metastasis, or benign lung nodule. Of note, we used 79 cases as training data for constructing the inference model and the remaining 100 cases as test data for evaluating the performance of the model. Without the knowledge of the clinical diagnosis, two radiologists (A and B) analyzed a representative nodule for each case and recorded 49 types of imaging findings as ordinal or nominal data. In addition, 37 clinical data types, including laboratory data and patient history of malignancies, were collected from patients’ electronic medical records; these data were used as the input information for the inference models (see Supporting information, S1 File). The clinical diagnosis data were used as reference for evaluating inference accuracy; based on these diagnoses, two other radiologists (C and D) selected a set of 1–7 imaging findings and/or clinical data as the reference explanations for diagnosis (e.g., “shape is polygon,” “diameter is small and cavitation exists,” and “satellite lesion exists”).

Inference model

The inference computational model infers diagnosis (primary lung cancer, lung metastasis, or benign lung nodule in our study) from the input image findings and clinical data. As the inference model, we employed a Bayesian network, a directed acyclic graphical model that includes nodes and directed links. Fig 1 provides an example of a Bayesian network (directed acyclic graphical model). Each node represents a random variable, and each directed link represents relationship between variables.
Fig 1

An example of a Bayesian network (directed acyclic graphical model).

The Bayesian network has nodes (circles) and directed links (arrows). Each node and directed link represent a random variable and relationship, respectively. Each node can have a discriminate value (state).

An example of a Bayesian network (directed acyclic graphical model).

The Bayesian network has nodes (circles) and directed links (arrows). Each node and directed link represent a random variable and relationship, respectively. Each node can have a discriminate value (state). In Fig 1, D denotes diagnosis as the inference target node, and X (j = 1, 2, …, N) denotes the imaging findings (e.g., shape) and clinical data (e.g., tumor marker) as the other nodes. Each node can have a discriminate value of d or x; for example, D takes d (i = 1, 2, 3; e.g., d1: primary lung cancer) and X takes x (e.g., for shape, k = 1, 2, …, 8 and j = 3, with x = irregular). Further, maximum value of k ranges from two to eight depending on X. Next, E denotes Evidence [20] as a set of x used for information input in the inference models. The posterior probability of d is denoted by p(d|E) when E is set in the inference model, and d indicates the inference diagnosis with the highest posterior probability among p(d|E). The inference result can be calculated based on the Bayesian network structure using a probability propagation algorithm [20]. With a change in the Bayesian network structure (graphical model), the probability propagation path is also changed, indicating that the structure of the graphical model affects the inference result. We obtained the prior probability distributions for each node from the training data and calculated the conditional probabilities for each node based on links.

Reason derivation

Herein, we illustrate how the reasons are derived from Evidence (E) to justify the inference results. First, the notation for deriving reasons and the examples of notation usage are explained. E is given as a set of x, and R (reason candidate) is defined as a proper subset of E that can be selected as a reason, e.g., when the graphical model comprises D (diagnosis), X (nodule size), and X (cavitation) as nodes, if x11 (diameter is small) and x21 (cavitation exists) are specified as E, then R comprises only these two elements, and {{x11, x21}, {x11}, {x21}} represent all possible values of R. This notation allows the reasons to be derived from E. The influence, I(R), is defined as a quantitative measure to select R based on the graphical model. Its calculation is summarized in S2 File; to summarize, it represents the influence of R on the inference result (d): I(R) > 0 indicates a positive influence, whereas I(R) < 0 indicates a negative influence. I(R) is defined by the following equations: In Eqs 1 and 2, p(d) denotes the prior probability of d, and |R| denotes the number of elements of R. As stated in the section detailing the inference model, p(d) is calculated from the training data. Based on these equations, when R comprises only one element (i.e., |R| = 1), I(R) equals p(R) and is simply defined as the difference between p(d|R) and p(d). For |R| > 1, I(R) is calculated from p(R) and an additional penalty term, f(R), introduced to consider possible synergy among the multiple elements. To explain the synergetic effect on I(R), we use the notation R (t = 1, 2, …) for the subset of R, with only one element of R, e.g., when all possible values of R are {{x11, x21}, {x11}, {x21}}, then R = {x11} and R = {x21}. Note that R is also the reason candidate in this example (|R| = 1). If p(R = {x11}) and p(R = {x21}) are comparatively higher than p(R = {x11, x21}), then f(R = {x11, x21}) is also high, and we regard that {x11} and {x21} are more adequate than {x11, x21} as the reasons (e.g., “diameter is small” is more adequate than the combination of “diameter is small” AND “cavitation exists”). By contrast, if p(R = {x11}) and p(R = {x21}) are comparatively lower than p(R = {x11, x21}), then f(R = {x11, x21}) is also low, and the combination of elements {x11, x21} is regarded as more adequate than {x11} and {x21} (e.g., the combination of “diameter is small” AND “cavitation exists” is more adequate than “cavitation exists”). f(R) is defined as follows by calculating an element-wise total positive effect (f) and a total negative effect (f): In Eq 3, sgn(∙) denotes a sign function, and f − f can be considered a net effect of the non-synergetic influence of each element in R. f(R), as a synergetic influence, is set to zero when the sign of the element-wise influence f—f is different from p(R). That is to say, f(R) can work as penalty term when the sign of f—f is equal to that of p(R). When the value of the element-wise influence is larger than p(R), the synergetic influence is considered negligible, and f(R) is set to p(R), providing an I(R) of zero. L2 regularization is frequently used as penalty term in machine learning algorithm (i.e., support vector machine [21]). The difference between L2 regularization and Eqs (4)–(5) is the separation based on the sign. Therefore, it is expected that effect of our penalty term is similar to that of L2 regularization. Based on Eqs 1 and 3, I(R) can then be rewritten as follows for |R| ≥ 2: The maximum number of elements in R (|R|) is set to two to reduce the computational complexity. Further, I(R) is calculated for all possible candidates of R with |R| = 1 or 2. At most, the best three reason candidates are selected as appropriate reasons for each model. If I(R) is <0.05 * p(d), the reason is rejected.

Effect of model structure on deriving reasons

The structure of the graphical model, comprising nodes and directed links, affects both the inference result and reason derivation for the Bayesian network. Fig 2 shows an example of the probability propagation for two different structures of the graphical models. These models have three nodes (diagnosis node, D, and two other nodes, X and X). Model A has two directed links, from X to X and from D to X. Similarly, model B has two directed links, from X to X and from X to D. The direction of the link between X and D is different between the two models. The results are different when Evidence is given to X in the networks; in model A, propagation occurs from X to X but not from X to D, whereas in model B, propagation occurs in both directions. Thus, because X does not influence D in model A, it is not selected as a reason. In this way, the model structure influences the probability propagation (inference result) and reasons.
Fig 2

An example of probability propagation.

Curved arrows represent the propagation direction, dotted curved arrow with an X indicates no propagation, and gray circle (X) represents a node where Evidence is given. (a) Model A: Propagation does not occur from X to D. (b) Model B: Propagation occurs from X to D.

An example of probability propagation.

Curved arrows represent the propagation direction, dotted curved arrow with an X indicates no propagation, and gray circle (X) represents a node where Evidence is given. (a) Model A: Propagation does not occur from X to D. (b) Model B: Propagation occurs from X to D.

Metric

To automatically construct the inference model, its performance has to be calculated. Moreover, because we regard the reasons as important for evaluating the inference results, we require the performance measure to reflect explanation adequacy and inference accuracy. Several studies have suggested a trade-off between explanation adequacy and inference accuracy [22, 23]. Based on the consensus of the radiologists in our study, we used the following metric to evaluate the inference models: Herein, S denotes an inference model, V(S) denotes the performance metric of S, V(S) denotes the explanation adequacy of S, and V(S) denotes the inference accuracy of S. The values of V(S), V(S), and V(S) range from zero to one. Remarkably, this metric considers the explanation adequacy and inference accuracy of S. We also employ F-measure, which is commonly used in information retrieval, as V(S), providing a harmonic mean of precision and recall (completeness). The relationships among F-measure, precision, and recall are as follows: Here, R denotes a reference set of standards for a case (i.e., 1–7 imaging findings and/or clinical data selected by the two radiologists), R denotes a set of reasons derived by the inference system, and |∙| denotes the number of elements of a set. R ∩ R denotes the intersection between R and R. For this, R is obtained from R with I(R) up to three R. If there are >3 possible R, then R is obtained as follows: where represents R with the highest, second highest, and third highest values of I(R), respectively, and “∪” denotes an operator of union. The accuracy of the inference model, V(S), is defined as m / n, where m denotes the number of cases correctly inferred by the models, and n denotes the total number of cases.

Automatic model construction

The number of possible Bayesian network structures dramatically increases as the number of nodes increases; from these, structures with high performance must be effectively searched. Therefore, we use the Markov chain Monte Carlo (MCMC) method [24] to construct the model, S, and iteratively find the most appropriate model, i.e., with the maximum value of V(S). We use the metric and MCMC method to automatically construct the Bayesian model as follows: Set an initial model to the current model (Scurrent), and initialize the iteration count (M = 1). Create a temporary model (Stemp) by updating Scurrent. The update action is probabilistically selected as one of the following, with a probability based on the Scurrent structure: (1) deleting a link, (2) reversing a link, or (3) creating a new link (see Fig 3). If the action is not appropriate (e.g., Stemp has a cyclic loop in its structure), Step 2 is iterated.
Fig 3

Three types of update to the graphical model.

Delete denotes unlinking an existing link, reverse denotes reversing an existing link, and join denotes creating a new link.

Calculate V(Stemp) with 5-fold cross validation of the training data. Probabilistically replace Scurrent with Stemp with the following probability (P): where β represents the damping ratio (0 < β < 1). Note that Pm is small (difficult to replace) when V(Scurrent) > V(Stemp) or when M is large. If M reaches the iteration limit (M) or Scurrent has not been replaced M times, then Scurrent is output as the final model. If not, M = M + 1 is set, and the process returns to Step 2.

Three types of update to the graphical model.

Delete denotes unlinking an existing link, reverse denotes reversing an existing link, and join denotes creating a new link. In Step 2, Stemp is created with a probability based on the current model Scurrent, enabling setting a different Stemp at another trial even while using the same Scurrent. In this process, we set the core values as follows: β = 0.999, M = 10000, and M = 2500. If the inference accuracy V(S) of the final model is <0.70 for the training data, the model is discarded because the low inference accuracy is expected to negatively influence the model’s acceptability by the radiologists. For the same reason, if V(S) is <0.70, we set V(S) to V(S) and V(S) to 0 in Step 3, which eliminates the time-consuming calculation of V(S). The number of parent nodes is limited to no more than two because of limited computational resources.

Initial model with and without prior knowledge of radiologists

The final model depends on the initial model and metric V(S). To evaluate the effect of the initial model on the performance of the final model, we examined initial models with and without the radiologists’ expert knowledge. The radiologists’ knowledge is represented as links between the diagnostic node and other nodes in the initial model. When no prior knowledge is included, no link is present in the initial model. We conducted multiple trials of model construction with the same initial model because each trial could experience different paths, as already described.

Subjective evaluation of inference model

The two radiologists (A and B, who did not set the reference standards) were asked to subjectively evaluate the model with the best performance. Based on the clinical diagnosis, inference result, and derived reasons, a subjective rank was assigned to each case on a 5-point scale, wherein ranks 5, 3, and 1 represented beneficial, appropriate, and detrimental, respectively.

Results

Finally, 13 models with prior knowledge and five without prior knowledge were constructed after 37 trials. The remaining 19 models were discarded because they did not meet our predefined criteria. Table 2 shows the performance of the best three models with and without prior knowledge. S1 and S2 Tables show the performance of the other 10 and 2 models with and without prior knowledge, respectively. Among the 13 models with prior knowledge, the performance of the best model with the test data was as follows: F-measure (V) = 0.411, accuracy (V) = 72.0%, and metric (V) = 0.566. Among the five models without prior knowledge, the performance of the best model with the test data was as follows: F-measure (V) = 0.274, accuracy (V) = 65.0%, metric (V) = 0.462.
Table 2

Performance of the best three inference models with and without prior knowledge.

Training dataTest data
Prior knowledgeModelF-measure (Vr)Accuracy (Vi) (%)Metric (V)F-measure (Vr)Accuracy (Vi) (%)Metric (V)
withBest0.39975.90.5790.41172.00.566
2nd0.32470.90.5160.32576.00.542
3rd0.36370.90.5360.32874.00.534
withoutBest0.34272.20.5320.27465.00.462
2nd0.31474.70.5300.22263.00.426
3rd0.36177.20.5660.25060.00.425
According to Table 2, although the accuracy of three models without prior knowledge was comparable to that of three models with prior knowledge when applied to the training data, their performance (F-measure, accuracy, and metric) without prior knowledge was worse than that with prior knowledge when using the test data. Iteration numbers for the MCMC method in the three best models with prior knowledge were 2934, 2948, and 3126, while the corresponding numbers in those without knowledge were 2873, 5567, and 8642. Based on Table 2, we selected the best model constructed with prior knowledge (metric = 0.566) for the subjective evaluation. The average subjective ranks obtained from the two radiologists were 3.97 and 3.76. Fig 4 shows the frequencies of ranks recorded by the two radiologists, indicating that the mode of the ranks for each radiologist was 5. Rank 1 had the lowest frequency for Radiologist A, whereas rank 3 was less frequent than rank 1 as per Radiologist B. Fig 5 illustrates an example of misclassification by the inference system, in a case where a benign lung nodule was classified as a metastasis, and the three reasons for this were “shape is round,” “contour is smooth,” and “patient was diagnosed with malignancy during the past five years.” Both radiologists gave this a rank of 1.
Fig 4

Frequencies of subjective ranks recoded by two radiologists.

Note: Ranks 5, 3, and 1 in the 5-point scale represent beneficial, appropriate, and detrimental, respectively.

Fig 5

An example of misclassification and inadequate reasoning by the inference system.

A benign lung nodule (arrow) was classified as metastasis.

Frequencies of subjective ranks recoded by two radiologists.

Note: Ranks 5, 3, and 1 in the 5-point scale represent beneficial, appropriate, and detrimental, respectively.

An example of misclassification and inadequate reasoning by the inference system.

A benign lung nodule (arrow) was classified as metastasis. To compare our Bayesian-network-based method, inference and reasoning of lung nodules were performed using gradient tree boosting (xgboost) [25,26]. Please refer to the Supporting information (S3 File) for the comparison.

Discussion

We proposed a method for the automatic construction of a CADx system that could provide justification for its inference results. By incorporating radiologists’ knowledge into the model construction, we found that explanation adequacy and inference accuracy improved. To the best of our knowledge, few studies in radiology have sought to develop a CADx system that can provide valid reasoning for its inference results. Indeed, Langlotz et al. suggested that physicians not only based their trust in an inference system on good prediction performance but also on whether they understood the reasoning behind the predictions [27]. However, although Green et al. proposed an inference system for electrocardiogram that presented its reasoning [14], the same has not been proposed for radiology. Accordingly, our CADx system for lung nodules was capable of providing valid reasons for its inferences, with high explanation adequacy and high inference accuracy. As shown in Table 2, the models with prior knowledge from radiologists were more robust and superior to those without prior knowledge. Because the accuracies of the models without prior knowledge were comparable to those with prior knowledge in the training data, we speculate that the models without prior knowledge overfitted the training data. In effect, the radiologists’ knowledge prevented overfitting and improved the generalizability of the system. Consistent with this, a previous study showed that the Bayesian network performance in assessing mammograms was improved by incorporating experts’ knowledge [28]. The present study gained other benefits from expert involvement, e.g., the iteration numbers for the MCMC method were smaller in the models with prior knowledge than in those without prior knowledge. In addition to improving the model robustness, prior knowledge boosted the convergence speed of our inference models. The two radiologists (A and B) subjectively evaluated the model we constructed, giving an average rank more than 3. A large controlled study in the non-medical domain [29] has shown that providing reasoning and trace explanations for context-aware applications could improve user understanding and trust in the system. In line with this finding and the acceptability of our method to the radiologists participating in the present study, we expect that our CADx system could, at least, diminish an important barrier to the uptake of CADx systems. Several methods for automatic model construction were proposed in previous studies. These methods can be divided into two types [20,30]: (i) constraint-based methods [16,20,30,31] and (ii) search-and-score methods [19,20,30,32]. In constraint-based methods, relationship between nodes, such as conditional independency [31] or mutual information [16], are used to construct Bayesian network structure automatically. That is to say, if conditional independency is indicated or values of mutual information meet predefined criteria, existence of the links between the nodes are judged. For example, PC algorithm utilizes conditional independency for judging whether links are deleted or connected in Bayesian network structure [31]. Because constraint-based methods evaluate the relationship between nodes using training data, efficiency of entire Bayesian network structure is not assured when using the constraint-based methods. In search-and-score methods, Bayesian network structure is evaluated by score such as Bayesian score function, BIC, MDL, and MML (V, V, and V in our study). Based on the scores obtained from entire Bayesian network structures, better structure is searched or selected. MCMC is used for searching Bayesian network structure in our method and the previous study [19], and greedy algorithm is used in K2 algorithm [32]. In general, greedy algorithm, such as K2 algorithm, frequently sticks in local minimum/maximum, and cannot reach global minimum/maximum [19]. MCMC can break out of this local minimum/maximum and obtain better score [19]. As shown, our proposed method is classified as search-and-score methods. It is possible to use hybrid method of constraint-based methods and search-and-score methods. For example, in MCMC step of our method, the links between two nodes where conditional independency is indicated can be ignored when updating Bayesian network structure, which will make convergence speed of our proposed method faster. In conventional methods of structure learning, inference accuracy is mainly optimized, and explanation adequacy is frequently ignored. We focused on both inference accuracy and explanation adequacy in our study. In addition, our proposed method can speculate reasoning for prediction of one particular lung nodule. These two points are the major differences between our proposed method and conventional methods of Bayesian network structure learning/conventional CAD. There are several limitations in the current study. First, the number of parent nodes for the Bayesian network is limited to no more than two. In the case of a directed graph, such as Bayesian network, the number of possible structures can reach 3 (where B = C2 and N is the number of nodes). By limiting the number of parent nodes, it is possible to decrease computational cost for model construction, but this might have been at the expense of missing the optimal model. Second, despite restricting the number of model structures, the number of model candidates is still huge. Consequently, the automatic model construction in the MCMC process can reach the local minimum. Although our inference models converged to a reasonable model for the radiologists, this might not have been the optimal model. Third, the computational cost of model construction is large, with one trial requiring 30–40 hours to complete, making it difficult to construct models with different random seeds. Finally, the two radiologists only evaluated the best model. In future research, it will be preferable to perform subjective evaluation of more inference models.

Conclusions

In conclusion, we have proposed a method of automatic model construction for CADx of lung nodules that had high explanation adequacy and high inference accuracy. Notably, not only were the models constructed with prior knowledge from radiologists superior to those constructed without prior knowledge but the radiologists also considered the reasons provided for the inference results to be acceptable. Overall, these results suggest that our proposed CADx system might be acceptable in clinical practice and could eliminate the usual distrust of such systems among radiologists. We will perform further observational studies using our CAD system.

Performance of the ten inference models constructed with prior knowledge.

Except one model, performance of these models with prior knowledge was better than that of the best three models without prior knowledge (please compare S1 Table with Table 2). (DOCX) Click here for additional data file.

Performance of the two inference models constructed without prior knowledge.

(DOCX) Click here for additional data file.

List of imaging findings and clinical data.

(DOCX) Click here for additional data file.

Calculation of I(R).

(DOCX) Click here for additional data file.

Inference and reasoning using gradient tree boosting.

To compare our Bayesian-network-based method, inference and reasoning were performed using gradient tree boosting. (DOCX) Click here for additional data file.
  17 in total

1.  Computer-aided diagnosis to distinguish benign from malignant solitary pulmonary nodules on radiographs: ROC analysis of radiologists' performance--initial experience.

Authors:  Junji Shiraishi; Hiroyuki Abe; Roger Engelmann; Masahito Aoyama; Heber MacMahon; Kunio Doi
Journal:  Radiology       Date:  2003-05       Impact factor: 11.105

2.  Stochastic relaxation, gibbs distributions, and the bayesian restoration of images.

Authors:  S Geman; D Geman
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  1984-06       Impact factor: 6.226

3.  Neural network ensemble-based computer-aided diagnosis for differentiation of lung nodules on CT images: clinical evaluation.

Authors:  Hui Chen; Yan Xu; Yujing Ma; Binrong Ma
Journal:  Acad Radiol       Date:  2010-02-18       Impact factor: 3.173

4.  Computer-aided diagnosis for improved detection of lung nodules by use of posterior-anterior and lateral chest radiographs.

Authors:  Junji Shiraishi; Feng Li; Kunio Doi
Journal:  Acad Radiol       Date:  2007-01       Impact factor: 3.173

5.  Exploring new possibilities for case-based explanation of artificial neural network ensembles.

Authors:  Michael Green; Ulf Ekelund; Lars Edenbrandt; Jonas Björk; Jakob Lundager Forberg; Mattias Ohlsson
Journal:  Neural Netw       Date:  2008-10-17

Review 6.  Anniversary paper: History and status of CAD and quantitative image analysis: the role of Medical Physics and AAPM.

Authors:  Maryellen L Giger; Heang-Ping Chan; John Boone
Journal:  Med Phys       Date:  2008-12       Impact factor: 4.071

7.  Computer-aided differentiation of malignant from benign solitary pulmonary nodules imaged by high-resolution CT.

Authors:  Shingo Iwano; Tatsuya Nakamura; Yuko Kamioka; Mitsuru Ikeda; Takeo Ishigaki
Journal:  Comput Med Imaging Graph       Date:  2008-05-22       Impact factor: 4.790

8.  Computer-aided diagnosis of lung nodules on CT scans: ROC study of its effect on radiologists' performance.

Authors:  Ted Way; Heang-Ping Chan; Lubomir Hadjiiski; Berkman Sahiner; Aamer Chughtai; Thomas K Song; Chad Poopat; Jadranka Stojanovska; Luba Frank; Anil Attili; Naama Bogot; Philip N Cascade; Ella A Kazerooni
Journal:  Acad Radiol       Date:  2010-03       Impact factor: 3.173

9.  Application of an artificial neural network to high-resolution CT: usefulness in differential diagnosis of diffuse lung disease.

Authors:  Aya Fukushima; Kazuto Ashizawa; Tetsuji Yamaguchi; Naohiro Matsuyama; Hideyuki Hayashi; Isao Kida; Yoshihiro Imafuku; Akiko Egawa; Seigo Kimura; Kenji Nagaoki; Sumihisa Honda; Shigehiko Katsuragawa; Kunio Doi; Kuniaki Hayashi
Journal:  AJR Am J Roentgenol       Date:  2004-08       Impact factor: 3.959

10.  On the interplay of machine learning and background knowledge in image interpretation by Bayesian networks.

Authors:  Marina Velikova; Peter J F Lucas; Maurice Samulski; Nico Karssemeijer
Journal:  Artif Intell Med       Date:  2013-02-07       Impact factor: 5.326

View more
  4 in total

Review 1.  Evolving the pulmonary nodules diagnosis from classical approaches to deep learning-aided decision support: three decades' development course and future prospect.

Authors:  Bo Liu; Wenhao Chi; Xinran Li; Peng Li; Wenhua Liang; Haiping Liu; Wei Wang; Jianxing He
Journal:  J Cancer Res Clin Oncol       Date:  2019-11-30       Impact factor: 4.553

2.  Ada-GridRF: A Fast and Automated Adaptive Boost Based Grid Search Optimized Random Forest Ensemble model for Lung Cancer Detection.

Authors:  Ananya Bhattacharjee; R Murugan; Badal Soni; Tripti Goel
Journal:  Phys Eng Sci Med       Date:  2022-06-30

3.  A systematic review and meta-analysis of diagnostic performance and physicians' perceptions of artificial intelligence (AI)-assisted CT diagnostic technology for the classification of pulmonary nodules.

Authors:  Guo Huang; Xuefeng Wei; Huiqin Tang; Fei Bai; Xia Lin; Di Xue
Journal:  J Thorac Dis       Date:  2021-08       Impact factor: 3.005

4.  Detecting neonatal acute bilirubin encephalopathy based on T1-weighted MRI images and learning-based approaches.

Authors:  Miao Wu; Xiaoxia Shen; Can Lai; Weihao Zheng; Yingqun Li; Zhongli Shangguan; Chuanbo Yan; Tingting Liu; Dan Wu
Journal:  BMC Med Imaging       Date:  2021-06-22       Impact factor: 1.930

  4 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.