| Literature DB >> 31190176 |
Gary J R Cook1,2, Vicky Goh3,4.
Abstract
While molecular imaging with positron emission tomography or single-photon emission computed tomography already reports on tumour molecular mechanisms on a macroscopic scale, there is increasing evidence that there are multiple additional features within medical images that can further improve tumour characterization, treatment prediction and prognostication. Early reports have already revealed the power of radiomics to personalize and improve patient management and outcomes. What remains unclear is how these additional metrics relate to underlying molecular mechanisms of disease. Furthermore, the ability to deal with increasingly large amounts of data from medical images and beyond in a rapid, reproducible and transparent manner is essential for future clinical practice. Here, artificial intelligence (AI) may have an impact. AI encompasses a broad range of 'intelligent' functions performed by computers, including language processing, knowledge representation, problem solving and planning. While rule-based algorithms, e.g. computer-aided diagnosis, have been in use for medical imaging since the 1990s, the resurgent interest in AI is related to improvements in computing power and advances in machine learning (ML). In this review we consider why molecular and cellular processes are of interest and which processes have already been exposed to AI and ML methods as reported in the literature. Non-small-cell lung cancer is used as an exemplar and the focus of this review as the most common tumour type in which AI and ML approaches have been tested and to illustrate some of the concepts.Entities:
Keywords: Artificial intelligence; Machine learning, deep learning; Molecular imaging; Radiomics
Mesh:
Year: 2019 PMID: 31190176 PMCID: PMC6879441 DOI: 10.1007/s00259-019-04370-z
Source DB: PubMed Journal: Eur J Nucl Med Mol Imaging ISSN: 1619-7070 Impact factor: 9.236
Fig. 1Convolutional neural network architecture for oesophageal cancer 18F-FDG PET data in a vector composed from four convolutional (U) and four max-pooling (V) layers. Differently coloured arrows in the first convolutional layer represent different learnable weight matrices. Coloured squares in the feature maps represent elements that include local spatial information from the previous layer. In the max-pooling layers, 2 × 2 element windows represent non-overlapping grids from which the maximum element to down-sample the feature maps are chosen (h hidden layer, y responder, y nonresponder). From Ypsilantis et al. [38]