Literature DB >> 34268435

Artificial intelligence for molecular neuroimaging.

Amanda J Boyle1, Vincent C Gaudet2, Sandra E Black3, Neil Vasdev1,4, Pedro Rosa-Neto5, Katherine A Zukotynski6.   

Abstract

In recent years, artificial intelligence (AI) or the study of how computers and machines can gain intelligence, has been increasingly applied to problems in medical imaging, and in particular to molecular imaging of the central nervous system. Many AI innovations in medical imaging include improving image quality, segmentation, and automating classification of disease. These advances have led to an increased availability of supportive AI tools to assist physicians in interpreting images and making decisions affecting patient care. This review focuses on the role of AI in molecular neuroimaging, primarily applied to positron emission tomography (PET) and single photon emission computed tomography (SPECT). We emphasize technical innovations such as AI in computed tomography (CT) generation for the purposes of attenuation correction and disease localization, as well as applications in neuro-oncology and neurodegenerative diseases. Limitations and future prospects for AI in molecular brain imaging are also discussed. Just as new equipment such as SPECT and PET revolutionized the field of medical imaging a few decades ago, AI and its related technologies are now poised to bring on further disruptive changes. An understanding of these new technologies and how they work will help physicians adapt their practices and succeed with these new tools. 2021 Annals of Translational Medicine. All rights reserved.

Entities:  

Keywords:  Artificial intelligence (AI); machine learning (ML); medical imaging; neuroimaging

Year:  2021        PMID: 34268435      PMCID: PMC8246223          DOI: 10.21037/atm-20-6220

Source DB:  PubMed          Journal:  Ann Transl Med        ISSN: 2305-5839


Introduction

Machine learning (ML) is a branch of AI that allows machines to perform tasks characteristically requiring human intelligence. Specifically, ML encompasses the set of algorithms and techniques enabling computers to process data without explicit programmed instructions dictating how the data is processed. Rather, the machine determines its own rules based on exposure to training data that can then be applied to process new data. For example, in terms of molecular brain imaging, ML algorithms might be used for applications such as: (I) generating computed tomography (CT) scans for attenuation correction (AC), (II) segmentation, (III) diagnosing disease or (IV) making outcome predictions. ML algorithms are typically divided according to whether learning (i.e., training) is supervised, unsupervised or semi-supervised (reinforcement learning). In supervised learning an algorithm is given a training dataset that includes the classification or outcome for each item in the dataset so that the algorithm may infer distinguishing features. The idea is that once these distinguishing features are learned by the algorithm, the algorithm will be able to correctly classify new data according to the information learned from the input data. Common supervised algorithms are, for example, random forests (RFs) and support vector machines (SVMs). In unsupervised learning, the algorithm is provided with a dataset where the classification of the data is not given. The algorithm seeks common features and clusters the data according to these features. Common unsupervised algorithms include, for example, k-means clustering. In certain less common instances, algorithms may use a combination of supervised and unsupervised techniques. It is said that these algorithms have a semi-supervised learning approach. Deep learning (DL) represents a subset of ML algorithms. In DL, the algorithms used are typically artificial neural networks (ANNs) where the ANNs are structured in such a way that input data is processed at several levels of abstraction. A greater number of levels translates into “deeper” (more advanced) learning. Convolutional neural networks (CNNs) are a type of ANN that is structured to process images, and algorithms that use CNNs are increasingly being applied to tasks in medical imaging. Further details regarding the principles of AI and ML are described elsewhere (1,2). AI in molecular brain imaging is a rapidly evolving field with several notable publications across a broad spectrum of journals in the last few years (3-6). In the context of positron emission tomography (PET) and single photon emission computed tomography (SPECT), ML algorithms can be used to improve image quality, reduce scan time, delineate/segment brain tumors and metastases as well as assisting with image classification in the setting of neurodegenerative disease. This paper presents a review of recent applications of AI in molecular brain imaging, including neurodegenerative diseases, namely Alzheimer’s disease (AD) and Parkinson’s disease (PD), neuro-oncology, and other neurological abnormalities.

CT scan generation and AC

Today, attenuation and scatter correction for PET is typically performed using CT. In the absence of a concurrent CT, such as when dedicated brain PET scanners or scanners incorporating PET and magnetic resonance (MR) imaging (MRI) systems (PET/MR) are used, synthesis of CT data may be helpful for the purposes of attenuation and scatter correction. A few studies have shown CT data can be derived from MR images using DL algorithms (7,8). In an exploratory study, a deep convolutional auto-encoder (CAE) network was trained to identify air, bone, and soft tissue using 30 MR co-registered to CT imaging and to generate synthetic CT data. The CAE is essentially a CNN with 13 convolutional layers that takes the MR as input, progressively reduces resolution, i.e., encode into features, followed by another 13 convolutional layers that progressively reconstruct images, i.e., decode, until a synthetic CT of original resolution is obtained. This algorithm was then evaluated in ten patients by comparing the synthesized CT to the acquired CT scans (8). The synthesized CT scans were then used to perform AC for an additional PET scans in five healthy subjects. The results were compared to standard MR- and CT-based AC. The authors found the PET reconstruction error was significantly lower using the DL algorithm than using conventional MR-based attenuation approaches (8). In another study, CT scans were generated by training a CNN model, also with encoding and decoding phases organized in a Unet-style structure (i.e., encoder layers followed by decoder layers), to map MR images to their corresponding CT images based upon CT and MR data from 18 patients with brain tumors (7). This method outperformed an atlas-based approach in less than one hundredth the amount of time. CT data may also be derived from PET emission data. Arabi and Zaidi developed a method to predict AC factors using time-of-flight (TOF) PET emission data from [18F]FDG brain scans and DL (9). With their approach, TOF PET emission sinogram data was used to train a CNN to estimate AC factors with CT-based AC data serving as a reference. The estimated AC factors were then used for AC during PET image reconstruction and compared to a two-tissue class segmentation AC method (background-air and soft-tissue). This approach had significantly higher accuracy than the standard method, in terms of relative mean absolute error and root mean square error (9). This research group subsequently investigated assessing attenuation and scatter correction with only TOF PET emission for [18F]FDG, in comparison to [18F]DOPA, [18F]flortaucipir, and [18F]flutemetamol, which target dopamine receptors, tau pathology, and amyloid pathology, respectively (10). This work, based on the same CNN structure as in (9), showed moderate performance of their DL AC algorithm at generating PET attenuation corrected images, with <9% absolute standardized uptake value (SUV) bias for the radiotracers studied, and the best performance seen with [18F]FDG. Overall, this method was found to be vulnerable to outliers, still the most intriguing result shown is the greater bias observed with the radiotracers that have greater signal-to-noise ratios compared to [18F]FDG. Hwang et al. also used the recent advancement of TOF PET technology to augment their DL approach to AC when investigating patient datasets using the radiopharmaceutical 18F-fluorinated-N-3-fluoropropyl-2-β-carboxymethoxy-3-β-(4-iodophenyl)nortropane {[18F]FP-CIT} for brain dopamine transporter imaging (11). This radiopharmaceutical leads to high crosstalk artifacts and noise due to high uptake in the striatum leading to regions of high contrast against background. The aforementioned study aimed to overcome these limitations with DL. Based on TOF PET data alone, they employed three different CNN methods and found that a hybrid network combining aspects of CAE and Unet generated estimated linear coefficient attenuation maps similar to those determined by CT. Shiri et al. also aimed to obtain attenuation-corrected PET images directly from PET emission data and investigated [18F]FDG-PET brain images from 129 patients (12). Using a deep convolutional encoder-decoder architecture, their network was designed to map non-AC PET images to pixel-wise continuously valued measured AC PET images (12). Their approach achieved comparable results relative to CT-based AC images, and further demonstrated the potential of DL methods for emission-based AC of PET images. Generative Adversarial Networks (GANs) are a recent DL innovation where two ANNs compete against each other, with the goal of learning how to generate images with characteristics similar to those in the training set. GANs have recently been used to generate AC images (13-15). Armanious et al. trained a GAN to generate pseudo CTs in a supervised approach that used pairs of [18F]FDG PET and their corresponding CT for training (13). The approach, which was found to have satisfactory accuracy, was tested by comparing PETs that were attenuation-corrected using both actual CTs and pseudo CTs.

AI in neuro-oncology

AI has been applied to molecular imaging in neuro-oncology to differentiate normal and abnormal images, provide automated tumor delineation and segmentation, predict survival outcomes, and aid in the development personalized treatment plans (16-24). The interpretation of [18F]FDG PET brain scans is challenging due to the high background glucose metabolism of gray matter which leads to low signal-to-noise ratios (25). AI has been shown to have the potential to mitigate this issue and improve the interpretation of [18F]FDG PET scans. CNNs use either two-dimensional (2D) or three-dimensional (3D) convolutional “kernels”, or matrixes, based on slice-to-slice data or a patch of all scans, respectively. 3D CNNs are computationally more intensive than 2D CNNs, yet are superior because they use voxel-wise rich data from 3D images which results in more accurate classification of scans. Nobashi et al. aimed to develop a 2D CNN based on a ResNet structure using multiple axes (transverse, coronal, and sagittal) with multiple window intensity levels to accurately classify [18F]FDG brain PET scans in cancer patients (16). This study showed that despite using less rich extraction of information from PET scans, 2D CNNs have the potential to differentiate normal and abnormal brain scans, particularly ensemble models (i.e., models that generate their overall classifications based on sub-classifications from multiple ML algorithms) with different window settings and axes compared to individual models. Given that accurate tumor delineation and segmentation is critical in diagnosing and staging cancer, and that manual contouring is time-consuming and open to subjectivity, AI applications for automated tumor delineation is an attractive goal in oncology (17-21). Comelli et al. have done extensive work in recent years towards this goal, investigating various algorithms based on 2D PET slices to obtain the 3D shape of the tumor with [18F]FDG and carbon-11 labeled methionine {[11C]MET} radiotracers in brain PET images (26,27). Recently, the same research group proposed a PET reconstruction based on a 3D system in which the segmentation is performed by evolving an active surface directly into the 3D space (28). The relevance of this work in tumor delineation is that it performs shape reconstruction on the entire stack of slices simultaneously, naturally leveraging cross-slice data that could not be exploited in the 2D system. This fully 3D approach was evaluated in a data set of 50 [18F]FDG and [11C]MET PET scans of lung, head and neck, and brain tumors and a benefit was demonstrated with their novel approach, and though fully automated tumor segmentation was not achieved at this time, their future work incorporating ML components and CT data is intriguing. [11C]MET PET brain scans have also been the target of supervised ML to predict survival outcomes in glioma patients (22). In this study, 16 predictive ML models were investigated with combinations of in vivo features from [11C]MET PET scans, ex vivo features, and patient features for their ability to predict 36-month survival, with the most successful ML methods involving data from all three features with survival prediction sensitivities and specificities ranging from 86–98% and 92–95%, respectively (22). In the case of glioblastoma, treatment plans involve MRI to delineate tumors for resection followed by radiotherapy and chemotherapy to target residual and occult tumor cells. Unique to glioblastoma is the infiltration of occult cancer cells into the surrounding area that is not detected by MRI. Current treatment plans address this issue by uniformly extending the region of radiotherapy beyond the volume of the visible tumor, which does not address that the extent of occult infiltration is variable among glioblastoma patients (29,30). To address this issue, Lipkova et al. described a Bayesian ML framework that integrates data from MRI and PET scans with [18F]fluoroethyl-tyrosine {[18F]FET}, which has been shown to have uptake values proportional to tumor cell density and is much better at delineating tumor regions with occult cells than MRI (23). In a preliminary clinical research study, the radiotherapy plans generated from the inferred tumor cell infiltration maps obtained with the proposed Bayesian ML framework spare healthy tissue from unnecessary irradiation while yielding comparable accuracy with the standard radiotherapy protocol, as shown in .
Figure 1

A comparison of (A) personalized dose distribution plan of radiotherapy generated from a Bayesian ML framework that integrates data from MRI and PET scans with the PET radiotracer [18F]FET and the dose distribution plan generated by standard protocols (B). Modified from Lipkova et al., 2019 (23). MRI, magnetic resonance imaging; PET, machine learning; ML, machine learning.

A comparison of (A) personalized dose distribution plan of radiotherapy generated from a Bayesian ML framework that integrates data from MRI and PET scans with the PET radiotracer [18F]FET and the dose distribution plan generated by standard protocols (B). Modified from Lipkova et al., 2019 (23). MRI, magnetic resonance imaging; PET, machine learning; ML, machine learning.

AI in neurodegenerative disease

AI has shown remarkable accuracy in diagnostic classification of neurodegenerative diseases such as AD and PD (4,6,31). Some approaches identify patterns and extract various features associated with AD and distinguish these from non-AD groups. In contrast, others aim to identify patterns in healthy controls and remove these from AD datasets. PET radiopharmaceuticals commonly used for studying AI applications in AD include [18F]FDG, to image glucose metabolism, as well as amyloid plaque imaging agents [18F]florbetapir, [18F]florbetaben, [11C]PiB, and [18F]flutemetamol, and tau imaging agents such as [18F]flortaucipir, among others. SPECT has also been studied for classifying AD with AI by examining cerebral blood flow, most commonly with the radiopharmaceutical 99m-technetium-ethyl cysteinate diethyl ester {[99mTc]ECD}. Identifying patients with AD and mild cognitive impairment (MCI) with cerebral amyloid deposition may be helpful and has been studied using ML algorithms. This process is currently performed by a designated imaging analyst and is subject to human interpretation. Zukotynski et al. demonstrated the ability of RFs, a supervised ML algorithm, to successfully classify amyloid brain PET scans into positive or negative with sensitivity, specificity, and accuracy of 86%, 92%, and 90%, respectively (32). This approach used a 10,000-tree RF, where each tree was created using 15 randomly selected cases and 20 randomly selected features of SUV ratio per region of interest (ROI) (32). Zukotynski et al. next strived to identify ROIs in amyloid {[18F]florbetapir} and [18F]FDG PET scans that may be associated with a patient’s Montreal Cognitive Assessment (MoCA) score, again using RFs (33). In this study, RFs with 1,000 trees were constructed, with each tree using a random subset of 16 training [18F]florbetapir PET scans and 20 ROIs; the RFs identified ROIs associated with MoCA. Additionally, RFs with 1,000 trees were similarly constructed based on [18F]FDG PET scans, and identified specific ROIs associated with MoCA scores (33). Another study aiming to classify AD with the radiotracer [18F]florbetapir examined two structurally identical CNNs trained separately on MRI and amyloid PET data and compared the classification of each separately, as well as a different network that combined the two sets of data (34). This study determined that the greatest classification accuracy of 92%±2% was obtained with the CNN that used both MRI and amyloid PET data. The majority of published studies focus on investigating a specific algorithm, hence comparing algorithms is challenging considering that the datasets in different studies vary in data acquisition, PET protocols, and data reconstructions. Brugnolo et al. addresses this gap in AI literature by reporting a head-to-head comparison of various automatic semi-quantification tools for their diagnostic performance based on brain [18F]FDG PET scans from prodromal AD patients and healthy controls (35). This aforementioned study compared three different statistical parametric mapping (SPM) approaches with a voxel-based tool, and a volumetric ROI SVM-based approach for their ability to differentiate prodromal AD from healthy control scans and found that the volumetric ROI-SVM-based approach outperformed all others studied with an area under the receiver operating characteristic (ROC) curve (AUC) of 0.95 compared to 0.84 and 0.83 for two different volumetric ROI-based methods, 0.79 for SPM maps, and 0.87 for the voxel-based tool (35). In a slightly different approach to classifying AD, Choi et al. developed an unsupervised DL model, based on variational autoencoder, trained only by normal [18F]FDG brain PET scans, to define an “Abnormality Score” depending on how far a given patient scan was from the normal data (36). They measured the accuracy of their model using AUC analysis and report an AUC of 0.90 for differentiating AD from MCI data. The model was also validated on a heterogeneous cohort of patients with non-dementia related abnormalities, including abnormal behavioral symptoms and epilepsy, and found that experts’ visual interpretation was aided by the model in identifying non-dementia related abnormal patterns, in 60% of cases that did not initially identify as abnormal without the model (36). AI has also been applied to molecular imaging data to predict patient outcomes. For example, to predict progression from MCI to dementia due to AD, Mathotaarachchi et al. implemented a random under sampling random forest (RUSRF) approach (37). This approach applied to the Alzheimer’s Disease Neuroimaging Initiative (ADNI) dataset provided a [18F]florbetapir prediction accuracy of 84% and an AUC of 0.91. A demonstration is available online at: http://predictalz.tnl-mcgill.com/PredictAlz_Amy. As is common when data sets are limited in scope and size, many studies investigating the use of ML algorithms for automated prediction and classification of neurological disorders randomly divide their data into input “training” and “testing” groups; this is repeated many times for cross-validation, such that they are training their networks from the same pool of data they test with. This does not address whether or not the algorithm will be successful with new, completely unseen data from independently recruited participants (38). Improving the potential of ML algorithms to generalize complex neurological diseases outside their ‘training’ and ‘testing’ datasets can be done by incorporating more diverse data with ensemble models and may be beneficial when stratifying patients for therapeutic clinical trials. For example, amyloid beta deposition is a hallmark of AD, and is believed to start to develop many years before the onset of dementia (39). Several clinical trials attempting to target and lower Aβ levels in the brain in MCI and early dementia due to AD have failed. This failure may be partly attributed to patient selection. For example, in the negative bapineuzumab and solanezumab studies, a high percentage of patients with dementia but without increased cerebral Aβ levels were enrolled, and such patients may not be expected to benefit from the treatment (22). More recently, a positive amyloid PET scan is required for enrolment in anti-amyloid antibody trials. Often several data-types are included for the purpose of predicting outcome. Franzmeier et al. (38) applied a support vector regression to AD biomarkers from cerebrospinal fluid, structural MRI, amyloid-PET ([11C]PiB), and [18F]FDG data. Ensemble models combining multiple classifiers are known to improve prediction performance compared to individual classifiers (40). By incorporating data from 2 separate databases, Franzmeier et al. developed a robust algorithm that was able to predict the 4-year rate of cognitive decline in sporadic prodromal AD patients. The impact of such an algorithm is a substantial reduction in the sample size required in AD clinical trials, because proper stratification of patients who are most likely to benefit could be selected based on a positive amyloid PET scan. (38). Similarly, Ezzati et al. developed a ML model to predict Aβ status by including demographic, ApoE4 status, neuropsychological tests, MRI volumetric data and CSF biomarkers and training the algorithm with [18F]florbetapir brain PET scans from the ADNI database (41). The ML models that combined CSF biomarkers significantly outperformed neuropsychological tests and MR volumetrics alone. An ensemble model described by Wu et al. (42) combined three different classifiers using weighted and unweighted schemes improved the accuracy of predicting AD outcomes based on [11C]PiB brain PET images and determined that weighted ensemble models outperformed K-Nearest Neighbors, RFs, and ANNs. A novel, yet simple, approach of denoising using unsupervised ML was employed to improve signal-to-noise ratio of AD pattern expression scores from [18F]FDG brain PET scans, using 220 MCI patient scans available from ADNI (43). This work used principal component analysis—an unsupervised learning approach—to identify patterns of non-pathological variance in healthy control scans and removing them from the patient data before calculating the AD pattern expression score. Their pattern expression score-based prediction of AD conversion significantly improved from AUC 0.80 to 0.88. When combined with neuropsychological features and ApoE4 genotype data, the prediction model further improved with AUC 0.93 (43). A few studies have examined combining [18F]FDG PET data with [99mTc]ECD SPECT data, which provides information about cerebral blood flow, to differentiate AD from healthy controls (44,45). Ferreira et al. compared applying a SVM model to data from [18F]FDG PET, [99mTc]ECD, and MRI images to determine which dataset would provide the greatest classification power (44). They achieved reasonable, though significantly greater accuracy with [18F]FDG PET data (68-71%), and [99mTc]ECD SPECT (68-74%), compared to MRI data (58%). In a later study, this research group examined applying a multiple kernel learning (MKL) method that would allow identification of the most relevant ROIs of classification based again on data from [18F]FDG PET, [99mTc]ECD SPECT, and MRI images (45). They compared using two different atlases to define ROIs from each imaging dataset, automated anatomical labeling and Brodmann’s area, and applied these to their MKL method, then compared these with the SVM model. They found much greater accuracy with their MKL method compared to SVM, the highest being for [18F]FDG PET with 92.50% accuracy found with both atlases (45). AI has also been shown to have potential to classify PD using PET data with [18F]FDG or L-3,4-dihydroxy-6-[18F]fluorophenylalanine ([18F]DOPA), to assess degeneration in the dopaminergic pathway. Glaab et al. investigated the diagnostic power of two different ML methods, SVM and RF, on metabolic data with or without [18F]FDG or [18F]DOPA imaging, at differentiating PD from healthy controls (46). This study found that diagnostic power was lowest for any of these datasets alone, and that the greatest diagnostic power was obtained with combined [18F]DOPA and metabolic datasets (AUC =0.98). One limitation of AI includes overfitting of data, Shen et al. aimed to address this issue by adding the Group Lasso Sparse model to a Deep Belief Network (47). They applied their model to [18F]FDG PET scans from 2 cohorts of PD and healthy control subjects, the first cohort was randomly divided into training, validation, and test datasets, while the second cohort was used only to test the model. Their model outperformed traditional Deep Belief Networks at classifying [18F]FDG PET scans as PD or non-PD. Another study investigated extracting high-order radiomic features from [18F]FDG PET scans to differentiate PD from healthy controls (48). ROIs were identified based on the atlas-method; then features were selected by autocorrelation and the Fisher score algorithm. A SVM was trained on these extracted features to classify PD and healthy controls, which achieved 88.1%±5.3% accuracy compared to 80.1%±3.1% achieved with the traditional voxel values method (48). SPECT is routinely performed to detect dopamine transporters with imaging agents such as 123-iodine-fluoropropylcarbomethoxyiodophenylnortropane ([123I]FP-CIT) (49). Several studies have investigated applying AI to [123I]FP-CIT SPECT datasets to achieve automated classification of PD. Choi et al. developed a CNN model trained on [123I]FP-CIT SPECT data from PD and healthy control subjects from the Parkinson’s Progression Markers Initiative (PPMI) database, then validated the model with an independent cohort of PD and non-parkinsonian tremor patients from the Seoul National University Hospital (50). Their model was able to overcome variability of human evaluation, and was able to accurately classify the atypical subgroup of PD in which dopaminergic deficit is absent in SPECT scans despite a clinical diagnosis of PD (50). Nicastro et al. also aimed to provide an automated classification model for PD combining semiquantitative striatal [123I]FP-CIT SPECT indices and SVM analysis in an approach that was able to differentiate PD from atypical and non-PD syndromes (51). In another study, researchers extracted seven features from the brain scans and assessed their ability to diagnose PD separately and in combination using SVM, k-nearest neighbors, and logistic regression (52). The greatest classification potential, 97.9% accuracy, 98% sensitivity, and 97.6% specificity, was found to be with SVM applied to all features simultaneously. In a similar study, [123I]FP-CIT SPECT scans were pre-processed using automated template-based registration followed by computation of the radiopharmaceutical binding potential at a voxel level, then using these data to classify images using a SVM model (53). This approach is currently being studied in clinical trials (ClinicalTrial.gov identifier NCT01141023).

Challenges, current status, and future prospects

The field of AI has grown substantially in recent years and has been applied to brain molecular imaging to improve the quality of images, reduce the amount of time a patient must be on a scanner, and assist in physician decision making and clinical interpretation of images. To improve the quality of images obtained from PET imaging systems that lack a concurrent CT scanner, AI has been applied to either MR or PET data to determine accurate AC and create synthetic CT data. By applying AI to PET data to obtain AC a patient may endure shorter scan times. AI has shown potential in assisting decision making in clinical care by automating the delineation of brain tumors and metastases from reconstructed imaging data, as well as automated classification of neurological diseases, such as AD and PD. Current clinical trials indicate that the use of AI in brain molecular imaging is indeed a growing field (ClinicalTrials.gov identifiers NCT00330109, NCT04357236, NCT04174287). A rising trend in AI in brain molecular imaging is studies investigating ensemble models that involve multiple classifiers such as biomarkers and patient features. Neurological and neurodegenerative diseases are diagnosed by examining different features such as biomarkers in blood samples and scores on cognitive tests. Ensemble models that include these features with nuclear imaging data in their algorithms are showing greater classification and diagnostic potential compared to individual models. A gap exists in the literature regarding a lack of AI applied to novel radiopharmaceuticals; this is likely due to one of the major challenges of AI in nuclear imaging, which is the need for large training datasets. A typical clinical study with novel radiopharmaceuticals is unlikely to produce a large enough dataset to adequately train an AI algorithm. Another limitation is the potential biases of AI algorithms. Consider that the training, testing, and validation datasets are often selected randomly from the same larger dataset, this introduces bias into the validation of results. The ability of AI algorithms to perform tasks, such as the classification of images, and to have potential to be beneficial in a clinical setting is more evident if the models have been validated on independent cohorts from separate institutions. Additionally, it is difficult to compare AI algorithms used in different publications in the literature, for example, if considering algorithms for automated classification of AD, results will vary depending on the datasets used for validation. For this reason, the most robust studies on the development of new AI models will include a comparison to a standard AI model. Despite these challenges, FDA-approved AI technologies are moving into clinics. Siemens has recently received FDA approval for use on their Biograph family of PET/CT imaging systems (54). These AI technologies, collectively named AIDAN, are comprised of propriety DL algorithms that allow four new features: FlowMotion AI, OncoFreeze AI, PET FAST Workflow AI, and Multiparametric Pet Suite AI. FlowMotion AI technology allows fast tailoring of PET/CT protocols in different patients by automatically detecting anatomical structures whose locations and sizes vary greatly between patients. PET FAST WorkFlow AI is applied to post-scan tasks to permit faster interpretation of results and reducing possibility of error. OncoFreeze AI applies AI to PET/CT acquisitions to correct for patient respiration motion, which can compromise image quality. Multiparametric PET Suite AI extracts arterial input function from PET/CT images to eliminate the pain and risk associated with arterial sampling during PET/CT scans. Together, these technologies permit faster, more accurate PET/CT acquisitions and image interpretations while allowing a more comfortable experience for the patient. Additionally, a distribution partnership was recently announced between CorTechs Labs and Subtle Medical involving Subtle Medical’s proprietary software, SubtleMRTM and SubtlePETTM, which are the first FDA-approved AI applications for medical imaging enhancement. These will be combined with CorTech Labs’ software, NeuroQuant® and PETQuantTM, which may allow PET/MRI facilities to reduce scan time without foregoing quantitative volumetric image analyses (55). SubtleMRTM improves image quality and denoises MR scans, while SubtlePETTM is able to denoise PET scans that are carried out in 25% of the time of a typical PET scan. NeuroQuant® and PETQuantTM are applications that provide automated and rapid analysis of MR and PET images, respectively. The reduced acquisition times with quantitative analysis of images will likely improve the patient experience and aid in data interpretation. It is hard to say how the future of AI in molecular imaging will evolve. There remains the need for careful evaluation of the algorithms that are being used and developed, as well as where their use best fits in our clinical applications. The article’s supplementary files as
  52 in total

1.  Hybrid 11C-MET PET/MRI Combined With "Machine Learning" in Glioma Diagnosis According to the Revised Glioma WHO Classification 2016.

Authors:  Sied Kebir; Manuel Weber; Lazaros Lazaridis; Cornelius Deuschl; Teresa Schmidt; Christoph Mönninghoff; Kathy Keyvani; Lale Umutlu; Daniela Pierscianek; Michael Forsting; Ulrich Sure; Martin Stuschke; Christoph Kleinschnitz; Björn Scheffler; Patrick M Colletti; Domenico Rubello; Christoph Rischpler; Martin Glas
Journal:  Clin Nucl Med       Date:  2019-03       Impact factor: 7.794

2.  ATLAAS: an automatic decision tree-based learning algorithm for advanced image segmentation in positron emission tomography.

Authors:  Beatrice Berthon; Christopher Marshall; Mererid Evans; Emiliano Spezi
Journal:  Phys Med Biol       Date:  2016-06-08       Impact factor: 3.609

3.  Head-to-Head Comparison among Semi-Quantification Tools of Brain FDG-PET to Aid the Diagnosis of Prodromal Alzheimer's Disease.

Authors:  Andrea Brugnolo; Fabrizio De Carli; Marco Pagani; Slivia Morbelli; Cathrine Jonsson; Andrea Chincarini; Giovanni B Frisoni; Samantha Galluzzi; Robert Perneczky; Alexander Drzezga; Bart N M van Berckel; Rik Ossenkoppele; Mira Didic; Eric Guedj; Dario Arnaldi; Federico Massa; Matteo Grazzini; Matteo Pardini; Patrizia Mecocci; Massimo E Dottorini; Matteo Bauckneht; Gianmario Sambuceti; Flavio Nobili
Journal:  J Alzheimers Dis       Date:  2019       Impact factor: 4.472

4.  Identifying incipient dementia individuals using machine learning and amyloid imaging.

Authors:  Sulantha Mathotaarachchi; Tharick A Pascoal; Monica Shin; Andrea L Benedet; Min Su Kang; Thomas Beaudry; Vladimir S Fonov; Serge Gauthier; Pedro Rosa-Neto
Journal:  Neurobiol Aging       Date:  2017-07-11       Impact factor: 4.673

5.  Extraction, selection and comparison of features for an effective automated computer-aided diagnosis of Parkinson's disease based on [123I]FP-CIT SPECT images.

Authors:  Francisco P M Oliveira; Diogo Borges Faria; Durval C Costa; Miguel Castelo-Branco; João Manuel R S Tavares
Journal:  Eur J Nucl Med Mol Imaging       Date:  2017-12-23       Impact factor: 9.236

6.  Head and neck target delineation using a novel PET automatic segmentation algorithm.

Authors:  B Berthon; M Evans; C Marshall; N Palaniappan; N Cole; V Jayaprakasam; T Rackley; E Spezi
Journal:  Radiother Oncol       Date:  2017-01-23       Impact factor: 6.280

7.  Active contour algorithm with discriminant analysis for delineating tumors in positron emission tomography.

Authors:  Albert Comelli; Alessandro Stefano; Samuel Bignardi; Giorgio Russo; Maria Gabriella Sabini; Massimo Ippolito; Stefano Barone; Anthony Yezzi
Journal:  Artif Intell Med       Date:  2019-01-08       Impact factor: 5.326

8.  Predicting Amyloid-β Levels in Amnestic Mild Cognitive Impairment Using Machine Learning Techniques.

Authors:  Ali Ezzati; Danielle J Harvey; Christian Habeck; Ashkan Golzar; Irfan A Qureshi; Andrea R Zammit; Jinshil Hyun; Monica Truelove-Hill; Charles B Hall; Christos Davatzikos; Richard B Lipton
Journal:  J Alzheimers Dis       Date:  2020       Impact factor: 4.472

Review 9.  Clinical and Imaging Markers of Prodromal Parkinson's Disease.

Authors:  Eldbjørg Hustad; Jan O Aasly
Journal:  Front Neurol       Date:  2020-05-08       Impact factor: 4.003

10.  Predicting sporadic Alzheimer's disease progression via inherited Alzheimer's disease-informed machine-learning.

Authors:  Nicolai Franzmeier; Nikolaos Koutsouleris; Tammie Benzinger; Alison Goate; Celeste M Karch; Anne M Fagan; Eric McDade; Marco Duering; Martin Dichgans; Johannes Levin; Brian A Gordon; Yen Ying Lim; Colin L Masters; Martin Rossor; Nick C Fox; Antoinette O'Connor; Jasmeer Chhatwal; Stephen Salloway; Adrian Danek; Jason Hassenstab; Peter R Schofield; John C Morris; Randall J Bateman; Michael Ewers
Journal:  Alzheimers Dement       Date:  2020-02-11       Impact factor: 21.566

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.