| Literature DB >> 33837678 |
Handi Deng1, Hui Qiao2,3,4, Qionghai Dai2,3,4, Cheng Ma1,5.
Abstract
SIGNIFICANCE: Photoacoustic (PA) imaging can provide structural, functional, and molecular information for preclinical and clinical studies. For PA imaging (PAI), non-ideal signal detection deteriorates image quality, and quantitative PAI (QPAI) remains challenging due to the unknown light fluence spectra in deep tissue. In recent years, deep learning (DL) has shown outstanding performance when implemented in PAI, with applications in image reconstruction, quantification, and understanding. AIM: We provide (i) a comprehensive overview of the DL techniques that have been applied in PAI, (ii) references for designing DL models for various PAI tasks, and (iii) a summary of the future challenges and opportunities. APPROACH: Papers published before November 2020 in the area of applying DL in PAI were reviewed. We categorized them into three types: image understanding, reconstruction of the initial pressure distribution, and QPAI.Entities:
Keywords: convolutional neural network; deep learning; photoacoustic imaging
Year: 2021 PMID: 33837678 PMCID: PMC8033250 DOI: 10.1117/1.JBO.26.4.040901
Source DB: PubMed Journal: J Biomed Opt ISSN: 1083-3668 Impact factor: 3.170
Fig. 1The architecture of U-Net. Each blue box corresponds to a multi-channel feature map. White boxes represent copied feature maps. Gray arrows indicate straight—connected path on the same level.
Data sets commonly used in DL-based PAI
| Data set | Descriptions |
|---|---|
| Mammography image database from LAPIMO EESC/USP | The database consisted of around 1400 screening mammography images from around 320 patients. |
| DRIVE dataset | The database was used for comparative study of vascular segmentation in retinal images. It consisted of 40 photographs, 7 of which showed signs of mild early diabetic retinopathy. |
| Optical and acoustic breast database (OA-breast) | The database includes a collection of numerical breast phantoms generated from clinical magnetic resonance angiography data collected from Washington University in St. Louis School of Medicine. |
| Digital mouse | The database includes a 3D whole body mouse atlas from coregistered high-quality PET x-ray CT and cryosection data of a normal nude male mouse. |
| Shepp–Logan phantom | The Shepp–Logan phantom is a standard test image created by Larry Shepp and Benjamin F. Logan for their 1974 paper “The Fourier Reconstruction of a Head Section.” |
| 3D volume of CBA mouse brain vasculature | The database includes a high-resolution volumetric and vasculature atlas on CBA mouse brain based on a combination of magnetic resonance imaging and x-ray CT. |
| ELCAP public lung image database | The database consists of an image set of 50 low-dose whole-lung CT scans. The CT scans were obtained in a single-breath hold with a 1.25-mm slice thickness. The locations of nodules detected by the radiologist are also provided. |
| Big data from CT scanning | CT scans of several cadavers are provided. The data are collected at Massachusetts General Hospital at multiple different radiation dose levels for different x-ray spectra and with representative reconstruction techniques. |
| Tumor phantom in mouse brain | The database is based on segmentation of a micro-CT scan of a mouse brain into gray mater, vasculature, and dura mater. An artificial cancer tissue was created by a stochastic growth process. |
| VICTRE project | A series of toolkits include breast Phantom, breastCompress, and breastCrop. Using breast Phantom, the digital breast with varying patient characteristics (breast shape, glandularity and density, and size) can be generated. |
| CBIS-DDSM (curated breast imaging subset of DDSM) | The data set contains 2620 scanned film mammography studies including normal, benign, and malignant cases with verified pathology information. |
| Mouse PACT | The database consists of six athymic nude-Fox1nu mice (Harlan Laboratories) |
Fig. 2Diagram showing the applications of DL in PAI, and the structure of this review follows this diagram.
Network architectures used in PA image understanding.
| General task | Specific task | Network architecture |
|---|---|---|
| Image understanding | Classification | Simple NN; |
| Segmentation | U-Net | |
| Motion correction | Simple CNN |
Network architectures used in PA image reconstruction
| Non-ideal condition | Preprocessing | Postprocessing | Direct reconstruction | Combined reconstruction | Embedded in traditional reconstruction | |
|---|---|---|---|---|---|---|
| Single-non-ideal detection | Limited bandwidth | Simple NN | | U-Net | | |
| Sparse sampling | | U-Net | | | Simple CNN | |
| Limited view | | U-Net | U-Net | Multiple branches autoencoder | | |
| Non-uniform SoS | | Faster R-CNN | Simple CNN | | | |
| Multiple non-ideal conditions | U-net | U-Net; | U-Net | Ki-GAN (based on multiple branches autoencoder) | U-Net; | |
| Use of DL in economical and portable PAI devices | Single-channel DAQ | | | Autoencoder (based on LSTM) | | |
| LED-PAI | Complex CNN | U-Net | ||||
Network architectures used in QPAI
| Task categories | Network architecture |
|---|---|
| Calculate | U-Net, |
| Generate mask for trusted zones | U-Net |
Fig. 3PA reconstruction quality is influenced by different non-ideal conditions: (a) gold standard; (b) good-quality reconstruction (DAS); reconstruction with reduced quality due to (c) limited bandwidth; (d) sparse sampling; (e) limited view; and (f) wrong SoS.
Fig. 4The results of Davoudi et al.’s method. (a) Reconstructed image with under sampled (32 projections) data versus its artifact-free counterpart obtained with the trained network. TP, true positive; FP, false positive; TN, true negative; and FN, false negative. (b) Another example for under sampled data with 128 projections. (c) Top: image reconstructed with 60-deg angular coverage and its respective amplitude spectrum. Bottom: output image of the network and its corresponding amplitude spectrum. (d) Top: image reconstructed with 135-deg angular coverage and its respective amplitude spectrum. Bottom: output image of the network and its corresponding amplitude spectrum.
Fig. 5The architecture of Lan et al.’s model. (a) The overall architecture of Ki-GAN; KEB represents convolutional layers; DAS: delay and sum reconstruction. (b) The detailed architecture of KEB. (c) PA images of rat thigh reconstructed by iterative algorithm with 10 iterations (column 1), U-Net (column 2), and Ki-GAN (column 3).
Fig. 6The architecture of (a) Anas et al.’s network and (b) Johnstonbaugh et al.’s network. Details of the residual convolution module and the upsampling module are also provided.
Fig. 7(a) The ResU-Net architecture implemented by Cai et al. (b) Simultaneously reconstructed map (gray) and ICG concentration (color). (c) Absolute error (gray) and relative ICG concentration error (color). Scale bars: 5 mm.