Deep-learning prediction of amyloid deposition from early-phase amyloid positron emission tomography imaging
Issued Date
2022-10-01
Resource Type
ISSN
09147187
eISSN
18646433
Scopus ID
2-s2.0-85135252289
Pubmed ID
35913591
Journal Title
Annals of Nuclear Medicine
Volume
36
Issue
10
Start Page
913
End Page
921
Rights Holder(s)
SCOPUS
Bibliographic Citation
Annals of Nuclear Medicine Vol.36 No.10 (2022) , 913-921
Suggested Citation
Komori S., Cross D.J., Mills M., Ouchi Y., Nishizawa S., Okada H., Norikane T., Thientunyakit T., Anzai Y., Minoshima S. Deep-learning prediction of amyloid deposition from early-phase amyloid positron emission tomography imaging. Annals of Nuclear Medicine Vol.36 No.10 (2022) , 913-921. 921. doi:10.1007/s12149-022-01775-z Retrieved from: https://repository.li.mahidol.ac.th/handle/20.500.14594/85508
Title
Deep-learning prediction of amyloid deposition from early-phase amyloid positron emission tomography imaging
Other Contributor(s)
Abstract
Objective: While the use of biomarkers for the detection of early and preclinical Alzheimer's Disease has become essential, the need to wait for over an hour after injection to obtain sufficient image quality can be challenging for patients with suspected dementia and their caregivers. This study aimed to develop an image-based deep-learning technique to generate delayed uptake patterns of amyloid positron emission tomography (PET) images using only early-phase images obtained from 0–20 min after radiotracer injection. Methods: We prepared pairs of early and delayed [11C]PiB dynamic images from 253 patients (cognitively normal n = 32, fronto-temporal dementia n = 39, mild cognitive impairment n = 19, Alzheimer’s disease n = 163) as a training dataset. The neural network was trained with the early images as the input, and the output was the corresponding delayed image. A U-net convolutional neural network (CNN) and a conditional generative adversarial network (C-GAN) were used for the deep-learning architecture and the data augmentation methods, respectively. Then, an independent test data set consisting of early-phase amyloid PET images (n = 19) was used to generate corresponding delayed images using the trained network. Two nuclear medicine physicians interpreted the actual delayed images and predicted delayed images for amyloid positivity. In addition, the concordance of the actual delayed and predicted delayed images was assessed statistically. Results: The concordance of amyloid positivity between the actual versus AI-predicted delayed images was 79%(κ = 0.60) and 79% (κ = 0.59) for each physician, respectively. In addition, the physicians’ agreement rate was at 89% (κ = 0.79) when the same image was interpreted. And, the actual versus AI-predicted delayed images were not readily distinguishable (correct answer rate, 55% and 47% for each physician, respectively). The statistical comparison of the actual versus the predicted delated images indicated that the peak signal-to-noise ratio (PSNR) was 21.8 dB ± 2.2 dB, and the structural similarity index (SSIM) was 0.45 ± 0.04. Conclusion: This study demonstrates the feasibility of an image-based deep-learning framework to predict delayed patterns of Amyloid PET uptake using only the early phase images. This AI-based image generation method has the potential to reduce scan time for amyloid PET and increase the patient throughput, without sacrificing diagnostic accuracy for amyloid positivity.