Feature Integration of [18F]FDG PET Brain Imaging Using Deep Learning for Sensitive Cognitive Decline Detection.
Overview
abstract
BACKGROUND: Distinguishing individuals with cognitive decline (CD), including early Alzheimer's disease, from cognitively normal (CN) individuals is essential for improving diagnostic accuracy and enabling timely intervention. Positron emission tomography (PET) captures functional brain alterations associated with CD, but its broader application is often limited by cost and radiation exposure. To enhance the clinical utility of PET while addressing data limitations, we propose a multi-representational learning framework that leverages both imaging data and region-level quantification in a data-efficient manner. METHODS: Voxel-level features were extracted using convolutional neural networks (CNN) or principal component analysis networks (PCANet) from [¹⁸F]FDG PET imaging. Region-level features were derived from standardized uptake value ratio measurements across predefined brain regions and processed using a deep neural network (DNN). These voxel- and region-level information are integrated through direct concatenation. For final prediction, different machine learning models and ensemble technique were applied. The models were trained and validated using 5-fold cross-validation on PET scans from 252 participants in the Alzheimer's Disease Neuroimaging Initiative (ADNI), comprising 118 CN and 134 CD subjects. Additional correlation analysis and disease classification comparison with the Mini-Mental State Examination (MMSE) were also performed. RESULTS: In 5-fold cross-validation, CNN, PCANet, and DNN models achieved classification accuracies of 0.69 ± 0.04, 0.69 ± 0.06, and 0.82 ± 0.06, respectively. The integrated DNN-CNN model using direct concatenation yielded the highest accuracy (0.87 ± 0.05), with a 6.10% improvement in accuracy and reduced standard deviation relative to the DNN-only model. Moreover, there were an increase of 14.29% in Recall (0.77 to 0.88) and an increase of 7.32% in F1-Score (0.82 to 0.88). Moreover, the model output showed a significant level of relation with MMSE, and it outperformed the MMSE-based classification in accuracy, recall, and f1, except precision. CONCLUSION: Combining PET imaging with region-level quantification and deep learning improves diagnostic performance over single-feature based models. Notably, fusion-based approaches enhanced sensitivity to cognitive decline. This multimodal strategy offers a more data-efficient and accurate approach for classifying cognitive decline and supports broader PET application in clinical settings.