Alzheimer's disease (AD) develops for an unknown and variable amount of time before its symptoms fully manifest. But, when the symptoms become clinically observable, a significant neurodegeneration has already taken place. Thus, there is a largely unmet need for technologies that can aid the effective early diagnosis and prognosis of AD in an in vivo and more objective manner. The goal of this renewal project is to develop a set of advanced machine-learning techniques for precise in vivo quantification of pathological changes of brains with multimodality neuroimaging for both early diagnosis and prognosis of AD. AD is a highly heterogeneous neurodegenerative disorder with complex pathophysiology, thus very challenging to pinpoint its subtle pathologies without any aid from advanced computational technologies. To this end, we propose the following four specific aims to identify those subtle disease-induced alterations, derive robust diagnostic conclusions, and predict future disease trajectories. Specifically, in Aim 1, we will develop a multi-view feature representation technique to robustly extract complementary information from neuroimaging data with multiple representative atlases, and then identify a small subset of most discriminative features for AD diagnosis. This novel multi-atlas technique will deviate from the conventional single-atlas approaches in feature representation, which are often susceptible to inter-subject structural variability, registration error, and atlas selection bias.
In Aim 2, we will further devlop two novel multi-view feature mapping techniques for collaborative fusion of multimodality information by explicitly considering the distribution heterogeneity of different categories of features extracted from different modalities. This will significantly avoid the unnecessary complexity of feature distributions after our collaborative fusion, thus increasing the efficacy of subsequent diagnostic classifiers. Specifically, a deep learning technique (with deep multi-layered architecture) will be adopted to hierarchically mine multimodality information that resides nonlinearly both within each modality and between different modalities.
In Aim 3, we will develop a novel multi-task sparse learning technique for joint prediction of diagnostic status and clinical scores (e.g., ADAS-Cog and MMSE) by considering the inherent correlations between features and between training samples. This will also allow us to exploit the latent structure underlying the data for robust estimation of these highly variable clinical scores. Finally, in Aim 4, we will jointly predict clinical scores of each given subject in multiple future time points, by developing coupled random forests that can take advantage of all training subjects with complete or even incomplete multimodality data and further enforce temporal consistency of those estimated clinical scores. All the above-proposed techniques will be evaluated by a large image set of elderly subjects in ADNI. We expect that the successful completion of this renewal project will result in a comprehensive and effective diagnosis/prognosis framework for improving early detection of AD. The respective software tools will be released freely to the research community, as we have done with our HAMMER software, which has been downloaded by >5200 users from >20 countries.
The goal of this renewal project is to develop a set of advanced machine-learning techniques for precise in vivo quantification of pathological changes, afforded by multimodality neuroimaging data, for both diagnosis and prognosis of Alzheimer's diseases (AD). Specifically, we will explicitly exploit the distribution complexity and hierarchical nature of the multimodality data, for identifying subtle disease-induced alterations, deriving robust diagnostic conclusions, and predicting the future disease trajectory.
|Zhao, Feng; Qiao, Lishan; Shi, Feng et al. (2017) Feature fusion via hierarchical supervised local CCA for diagnosis of autism spectrum disorder. Brain Imaging Behav 11:1050-1060|
|Yin, Qingbo; Hung, Sheng-Che; Wang, Li et al. (2017) Associations between Tumor Vascularity, Vascular Endothelial Growth Factor Expression and PET/MRI Radiomic Signatures in Primary Clear-Cell-Renal-Cell-Carcinoma: Proof-of-Concept Study. Sci Rep 7:43356|
|Rekik, Islem; Li, Gang; Yap, Pew-Thian et al. (2017) Joint prediction of longitudinal development of cortical surfaces and white matter fibers from neonatal MRI. Neuroimage 152:411-424|
|Chen, Xiaobo; Zhang, Han; Lee, Seong-Whan et al. (2017) Hierarchical High-Order Functional Connectivity Networks and Selective Feature Fusion for MCI Classification. Neuroinformatics 15:271-284|
|Chen, Xiaobo; Zhang, Han; Zhang, Lichi et al. (2017) Extraction of dynamic functional connectivity from brain grey matter and white matter for MCI classification. Hum Brain Mapp 38:5019-5034|
|An, Le; Adeli, Ehsan; Liu, Mingxia et al. (2017) A Hierarchical Feature and Sample Selection Framework and Its Application for Alzheimer's Disease Diagnosis. Sci Rep 7:45269|
|Zu, Chen; Wang, Zhengxia; Zhang, Daoqiang et al. (2017) Robust multi-atlas label propagation by deep sparse representation. Pattern Recognit 63:511-517|
|Shen, Dinggang; Wu, Guorong; Suk, Heung-Il (2017) Deep Learning in Medical Image Analysis. Annu Rev Biomed Eng 19:221-248|
|Zhang, Jun; Liu, Mingxia; Le An et al. (2017) Alzheimer's Disease Diagnosis Using Landmark-Based Features From Longitudinal Structural MR Images. IEEE J Biomed Health Inform 21:1607-1616|
|Zhang, Yu; Zhang, Han; Chen, Xiaobo et al. (2017) Hybrid High-order Functional Connectivity Networks Using Resting-state Functional MRI for Mild Cognitive Impairment Diagnosis. Sci Rep 7:6530|
Showing the most recent 10 out of 213 publications