TY - GEN
T1 - Parkinson's Disease Classification Using Contrastive Graph Cross-View Learning with Multimodal Fusion of Spect Images and Clinical Features
AU - Ding, Jun En
AU - Hsu, Chien Chin
AU - Liu, Feng
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - Parkinson's Disease (PD) affects millions globally, impacting movement. Prior research utilized deep learning for PD prediction, primarily focusing on medical images, neglecting the data's underlying manifold structure. This work proposes a multimodal approach encompassing both image and non-image features, leveraging contrastive cross-view graph fusion for PD classification. We introduce a novel multimodal co-attention module, integrating embeddings from separate graph views derived from low-dimensional representations of images and clinical features. This enables extraction of more robust and structured features for improved multi-view data analysis. Additionally, a simplified contrastive loss-based fusion method is devised to enhance cross-view fusion learning. Our graph-view multimodal approach achieves an accuracy of 91% and an AUC of 92.8% in five-fold cross-validation. It also demonstrates superior predictive capabilities on non-image data compared to solely machine learning-based methods.
AB - Parkinson's Disease (PD) affects millions globally, impacting movement. Prior research utilized deep learning for PD prediction, primarily focusing on medical images, neglecting the data's underlying manifold structure. This work proposes a multimodal approach encompassing both image and non-image features, leveraging contrastive cross-view graph fusion for PD classification. We introduce a novel multimodal co-attention module, integrating embeddings from separate graph views derived from low-dimensional representations of images and clinical features. This enables extraction of more robust and structured features for improved multi-view data analysis. Additionally, a simplified contrastive loss-based fusion method is devised to enhance cross-view fusion learning. Our graph-view multimodal approach achieves an accuracy of 91% and an AUC of 92.8% in five-fold cross-validation. It also demonstrates superior predictive capabilities on non-image data compared to solely machine learning-based methods.
KW - Contrastive learning
KW - Multi-view clustering
KW - Multi-view graph learning
KW - Multimodal fusion
UR - http://www.scopus.com/inward/record.url?scp=85203340234&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85203340234&partnerID=8YFLogxK
U2 - 10.1109/ISBI56570.2024.10635712
DO - 10.1109/ISBI56570.2024.10635712
M3 - Conference contribution
AN - SCOPUS:85203340234
T3 - Proceedings - International Symposium on Biomedical Imaging
BT - IEEE International Symposium on Biomedical Imaging, ISBI 2024 - Conference Proceedings
T2 - 21st IEEE International Symposium on Biomedical Imaging, ISBI 2024
Y2 - 27 May 2024 through 30 May 2024
ER -