Decoding Alzheimer's: Interpretable Visual and Logical Attention in Picture Description Tasks

  • Ning Wang
  • , Bingyang Wen
  • , Minghui Wu
  • , Yang Sun
  • , Zongru Shao
  • , Haojie Zhou
  • , K. P. Subbalakshmi

Research output: Contribution to journalConference articlepeer-review

Abstract

Recent studies have started to incorporate imagery information from picture-description tasks in clinical interviews to automate Alzheimer's disease detection in the elderly. However, the high-level logical flow of visual-attention cognition mechanisms has not yet been investigated for enhanced interpretability. In this study, we systematically analyze the elements of picture-description tasks and propose a set of top-to-bottom human-interpretable features to describe the cognitive behaviors of patients, focusing on visual attention patterns, description quality, and repetition characteristics. These features achieve 85% accuracy in AD detection without specialized equipment, offering valuable insights for clinical practices and non-expert caregivers. Our results demonstrate that these high-level descriptive features, particularly those related to visual attention and the logical flow of speech, serve as effective biomarkers for AD detection.

Original languageEnglish
Pages (from-to)2043-2047
Number of pages5
JournalProceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH
DOIs
StatePublished - 2025
Event26th Interspeech Conference 2025 - Rotterdam, Netherlands
Duration: 17 Aug 202521 Aug 2025

Keywords

  • Alzheimer's Disease
  • Interpretable AI
  • Neural Additive Model
  • Visual Attention

Fingerprint

Dive into the research topics of 'Decoding Alzheimer's: Interpretable Visual and Logical Attention in Picture Description Tasks'. Together they form a unique fingerprint.

Cite this