Logo image
Automated semantic relevance as an indicator of cognitive decline: Out-of-sample validation on a large-scale longitudinal dataset
Journal article   Open access   Peer reviewed

Automated semantic relevance as an indicator of cognitive decline: Out-of-sample validation on a large-scale longitudinal dataset

Gabriela Stegmann, Shira Hahn, Samarth Bhandari, Kan Kawabata, Jeremy Shefner, Cayla Jessica Duncan, Julie Liss, Visar Berisha and Kimberly Mueller
Alzheimer's & dementia : diagnosis, assessment & disease monitoring, Vol.14(1), pp.e12294-n/a
2022
PMID: 35229018

Abstract

Clinical Neurology Life Sciences & Biomedicine Neurosciences Neurosciences & Neurology Science & Technology
We developed and evaluated an automatically extracted measure of cognition (semantic relevance) using automated and manual transcripts of audio recordings from healthy and cognitively impaired participants describing the Cookie Theft picture from the Boston Diagnostic Aphasia Examination. We describe the rationale and metric validation. We developed the measure on one dataset and evaluated it on a large database (>2000 samples) by comparing accuracy against a manually calculated metric and evaluating its clinical relevance. The fully automated measure was accurate (r = .84), had moderate to good reliability (intra-class correlation = .73), correlated with Mini-Mental State Examination and improved the fit in the context of other automatic language features (r = .65), and longitudinally declined with age and level of cognitive impairment. This study demonstrates the use of a rigorous analytical and clinical framework for validating automatic measures of speech, and applied it to a measure that is accurate and clinically relevant.
url
https://doi.org/10.1002/dad2.12294View
Published (Version of record) Open

Metrics

1 Record Views

Details

Logo image