Extending the usefulness of the verbal memory test: The promise of machine learning

•Evaluated a novel story recall test with 24 parallel forms that was deployed on mobile smart phones.•Employed machine learning and vector-based natural language processing methods to automate test scoring.•The validation of automated tests should test for incremental validity.•The validation of aut...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Psychiatry research 2021-03, Vol.297, p.113743-113743, Article 113743
Hauptverfasser: Chandler, Chelsea, Holmlund, Terje B., Foltz, Peter W., Cohen, Alex S., Elvevåg, Brita
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:•Evaluated a novel story recall test with 24 parallel forms that was deployed on mobile smart phones.•Employed machine learning and vector-based natural language processing methods to automate test scoring.•The validation of automated tests should test for incremental validity.•The validation of automated tests should test for criterion validity.•The validation of automated tests should test for parallel form reliability. The evaluation of verbal memory is a core component of neuropsychological assessment in a wide range of clinical and research settings. Leveraging story recall to assay neurocognitive function could be made more useful if it were possible to administer frequently (i.e., would allow for the collection of more patient data over time) and automatically assess the recalls with machine learning methods. In the present study, we evaluated a novel story recall test with 24 parallel forms that was deployed using smart devices in 94 psychiatric inpatients and 80 nonpatient adults. Machine learning and vector-based natural language processing methods were employed to automate test scoring, and performance using these methods was evaluated in their incremental validity, criterion validity (i.e., convergence with trained human raters), and parallel forms reliability. Our results suggest moderate to high consistency across the parallel forms, high convergence with human raters (r values ~ 0.89), and high incremental validity for discriminating between groups. While much work remains, the present findings are critical for implementing an automated, neuropsychological test deployable using remote technologies across multiple and frequent administrations.
ISSN:0165-1781
1872-7123
1872-7123
DOI:10.1016/j.psychres.2021.113743