Back to Guides

Vocal Biomarkers for Cognitive Screening: What SLPs Need to Know

March 5, 202616 min readDr. Jorge C. Lucero

Key Takeaways

  • Voice reveals cognition: Acoustic and linguistic features can detect MCI with AUC up to 0.89 in recent validation studies
  • Key acoustic markers: Jitter, shimmer, speech rate, pause duration, pitch variability, and spectral contrast correlate with cognitive status
  • Still emerging: No FDA-cleared vocal biomarker tests exist yet—these are screening aids, not diagnostic tools
  • SLP opportunity: Speech pathologists are uniquely positioned to bridge voice science and cognitive assessment
  • Standardization needed: Recording protocols, compression methods, and feature extraction all affect validity

We've long known that neurological conditions affect speech. Parkinson's disease produces hypophonia and monotone speech. Stroke can cause dysarthria or aphasia. But emerging research suggests something more profound: subtle changes in voice may precede obvious cognitive symptoms by years, potentially offering a window for early intervention.

The field of vocal biomarkers has exploded in the past two years. Recent studies demonstrate that AI-driven voice analysis can detect mild cognitive impairment (MCI) and early Alzheimer's disease with remarkable accuracy—sometimes from recordings as short as 25 seconds.

For speech-language pathologists, this represents both an opportunity and a responsibility. We already conduct voice assessments. We understand acoustic analysis. We work with aging populations. But vocal biomarkers for cognition operate differently than traditional voice quality measures—and understanding those differences is essential.

What Are Vocal Biomarkers?

A vocal biomarker is a measurable characteristic of speech or voice that indicates an underlying health condition. Unlike voice quality measures (which assess phonatory function), cognitive vocal biomarkers target how the brain coordinates speech production, linguistic processing, and motor control.

The voice signal carries information at multiple levels:

Acoustic Features

Properties of the sound signal itself, extracted without understanding content:

  • • Jitter and shimmer (motor control stability)
  • • Pitch mean and variability (prosodic regulation)
  • • Spectral tilt and contrast (vocal effort)
  • • Speech rate and pause patterns (cognitive processing)
  • • Voice onset time (motor planning)

Linguistic Features

Properties requiring speech-to-text transcription and language analysis:

  • • Vocabulary diversity (semantic access)
  • • Syntactic complexity (working memory)
  • • Word-finding pauses vs. other pauses
  • • Idea density (information packaging)
  • • Lexical frequency patterns

Most vocal biomarker research combines both acoustic and linguistic features. However, pure acoustic approaches offer a significant advantage: they work across languages without requiring transcription or NLP models trained on specific languages. This makes acoustic features particularly valuable for global deployment.

The Evidence: What Recent Research Shows

Several landmark studies published in 2025-2026 have validated vocal biomarkers for cognitive screening:

Lancet Regional Health – Western Pacific (2025)

A cross-sectional study of 1,461 community-dwelling adults in Japan developed prediction models using AI-derived voice biomarkers from unstructured conversations.

Key finding: The best model combining age, sex, education, and voice biomarkers achieved an AUC of 0.89 for detecting cognitive impairment—substantially higher than previous voice-based MCI prediction models (AUC 0.74).

npj Dementia (2025)

Researchers analyzed digital voice recordings from the Craft Story Recall task in the Longitudinal Early-onset AD Study (LEADS), comparing 120 patients with 68 controls.

Key finding: Using acoustic and linguistic features, the feature-engineered model achieved AUC of 0.945 for detecting MCI, while an end-to-end deep learning model reached AUC of 0.988.

Alzheimer's & Dementia (2026)

Using the ADReSS-M dataset (237 participants), researchers examined specific acoustic features during the "Cookie Theft" picture description task.

Key finding: Spectral contrast and pitch variability showed significant associations with MMSE scores. Participants with AD demonstrated reduced vocal clarity and flattened prosodic patterns compared to healthy controls.

Important Context

These impressive numbers come with caveats. Most studies use case-control designs (known cognitively impaired vs. known healthy), which inflate accuracy compared to real-world screening populations. MCI prediction is inherently harder than dementia detection due to symptom subtlety. And no vocal biomarker test has yet received FDA clearance for clinical use.

Which Voice Parameters Matter for Cognition?

While many of the same parameters we use for voice quality assessment appear in cognitive biomarker research, they're being interpreted differently. Here's what the literature associates with cognitive status:

ParameterVoice Quality InterpretationCognitive Biomarker Interpretation
Jitter/ShimmerVocal fold irregularity, laryngeal pathologyMotor control degradation, subclinical neurodegeneration
F0 VariabilityProsodic range, expressivenessEmotional regulation, executive function
Speech RateFluency, articulationProcessing speed, cognitive load
Pause DurationRespiratory support, fluencyWord retrieval, planning, working memory
Spectral TiltBreathiness, glottal closureVocal effort, compensatory strategies
HNRVoice quality, noise componentLaryngeal coordination, neural control

Key Insight: Connected Speech vs. Sustained Vowels

While sustained vowel measures (like AVQI and ABI) dominate voice quality assessment, cognitive biomarkers rely heavily on connected speech—picture descriptions, story recalls, or free conversation. This is because temporal-prosodic features (pause patterns, speech rate, rhythmic variability) require running speech to measure, and these features reflect cognitive processing more directly than phonatory function.

The Standardization Challenge

A 2025 review in Frontiers in Digital Health identified a critical barrier: the lack of standardized protocols for vocal biomarker development. This affects everything:

Recording Conditions

Smartphone microphones, clinical-grade equipment, and professional studio setups produce different acoustic signatures. Background noise affects feature extraction. Mouth-to-microphone distance impacts amplitude measures. Without standardization, models trained in one environment may fail in another.

Audio Compression

Research shows that compression algorithms (MP3 vs. M4A vs. WMA) distort acoustic features—particularly jitter and shimmer, which are crucial for both voice quality and cognitive biomarkers. A model validated on uncompressed audio may give different results on compressed smartphone recordings.

Speech Tasks

The "Cookie Theft" picture description, Craft Story Recall, reading passages, free conversation, and verbal fluency tasks all elicit different vocal behaviors. Features extracted from one task type may not transfer to another.

Feature Extraction Software

Praat, openSMILE, Parselmouth, and proprietary tools calculate "the same" features differently. Algorithm versions matter—Praat 6.1 and Praat 6.4 may produce slightly different values. Cross-study comparisons become difficult.

The Bridge2AI-Voice Consortium is working to address these issues, developing evidence-based recommendations for recording protocols. But widespread standardization remains years away.

What This Means for Speech Pathologists

The emergence of vocal biomarkers creates unique opportunities for SLPs:

1

You Already Collect the Data

If you're doing voice assessments with acoustic analysis, you're already capturing parameters that research links to cognitive status. The infrastructure exists—the interpretation framework is what's evolving.

2

Cross-Domain Expertise

SLPs understand both voice production (acoustic side) and cognitive-communication disorders (cognitive side). This positions us to bridge disciplines that often work in silos—computer scientists developing algorithms and clinicians who understand patient populations.

3

Early Detection Opportunity

Voice clinicians see patients for voice complaints who may also be in early cognitive decline. Awareness of vocal biomarkers allows you to notice patterns that warrant further cognitive evaluation—not diagnosis, but appropriate referral.

4

Telehealth Applications

Recent smartphone validation research supports remote voice assessment. As vocal biomarkers mature, SLPs providing telehealth services may be positioned to offer cognitive screening that requires only a voice recording—dramatically expanding access to early detection.

Scope of Practice Considerations

Vocal biomarkers for cognitive screening are not yet validated for clinical diagnosis. SLPs should not use voice analysis to diagnose dementia or MCI. However, we can: (1) be aware of patterns that warrant referral; (2) contribute to research; (3) advocate for standardized protocols; and (4) prepare for future tools that may fall within expanded practice areas.

What's Available Now vs. What's Coming

Available Now

  • Standard acoustic analysis tools (Praat, PhonaLab) that measure parameters linked to cognition
  • Published research establishing correlations between voice features and cognitive status
  • Commercial research platforms (not FDA-cleared) for vocal biomarker exploration
  • Smartphone validation showing that mobile recordings can support acoustic analysis

On the Horizon

  • FDA-cleared vocal biomarker screening tools (projected 2027-2028)
  • Telehealth platform integration with real-time cognitive screening
  • Ambient listening in clinical settings (with consent) for passive monitoring
  • Standardized multi-site protocols from Bridge2AI and similar initiatives

Frequently Asked Questions

Q: Can I use my current voice analysis software for cognitive screening?

Not directly. While tools like PhonaLab and Praat measure relevant parameters, validated cutoffs for cognitive impairment don't exist yet. You can observe patterns and document changes over time, but you shouldn't interpret values as cognitive indicators without validated norms.

Q: How is this different from traditional cognitive-communication assessment?

Traditional assessment focuses on what patients say—comprehension, naming, discourse coherence. Vocal biomarkers focus on how they say it—the acoustic and temporal properties of the signal itself. Both provide complementary information.

Q: Should I be measuring speech rate and pause patterns now?

If you work with aging populations, yes. These temporal-prosodic features are clinically relevant even without cognitive biomarker claims. Documenting baseline speech rate and pause patterns provides valuable longitudinal data for any patient at risk of cognitive decline.

Q: What recording protocols should I use?

Follow evidence-based recommendations: quiet environment (<45 dB background noise), consistent device positioning (15-30 cm mouth-to-microphone), high-quality recording settings (44.1 kHz minimum, WAV/FLAC preferred over compressed formats). For cognitive biomarker research specifically, include connected speech tasks alongside sustained vowels.

Q: Are there ethical concerns?

Yes, significant ones. Voice is personally identifiable. Storing voice recordings creates privacy risks. Using recordings for purposes beyond stated consent is problematic. And premature deployment of unvalidated screening could cause anxiety from false positives or false reassurance from false negatives. Always obtain specific consent for any research use.

Bottom Line: The Voice as a Window to the Brain

  1. 1Vocal biomarkers are real—research consistently shows voice changes precede and correlate with cognitive decline
  2. 2Clinical tools aren't ready yet—no FDA-cleared products exist; current use is research and screening, not diagnosis
  3. 3Standardization is the bottleneck—recording protocols, compression, and feature extraction all need harmonization
  4. 4SLPs are uniquely positioned—we bridge voice science and cognitive-communication assessment
  5. 5Start documenting now—baseline speech rate, pause patterns, and acoustic parameters create valuable longitudinal records

The convergence of AI, digital health, and acoustic analysis is transforming what's possible. Within this decade, a smartphone voice recording may become a routine part of cognitive screening—scalable, non-invasive, and accessible. Speech pathologists who understand both the promise and limitations of vocal biomarkers will be essential to implementing these tools responsibly.

📊 Analyze Voice Acoustics with PhonaLab

While PhonaLab focuses on voice quality assessment rather than cognitive screening, the same acoustic parameters linked to cognition—jitter, shimmer, speech rate, and spectral measures—are available in our free analysis tools. Start building documentation practices that prepare you for the vocal biomarker future.

Try Free Voice Analyzer →

F0, jitter, shimmer, HNR, CPP/CPPS • PDF reports • Longitudinal tracking

⚠️ Important Disclaimer

This article is provided for educational purposes only. Vocal biomarkers for cognitive screening are an emerging research area without FDA-cleared clinical applications. The information presented should not be used to diagnose cognitive impairment or dementia. Voice analysis tools, including PhonaLab, are designed for voice quality assessment, not cognitive diagnosis. All cognitive concerns should be referred to appropriate specialists for comprehensive evaluation.

References & Further Reading

  • Kalia A, Boyer M, Fagherazzi G, Bélisle-Pipon JC, Bensoussan Y. (2025). Master protocols in vocal biomarker development to reduce variability and advance clinical precision: a narrative review. Frontiers in Digital Health, 7, 1619183.
  • Rodrigo I, Duñabeitia JA. (2025). Listening to the Mind: Integrating Vocal Biomarkers into Digital Health. Brain Sciences, 15(7), 762.
  • Awan SN, Bensoussan Y, Watts S, et al. (2025). Influence of recording instrumentation on measurements of voice in sentence contexts: use of smartphones and tablets. Frontiers in Digital Health, 7, 1610772.
  • Kiyoshige E, Ogata S, Kwon N, et al. (2025). Developing and testing AI-based voice biomarker models to detect cognitive impairment among community dwelling adults. The Lancet Regional Health – Western Pacific, 59.
  • Rezaii N, Wong B, Aisen P, et al. (2025). Voiceprints of cognitive impairment: analyzing digital voice for early detection of Alzheimer's and related dementias. npj Dementia, 1(1), 35.
  • Azami H, et al. (2026). Speech-Based Detection of Alzheimer's Disease: Leveraging Spectral Contrast and Pitch Variability as Potential Diagnostic Markers. Alzheimer's & Dementia, 21, e106182.

Dr. Jorge C. Lucero

Professor of Computer Science, University of Brasília

Dr. Lucero has 30+ years researching voice production, vocal fold dynamics, and acoustic analysis. He created PhonaLab to make professional voice analysis accessible to clinicians worldwide, and follows emerging vocal biomarker research as it intersects with traditional voice science.