» Articles » PMID: 39957549

Impact of ERP Reliability Cutoffs on Sample Characteristics and Effect Sizes: Performance-Monitoring ERPs in Psychosis and Healthy Controls

Overview
Specialty Psychiatry
Date 2025 Feb 17
PMID 39957549
Authors
Affiliations
Soon will be listed here.
Abstract

In studies of event-related brain potentials (ERPs), it is common practice to exclude participants for having too few trials for analysis to ensure adequate score reliability (i.e., internal consistency). However, in research involving clinical samples, the impact of increasingly rigorous reliability standards on factors such as sample generalizability, patient versus control effect sizes, and effect sizes for within-group correlations with external variables is unclear. This study systematically evaluated whether different ERP reliability cutoffs impacted these factors in psychosis. Error-related negativity (ERN) and error positivity (Pe) were assessed during a modified flanker task in 97 patients with psychosis and 104 healthy comparison participants, who also completed measures of cognition and psychiatric symptoms. ERP reliability cutoffs had notably different effects on the factors considered. A recommended reliability cutoff of 0.80 resulted in sample bias due to systematic exclusion of patients with relatively few task errors, lower reported psychiatric symptoms, and higher levels of cognitive functioning. ERP score reliability lower than 0.80 resulted in generally smaller between- and within-group effect sizes, likely misrepresenting effect sizes. Imposing rigorous ERP reliability standards in studies of psychotic disorders might exclude high-functioning patients, which raises important considerations for the generalizability of clinical ERP research. Moving forward, we recommend examining characteristics of excluded participants, optimizing paradigms and processing pipelines for use in clinical samples, justifying reliability thresholds, and routinely reporting score reliability of all measurements, ERP or otherwise, used to examine individual differences, especially in clinical research.

References
1.
Baldwin S, Larson M, Clayson P . The dependability of electrophysiological measurements of performance monitoring in a clinical sample: A generalizability and decision analysis of the ERN and Pe. Psychophysiology. 2015; 52(6):790-800. DOI: 10.1111/psyp.12401. View

2.
Hedge C, Powell G, Sumner P . The reliability paradox: Why robust cognitive tasks do not produce reliable individual differences. Behav Res Methods. 2017; 50(3):1166-1186. PMC: 5990556. DOI: 10.3758/s13428-017-0935-1. View

3.
Nosek B, Ebersole C, DeHaven A, Mellor D . The preregistration revolution. Proc Natl Acad Sci U S A. 2018; 115(11):2600-2606. PMC: 5856500. DOI: 10.1073/pnas.1708274114. View

4.
Brand A, Bradley M . The Precision of Effect Size Estimation From Published Psychological Research: Surveying Confidence Intervals. Psychol Rep. 2018; 118(1):154-170. DOI: 10.1177/0033294115625265. View

5.
Flegal K, Kit B, Graubard B . Bias in Hazard Ratios Arising From Misclassification According to Self-Reported Weight and Height in Observational Studies of Body Mass Index and Mortality. Am J Epidemiol. 2018; 187(1):125-134. PMC: 5859975. DOI: 10.1093/aje/kwx193. View