From awesome-cognitive-and-neuroscience-skills
Guides power analysis for cognitive neuroscience studies with modality-specific effect sizes (behavioral, EEG, fMRI) and sample sizes from meta-analyses. For experiment design, grants, preregistrations.
npx claudepluginhub neuroaihub/awesome_cognitive_and_neuroscience_skills --plugin awesome-cognitive-and-neuroscience-skillsThis skill uses the workspace's default tool permissions.
This skill encodes domain-specific knowledge for planning adequately powered studies in cognitive science and neuroscience. It provides:
Guides sample-size planning for fMRI/EEG/MEG studies using effect-size benchmarks, simulations, and multiple-comparison adjustments. For new study design, grants, and power evaluation.
Guides statistical test selection, assumption checks, power analysis, hypothesis tests (t-tests, ANOVA, chi-square, regression, Bayesian), effect sizes, and APA-formatted reports for research data.
Selects statistical tests, interprets effect sizes and confidence intervals, conducts power analysis, verifies assumptions for quantitative research data analysis.
Share bugs, ideas, or general feedback.
This skill encodes domain-specific knowledge for planning adequately powered studies in cognitive science and neuroscience. It provides:
An AI agent needs this because generic power analysis advice (e.g., "use G*Power with d = 0.5") fails to capture the enormous variability in effect sizes across cognitive science paradigms, and because neuroimaging modalities have unique statistical considerations.
Before executing the domain-specific steps below, you MUST:
For detailed methodology guidance, see the research-literacy skill.
This skill was generated by AI from academic literature. All parameters, thresholds, and citations require independent verification before use in research. If you find errors, please open an issue.
Determine which modality and design type apply:
| Modality | Common Designs | Key Consideration |
|---|---|---|
| Behavioral | Between-groups, within-subjects, mixed | Effect sizes vary enormously by paradigm |
| EEG/ERP | Within-subjects repeated measures | Trial count matters as much as participant count |
| fMRI (task) | Within-subjects block/event-related | Whole-brain vs. ROI analysis affects power |
| fMRI (individual differences) | Correlational, between-subjects | Requires much larger N than task contrasts |
| Clinical/Developmental | Case-control, longitudinal | Recruitment constraints often limit N; adjust design |
Do not use generic benchmarks (Cohen's "small/medium/large"). Instead:
references/effect-sizes.md for a curated library organized by modality.Modality-level median effect sizes (use only when paradigm-specific estimates are unavailable):
| Modality | Median Effect Size | Source |
|---|---|---|
| Behavioral (cognitive psychology) | d = 0.40 | Brysbaert, 2019 |
| EEG/ERP component differences | d = 0.50 - 1.00 | Boudewyn et al., 2018; Clayson et al., 2019 |
| fMRI task activation | d = 0.75 - 1.00 (within-subject) | Poldrack et al., 2017 |
| fMRI brain-behavior correlation | r = 0.10 - 0.20 | Marek et al., 2022 |
| Clinical group differences | d = 0.30 - 0.80 | Leucht et al., 2015; Button et al., 2013 |
Critical warning: The median statistical power in neuroscience has been estimated at only 21% (Button et al., 2013, Nature Reviews Neuroscience). Many published effect sizes are inflated by publication bias. Always apply skepticism to effect sizes from underpowered, unreplicated studies.
Choose method based on design complexity:
Use analytic solutions via G*Power or pwr (R):
Target: 80% power (minimum) or 90% power (recommended)
Alpha: 0.05 (two-tailed unless directional hypothesis is justified)
pwr.t.test(d = effect_size, power = 0.80, sig.level = 0.05, type = "two.sample")pwr.t.test(d = effect_size_dz, power = 0.80, sig.level = 0.05, type = "paired")pwr.r.test(r = effect_size, power = 0.80, sig.level = 0.05)Use simulation-based power analysis:
references/sample-size-guide.md for worked examples.Use these as sanity checks, not replacements for formal power analysis:
| Modality | Minimum N (per group/condition) | Basis |
|---|---|---|
| Behavioral (medium effect, d ≈ 0.5) | n = 30-50 per group | Brysbaert, 2019 |
| Behavioral (small effect, d ≈ 0.2) | n = 80-100 per group | Brysbaert, 2019 |
| Behavioral (within-subjects, d_z ≈ 0.4) | n = 50-65 | Brysbaert, 2019 |
| EEG/ERP (within-subjects) | n = 25-40 | Boudewyn et al., 2018 |
| fMRI (task activation, within-subjects) | n = 30-50 | Cremers et al., 2017; Poldrack et al., 2017 |
| fMRI (individual differences / brain-behavior) | n = 100+ (ideally 200+) | Marek et al., 2022 |
| fMRI (clinical group comparison) | n = 30-50 per group | Button et al., 2013 |
| Clinical/patient studies | n = 20-30 per group (minimum) | Leucht et al., 2015 |
| Developmental (cross-sectional age groups) | n = 25-40 per age group | Mills & Tamnes, 2014 |
For preregistration and manuscripts, the power analysis section must include:
Template language:
"Based on the meta-analytic effect size of d = [X] reported by [Author, Year], a power analysis using [tool] indicated that N = [X] participants per group would be needed to detect this effect with [80/90]% power at alpha = .05 (two-tailed). Anticipating a [X]% attrition/exclusion rate, we plan to recruit N = [adjusted X]."
Using Cohen's generic benchmarks as effect size priors: Cohen (1988) himself warned these were rough guidelines. Cognitive science effects range from d = 0.1 to d = 3.0+ depending on the paradigm. Always use paradigm-specific estimates (Brysbaert, 2019).
Ignoring the distinction between d and d_z: Between-subjects Cohen's d and within-subjects d_z are not interchangeable. Within-subjects designs typically yield larger d_z due to reduced error variance. Confusing them leads to incorrect sample size estimates (Lakens, 2013).
Powering for whole-brain fMRI but reporting ROI results (or vice versa): Whole-brain analyses with multiple comparison correction require larger effects to survive thresholding. Power calculations must match the planned analysis (Mumford & Nichols, 2008).
Treating pilot effect sizes as population estimates: Pilot studies with N = 10-20 produce wildly variable effect size estimates. Apply a correction factor or use the lower bound of the CI (Albers & Lakens, 2018).
Ignoring trial count in EEG/ERP power: For ERP analyses, both participant N and trial count per condition affect statistical power. Insufficient trials per condition reduces signal-to-noise ratio regardless of participant count (Boudewyn et al., 2018; Luck, 2014).
Assuming brain-behavior correlations are large: Marek et al. (2022) demonstrated that brain-wide association studies require thousands of participants for reliable effects. Planning an fMRI individual-differences study with N = 30 is almost certainly underpowered.
| Question | Answer | Recommended Action |
|---|---|---|
| "How many subjects for a Stroop study?" | Within-subjects Stroop effect is very large (d ≈ 1.0-1.5) | N = 15-25 likely sufficient (Brysbaert, 2019) |
| "How many for an ERP study of N400?" | N400 semantic violation effect d ≈ 0.8-1.5 | N = 20-30 (Boudewyn et al., 2018) |
| "How many for fMRI brain-behavior correlation?" | True r likely 0.10-0.20 | N = 200+ minimum (Marek et al., 2022) |
| "How many for a patient vs. control comparison?" | Effects vary widely (d ≈ 0.3-0.8) | N = 30-80 per group depending on expected effect |
| "Can I use my pilot N=12 effect size?" | Pilot effect is unreliable | Use meta-analytic estimate instead; if unavailable, use lower CI bound of pilot |
See references/effect-sizes.md for the full effect size reference library and references/sample-size-guide.md for detailed sample size guidance by modality.