Statistical Analysis in Scientific Research: Core Concepts
Statistical analysis is the formal machinery that separates a scientific finding from a scientific guess. This page covers the foundational concepts, structural mechanics, and classification boundaries of statistical methods as applied in research contexts — from hypothesis testing and p-values to the tradeoffs that make even experienced researchers argue at conferences. The goal is clarity on both what these tools do and where they quietly fail.
- Definition and scope
- Core mechanics or structure
- Causal relationships or drivers
- Classification boundaries
- Tradeoffs and tensions
- Common misconceptions
- Checklist or steps
- Reference table or matrix
Definition and scope
A p-value of 0.049 has launched thousands of papers. A p-value of 0.051 has buried just as many. That single decimal point illustrates why statistical analysis is simultaneously indispensable and contentious in scientific research.
Statistical analysis is the collection of mathematical methods used to collect, organize, summarize, and interpret numerical data — with the specific aim of drawing defensible conclusions under conditions of uncertainty. It operates at every stage of research design and methodology: determining sample sizes before a study begins, validating measurements during data collection, and distinguishing signal from noise when results come in.
The scope is genuinely broad. Statistical methods appear in clinical trials evaluating drug efficacy (FDA guidance on adaptive trial designs), in ecological field studies counting species populations, in particle physics experiments at CERN requiring 5-sigma confidence thresholds before announcing a discovery, and in social science surveys with nationally representative samples. The unifying principle is that all of these disciplines are making inferences from incomplete data — and statistics is the formal language for doing that honestly.
Core mechanics or structure
Statistical analysis divides into two broad operational modes: descriptive statistics and inferential statistics.
Descriptive statistics summarize what is actually in a dataset — mean, median, standard deviation, range, and distribution shape. These numbers don't generalize beyond the data at hand. A mean blood pressure of 128 mmHg in a study sample describes that sample. It says nothing, on its own, about any other population.
Inferential statistics are where the action is. The goal is to use sample data to make probabilistic claims about a larger population. The mechanics rest on three pillars:
Probability distributions — mathematical models (normal, binomial, Poisson, chi-squared, etc.) that describe how data should behave under specific assumptions. The normal distribution, for instance, is defined entirely by its mean and standard deviation.
Hypothesis testing — a formal procedure in which a null hypothesis (typically, "no effect") is tested against an alternative. The test statistic — a t-value, F-value, chi-square value, z-score — quantifies how far the observed data deviate from what the null hypothesis predicts. The p-value then expresses the probability of observing data at least as extreme as the sample, assuming the null is true. As NIST/SEMATECH's e-Handbook of Statistical Methods documents, this framework was formalized primarily by Ronald Fisher and later extended by Jerzy Neyman and Egon Pearson in the 20th century.
Confidence intervals — a range of values, typically at the 95% level, within which the true population parameter is estimated to fall. A 95% confidence interval means that if the study were repeated 100 times using the same methods, approximately 95 of those intervals would contain the true value — not that there's a 95% probability the current interval is correct.
Effect sizes — Cohen's d, Pearson's r, odds ratios — quantify the magnitude of a finding, independent of sample size. A result can be statistically significant with a trivially small effect size in a large enough sample, which is precisely why effect sizes matter as much as p-values.
Causal relationships or drivers
Statistical correlation does not establish causation — this is one of the most repeated sentences in science, and one of the least consistently applied. Understanding what statistical analysis can and cannot prove requires understanding the causal hierarchy.
Randomized controlled trials (RCTs) sit at the top of the evidence hierarchy (NIH National Library of Medicine) because random assignment to treatment and control groups neutralizes confounding variables. When confounding is controlled, statistical differences between groups carry genuine causal weight.
Observational studies — cohort studies, case-control studies, cross-sectional surveys — rely on statistical adjustments (multivariate regression, propensity score matching, instrumental variables) to approximate causal inference. These methods reduce confounding but cannot eliminate it entirely. Directed acyclic graphs (DAGs), a tool from epidemiology and computer science, have become standard in quantitative vs qualitative research design for mapping assumed causal structures before analysis begins.
Power analysis is the mechanism that determines whether a study is capable of detecting an effect if one exists. Statistical power is defined as 1 − β, where β is the probability of a Type II error (failing to detect a true effect). A conventionally acceptable power level is 0.80 — meaning an 80% chance of detecting a real effect — though high-stakes medical research often targets 0.90 or higher.
Classification boundaries
Statistical tests are not interchangeable. The choice of test depends on four characteristics of the data:
- Measurement level — nominal, ordinal, interval, or ratio
- Number of groups or variables — one-sample, two-sample, k-sample
- Independence of observations — independent groups vs. paired/repeated measures
- Distribution assumptions — parametric (assumes normality) vs. non-parametric
Parametric tests (t-tests, ANOVA, Pearson correlation, linear regression) assume that the underlying data follow a normal distribution. Non-parametric alternatives (Mann-Whitney U, Kruskal-Wallis, Spearman correlation) make no such assumption and are appropriate for ordinal data or small samples where normality cannot be verified.
Multivariate methods — principal component analysis (PCA), factor analysis, structural equation modeling (SEM), cluster analysis — handle datasets with multiple dependent or independent variables simultaneously. These techniques are standard in genomics, neuroscience, and social science, where computational and data-driven research regularly involves datasets with thousands of variables and relatively modest sample sizes.
Bayesian statistics represent an alternative classification entirely. Rather than computing a p-value from a single dataset, Bayesian methods incorporate prior probability distributions and update them with observed data to produce posterior distributions. The philosophical divide between frequentist and Bayesian approaches remains active, with different research communities showing distinct preferences.
Tradeoffs and tensions
The replication crisis in science has a statistical signature. A 2015 replication attempt of 100 psychology studies, coordinated by the Open Science Collaboration and published in Science, found that only 36 of those studies produced statistically significant results when repeated under similar conditions (Open Science Collaboration, Science, 2015, Vol. 349). The original p-value threshold of 0.05 — treating 1-in-20 false positive risk as acceptable — is part of the diagnosis.
The core tension: lowering the alpha threshold (say, to 0.005, as proposed in a 2017 Nature Human Behaviour paper signed by 72 researchers) reduces false positives but increases false negatives, requiring larger and more expensive samples. Raising statistical power costs money and time. Research funding timelines and publication incentives create structural pressure toward underpowered studies and selective reporting — a problem that pre-registration (OSF Pre-registration) and mandatory data sharing are designed to address.
Multiple comparisons present a related problem. Testing 20 hypotheses simultaneously at α = 0.05 produces, on average, 1 false positive by chance alone. Corrections like Bonferroni adjustment (dividing alpha by the number of tests) reduce false positives but at the cost of reduced sensitivity to genuine effects.
Bayesian methods sidestep some of these issues but introduce their own: the choice of prior distribution is not neutral, and different priors can produce different conclusions from identical data. This subjectivity is a feature in some frameworks and a vulnerability in others.
The scientific landscape described at /index reflects ongoing debate about which statistical standards should govern which disciplines — a debate that has no clean resolution because the answer depends partly on what kinds of errors are most costly in each domain.
Common misconceptions
"A p-value of 0.05 means there is a 95% probability the result is true." This is incorrect. The p-value is a statement about data given a null hypothesis, not about hypotheses given data. As the American Statistical Association's 2016 statement on p-values explicitly states, a p-value does not measure the probability that the null hypothesis is true.
"Statistical significance equals practical significance." A drug that lowers blood pressure by 0.3 mmHg with p < 0.001 in a trial of 50,000 patients is statistically significant and clinically meaningless. Effect size and confidence intervals carry the practical weight.
"A non-significant result means no effect exists." Absence of evidence is not evidence of absence. Underpowered studies routinely fail to detect real effects. The correct interpretation is that the data were insufficient to reject the null hypothesis at the chosen threshold.
"More data always solves statistical problems." Larger samples reduce random error but cannot fix systematic bias. A survey with a structurally unrepresentative sample produces more precisely wrong estimates as n increases. Data collection methods in research directly determine the validity ceiling that no amount of statistical processing can exceed.
Checklist or steps
The following sequence describes the standard statistical analysis workflow in empirical research, as outlined in resources like the NIST/SEMATECH e-Handbook:
- Define the research question and hypothesis — specify the null and alternative hypotheses before data collection
- Determine measurement levels — identify whether variables are nominal, ordinal, interval, or ratio
- Conduct a power analysis — calculate required sample size based on target power (typically 0.80), expected effect size, and alpha level
- Select the appropriate statistical test — match test choice to data type, distribution assumptions, and study design
- Check assumptions — verify normality, homogeneity of variance, independence of observations, and linearity as applicable to the chosen test
- Run descriptive statistics first — examine distributions, identify outliers, and verify data quality before inferential testing
- Apply corrections for multiple comparisons — use Bonferroni, Benjamini-Hochberg false discovery rate, or planned comparison frameworks where applicable
- Report effect sizes and confidence intervals alongside p-values — ASA guidelines and most major journals require this
- Pre-register the analysis plan — submit hypothesis, methods, and analysis plan to a registry (OSF, ClinicalTrials.gov) before data collection where feasible
- Conduct sensitivity analyses — test whether conclusions hold under alternative assumptions or with outliers removed
Reference table or matrix
Common Statistical Tests: Selection Matrix
| Test | Data Type | Groups | Assumptions | Typical Use |
|---|---|---|---|---|
| Independent samples t-test | Continuous | 2 independent | Normal distribution, equal variance | Comparing means of two groups |
| Paired samples t-test | Continuous | 2 paired | Normal difference scores | Pre/post measurements, matched pairs |
| One-way ANOVA | Continuous | ≥ 3 independent | Normal distribution, equal variance | Comparing means across multiple groups |
| Chi-square test | Categorical | ≥ 2 | Expected frequencies ≥ 5 | Testing association between categorical variables |
| Mann-Whitney U | Ordinal or non-normal | 2 independent | None (non-parametric) | Non-normal continuous or ordinal data |
| Pearson correlation | Continuous | N/A (bivariate) | Linearity, normality | Measuring linear association between two variables |
| Spearman correlation | Ordinal or ranked | N/A (bivariate) | Monotonic relationship | Ranked data or non-normal continuous data |
| Linear regression | Continuous (DV) | N/A | Linearity, homoscedasticity, normality of residuals | Predicting a continuous outcome from predictors |
| Logistic regression | Binary (DV) | N/A | Independence of observations | Predicting binary outcomes (yes/no, event/no event) |
| Structural equation modeling | Mixed | N/A | Large samples (n ≥ 200 typical) | Testing complex causal path models |
Error Types in Hypothesis Testing
| Error Type | Definition | Controlled By | Cost of Error |
|---|---|---|---|
| Type I (α) | Rejecting a true null hypothesis (false positive) | Alpha threshold (typically 0.05) | Acting on a non-existent effect |
| Type II (β) | Failing to reject a false null hypothesis (false negative) | Statistical power (1 − β) | Missing a real effect |
| Type III | Correctly rejecting null but for wrong reason | Study design and theory | Misleading conclusions about mechanism |