Why Rigorous Statistics Are Non-Negotiable in Modern Research

Reproducibility crises, retracted papers, and p-value manipulation are symptoms of the same underlying problem: researchers lack accessible tools for doing statistics correctly. Here is why that needs to change.

Tasuku Kimura April 3, 2026 2 min read Release Notes

The replication crisis has been called one of the most serious problems in modern science. Estimates suggest that more than 50% of published findings in psychology, medicine, and biology cannot be reproduced by independent labs. The causes are numerous, but one thread runs through almost every failed replication: incorrect or inadequate statistical analysis.

The Real Cost of Getting Statistics Wrong

When a clinical trial misinterprets a confidence interval, the consequences can reach patients who receive ineffective — or harmful — treatments. When an ecology paper uses the wrong test for paired samples, conservation policy may be misdirected. The stakes are real.

Yet most researchers are not statisticians. A molecular biologist who spends years mastering PCR and cell culture is not expected to also be fluent in mixed-effects models, multiple comparison corrections, and variance homogeneity checks. The existing tools — R, Python, SPSS — require substantial training and offer no guardrails against common mistakes.

What "Rigorous" Actually Means

Rigorous statistics is not about complexity. It is about:

  • Choosing the right test for your data structure and research question
  • Meeting test assumptions before drawing conclusions
  • Controlling family-wise error rate when running multiple comparisons
  • Reporting effect sizes alongside p-values
  • Being transparent about data exclusions and transformations

Each of these is a known, codified practice. The challenge is that applying them consistently requires expertise that most researchers simply do not have time to develop.

The Accessibility Gap

Software like SPSS was designed to make statistics accessible — and it did, in the 1980s. But modern research involves complex experimental designs, large datasets, and multi-level models that these tools handle awkwardly at best.

R and Python close the capability gap but open a new one: a programming literacy gap. Most researchers can follow a tutorial and run a t-test, but they cannot reliably judge whether their data meets the assumptions of that test, or whether a different approach would be more appropriate.

This is the accessibility gap: the space between what is statistically correct and what researchers can actually execute.

A New Approach: AI-Assisted Statistical Rigor

Licklider is built around a single premise: rigorous statistics should be the path of least resistance, not the path of most effort.

Instead of requiring researchers to navigate menus or write code, Licklider accepts plain-language descriptions of experimental designs and research questions. Its underlying engine — trained on decades of statistical methodology literature — selects appropriate tests, checks assumptions, applies necessary corrections, and generates publication-ready figures with accurate method descriptions.

The researcher focuses on the science. The statistics are handled correctly, every time.

Looking Forward

The tools we build shape the science we produce. If statistical analysis remains difficult, researchers will continue to reach for convenient shortcuts rather than correct methods. If we build tools that make rigor the default, we change the baseline quality of published research.

That is the problem Licklider is solving. Not by replacing statisticians, but by making their expertise available to every researcher who needs it.