- CAREER COLUMN
Some data practices can lead to statistically dubious findings. Here’s how to avoid them.
By
-
Benjamin Tsang
-
Benjamin Tsang is a PhD candidate in the Department of Cell and Systems Biology at the University of Toronto, Canada.
-
It can happen so easily. You’re excited about an experiment, so you sneak an early peek at the data to see if the P value — a measure of statistical significance — has dipped below the threshold of 0.05. Or maybe you’ve tried analysing your results in several different ways, hoping one will give you that significant finding. These temptations are common, especially in the cut-throat world of publish-or-perish academia. But giving in to them can lead to what scientists call P hacking.
Access options
Access Nature and 54 other Nature Portfolio journals
Get Nature+, our best-value online-access subscription
$32.99 / 30 days
cancel any time
Subscribe to this journal
Receive 51 print issues and online access
$199.00 per year
only $3.90 per issue
Rent or buy this article
Prices vary by article type
from$1.95
to$39.95
Prices may be subject to local taxes which are calculated during checkout
doi: https://doi.org/10.1038/d41586-025-01246-1
This is an article from the Nature Careers Community, a place for Nature readers to share their professional experiences and advice. Guest posts are encouraged.
Competing Interests
The author declares no competing interests.
Related Articles
-
‘Spell-checker for statistics’ reduces errors in the psychology literature
-
Smart software spots statistical errors in psychology papers
-
Retiring statistical significance would give bias a free pass
One statistical analysis must not rule them all
It’s time to talk about ditching statistical significance
Collection: Scientific data