Biohacking is basically a long series of guesses. Some are smart guesses. Some are expensive guesses. The problem is not guessing. The problem is thinking you proved something when you didn’t.
This is where correlation vs causation matters. If you don’t understand the difference, you will collect “wins” that are not real, build a stack that doesn’t deserve credit, and waste time chasing patterns that were never true.
Contents
Correlation Vs Causation In Plain English
Correlation means two things happen together. Example: you start taking magnesium and your sleep score improves that week.
Causation means one thing causes the other. Example: magnesium directly improved your sleep quality.
The problem is that many biohacking “results” are correlations, not proven causes.
Why Biohackers Are Especially Vulnerable To False Wins
False wins happen in many areas of life, but biohacking is a perfect storm for them:
- You change your behavior often.
- You track many metrics that naturally bounce around.
- Results are often subjective (sleep, mood, focus).
- Marketing and hype prime your expectations.
- Life itself changes (stress, travel, illness, seasons).
With that many moving parts, it is easy to accidentally assign credit to the wrong thing.
Five Common False Win Patterns
Pattern One: The Confounder
You start a supplement the same week your work stress drops. You sleep better and feel calmer. You credit the supplement. The real cause may have been lower stress.
Confounders are “hidden variables” that move outcomes. Common confounders include: alcohol, travel, illness, late meals, new workouts, relationship stress, and changes in schedule.
Pattern Two: Regression To The Mean
This is a fancy term for a simple idea: if you start experimenting when things are unusually bad, they often improve naturally over time.
Example: you try a new sleep product during a terrible week. The next week is more normal, so sleep improves. You credit the product. But the improvement may have happened anyway.
Pattern Three: The “New Routine” Effect
When you start a new intervention, you also start acting like a person who cares. You go to bed earlier. You drink less. You stop doomscrolling. Then you credit the new supplement or device.
The improvement may be real, but the cause may be your overall behavior shift, not the specific tool.
Pattern Four: Over-Tracking Noise
If you track enough metrics, something will look better by chance. A single week of higher HRV or deeper sleep can happen for random reasons. If you run many experiments and watch many numbers, you will find “proof” everywhere.
Pattern Five: Placebo And Storytelling
Belief changes perception. If you expect a new protocol to work, you may notice the good moments more and ignore the neutral ones. Over time, your memory becomes a story: “That supplement was amazing.”
How To Think Like A Skeptical Biohacker
You do not need to be paranoid. You just need a better standard for what counts as evidence.
Ask: “What Else Could Explain This?”
Before you claim a win, look for alternative explanations:
- Did my sleep schedule change?
- Did my stress level change?
- Did I change caffeine timing?
- Did I train differently?
- Did I eat later or drink alcohol?
If you can’t rule out these factors, your conclusion should be modest.
Look For Trends, Not Single Days
Single days are noisy. You can have a bad night of sleep for no clear reason. You can also have a great day after a mediocre night. Biohacking works better when you evaluate patterns over weeks.
A simple rule: do not declare victory based on one good day or one good night.
Change One Variable At A Time
This is the most important rule in self-experimentation. If you add three supplements and start a new training plan, any result is hard to interpret. If you change one thing, the signal is clearer.
Use A Baseline And A Clear Time Window
Without a baseline, you are guessing what “normal” looks like. Without a time window, your experiment becomes vague. A cleaner approach is:
- track a baseline for 10 to 14 days
- run one intervention for 14 to 30 days
- review weekly
Even simple structure prevents many false conclusions.
Use Stop Rules And Success Rules
Before you start, decide what would count as success and what would count as failure. Examples:
- Sleep Experiment Success: morning rested rating improves by 2 points on average and bedtime becomes more consistent.
- Failure: anxiety increases, sleep worsens, or side effects show up for more than a few days.
Rules keep you honest. Without them, you can always reinterpret the results.
Repeat The Experiment
The strongest personal evidence is repeatability. If something works once, it might be luck. If it works twice under similar conditions, it is more likely real.
Repeatability is also how you learn whether an intervention is fragile. If it only works when everything is perfect, it may not be worth it.
A Practical Example: The “New Supplement” Trap
Let’s say you start a new supplement for sleep.
- Week 1: you also decide to reduce screen time and go to bed earlier.
- Week 2: sleep improves.
- Conclusion: the supplement worked.
A more skeptical approach would be:
- Keep your routine stable for 2 weeks (baseline).
- Add the supplement for 2 to 4 weeks.
- Track a few metrics and review weekly.
- Stop the supplement for 1 week (washout) and observe.
- Restart if needed to see if the effect returns.
This approach is not perfect science, but it is much less likely to produce a false win.