← Back to Learn
QUANT CONCEPTS

Why Backtesting Alone Isn't Enough

2026-03-19 | PRUVIQ Research

The Gap Between Backtest and Reality

A strategy that returns +800% in a backtest and loses 30% live is not a surprise. It’s the expected outcome when the backtest is built on flawed foundations. There are three structural problems that cause this gap — and all three are solvable with the right process.

Problem 1: Look-Ahead Bias

Look-ahead bias is the most common and most dangerous flaw in crypto backtests. It happens when the backtest uses information that wouldn’t be available at the moment the trade is made.

The trap: backtest uses the current candle's close to make a decision
Reality: at 10:01 AM, the 10:00 candle is still forming

The current candle doesn’t have a close yet. Using it for signal logic means the backtest is “seeing the future” — it knows how the candle will close before it closes. This alone can transform a losing strategy into a winner.

A concrete example: a SHORT entry triggered when current_candle.close < open. In the backtest, this looks clean. In live trading, at 10:01, there is no close — only a price moving in real time. The logic cannot be replicated.

The fix: signal logic must only use the previous completed candle. The current candle is only valid for entry price.

Problem 2: Overfitting

Overfitting happens when you tune parameters until the strategy looks perfect on historical data — then discover it fails on any data it hasn’t seen.

Test 40 parameter combinations → pick the best → "PF 3.8, 72% win rate!"
This is not a strategy. This is memorization of the past.

The backtest period has a finite set of market conditions. A strategy optimized on 6 months of 2024 data has learned patterns specific to that period — including patterns that were random noise. It will fail when market conditions shift.

Signs of overfitting:

  • Very high win rate (>70%) on a single simple signal
  • Strategy only works on a specific narrow time window
  • Tiny parameter changes cause large performance swings
  • No out-of-sample period tested

Problem 3: Ignoring Real Execution Costs

Even a properly-coded backtest with no look-ahead bias can overstate performance if it ignores trading costs:

CostTypical AmountOften Ignored?
Exchange fees0.04–0.08% per sideFrequently
Slippage0.05–0.2% per fillAlmost always
Funding rates0.01–0.1% per 8hAlmost always
Partial fillsVariableAlmost always

A strategy with backtested PF 1.6 may produce PF 1.0–1.1 after costs — barely viable. These costs compound across hundreds of trades.

The Solution: Out-of-Sample Testing

Out-of-Sample (OOS) testing is the gold standard for validating a strategy. The process:

1. Split data: first 70% = training set, last 30% = test set
2. Develop and optimize entirely on the training set
3. Run the FINAL strategy (no further changes) on the test set
4. If performance holds up → evidence of real edge
5. If performance collapses → overfitting confirmed

The OOS period must remain untouched until the strategy is finalized. Peeking at it to make adjustments defeats the purpose.

A stronger version: walk-forward testing — repeatedly retrain on a rolling window and test on the next unseen period. This mimics what actually happens in live trading.

How PRUVIQ Addresses These Problems

PRUVIQ’s simulator is built around these three issues:

On look-ahead bias: every strategy signal is evaluated on the previous completed candle. The current candle’s close, volume ratio, and direction fields are locked out of signal logic. Only entry price uses the current candle open.

On overfitting: strategy rankings run across 500+ coins and multiple market periods simultaneously. A strategy that only works on one coin or one time window ranks poorly — the diversity of the test universe filters out overfit strategies naturally.

On execution costs: every backtest includes 0.08% fees per side and 0.1% slippage per fill by default. Strategies that look good only before costs are clearly labeled. PF in the rankings reflects post-cost performance.

Key Takeaway

Backtesting is necessary but not sufficient. A backtest without look-ahead prevention, overfitting controls, and real execution costs is closer to financial fiction than research. The threshold for confidence is a strategy that survives unseen data with realistic costs applied.

Open Strategy Builder →


This is educational content. Not financial advice.


Ready to test strategies yourself?

Simulate trading strategies on 570+ coins with 2+ years of data. Free.