Out-of-Sample Testing
Out-of-sample testing is the process of evaluating a model's performance on data that was not used during the training or parameter optimization phase. This is the ultimate test of a strategy's viability, as it simulates how the model would perform in the real world.
By keeping a portion of historical data completely separate until the final stage, researchers can verify if the model has truly learned a signal or just memorized the noise. In the volatile environment of digital assets, out-of-sample performance is often significantly lower than in-sample performance, which is a warning sign of overfitting.
This methodology is crucial for maintaining the integrity of the development process. If a model fails to perform on out-of-sample data, it must be discarded or re-evaluated, rather than being tweaked to fit the new data.
This strict discipline prevents the developer from falling into the trap of data snooping. It provides a realistic measure of expected performance and helps manage risk by identifying potential failure modes.
Reliable trading systems are defined by their ability to maintain performance across different, unseen market segments.