Trading Systems: A New Approach to System Development and Portfolio Optimisation - Extended Summary
Author: Emilio Tomasini and Urban Jaekle | Categories: Trading Systems, Algorithmic Trading, Quantitative Finance, Backtesting, Portfolio Optimisation
About This Summary
This is a PhD-level extended summary covering all key concepts from "Trading Systems: A New Approach to System Development and Portfolio Optimisation" by Emilio Tomasini and Urban Jaekle. This summary distills the complete systematic trading development lifecycle - from initial hypothesis formation through backtesting, walk-forward analysis, Monte Carlo simulation, and multi-system portfolio construction. For AMT/Bookmap daytraders seeking to systematize their approach, this book provides the statistical rigor needed to distinguish genuine edge from noise. Every concept is examined through the lens of practical implementation, with particular attention to the dangers of overfitting and the methodologies that protect against it.
Executive Overview
"Trading Systems," published by Harriman House in 2009, occupies a unique position in quantitative trading literature. While many books present trading systems as finished products - here is the entry rule, here is the exit rule, here are the profits - Tomasini and Jaekle concern themselves with a far more fundamental question: how do you know if a trading system actually works? Not whether it worked in the past, but whether it has a genuine, statistically defensible edge that will persist into the future.
This distinction matters enormously for daytraders using tools like Bookmap and Auction Market Theory frameworks. Discretionary traders who observe order flow patterns and value area dynamics often want to systematize their observations. The temptation is to backtest an idea, see positive results, and deploy capital. Tomasini and Jaekle demonstrate, with mathematical precision, why this naive approach is almost guaranteed to fail. The historical results you see in a backtest are contaminated by look-ahead bias, survivorship bias, and most critically, by the optimization process itself. Every parameter you choose, every filter you add, every condition you tune has been selected precisely because it performed well on the historical data you are testing. This is overfitting, and it is the single greatest destroyer of trading capital in the systematic space.
The authors propose a rigorous, multi-stage methodology that treats system development as a scientific endeavor. Hypotheses must be formulated before testing. Parameters must be optimized in-sample and validated out-of-sample through walk-forward analysis. Results must be stress-tested through Monte Carlo simulation to generate probability distributions of future performance rather than single-point estimates. And finally, individual systems must be combined into diversified portfolios that reduce drawdowns and smooth equity curves through low correlation.
What elevates this book above typical system-trading guides is its intellectual honesty. The authors do not promise easy profits. They repeatedly emphasize that most systems fail, that even robust systems experience severe drawdowns, and that the statistical tools they present are necessary but not sufficient conditions for success. The book is a vaccine against wishful thinking - and for traders who survive its sobering message, it provides a genuinely powerful framework for building and maintaining mechanical strategies.
Part I: Foundations of Systematic Trading
Chapter 1: What Is a Trading System?
The book opens by establishing what separates a mechanical trading system from discretionary trading. A trading system is a complete set of rules that generates entry signals, exit signals, position sizing, and risk management without any human judgment. The rules must be explicit, unambiguous, and fully testable on historical data. This last requirement is critical: if a rule cannot be precisely defined and applied to past data, it cannot be validated.
Tomasini and Jaekle argue that systematic approaches outperform discretionary trading for most market participants, not because mechanical rules are inherently superior, but because they eliminate the behavioral biases that consistently destroy discretionary performance. Fear, greed, recency bias, anchoring, overconfidence, and loss aversion all distort human decision-making under conditions of uncertainty. A mechanical system, once validated, executes its rules regardless of the trader's emotional state.
"The market does not care about your feelings. A well-designed system ensures that your feelings do not affect your trading."
This is not to say that discretionary insight has no value. For AMT/Bookmap traders, the ability to read order flow in real time and identify absorption, iceberg orders, and aggressive market orders provides information that is difficult to encode in mechanical rules. However, the authors would argue that even discretionary traders benefit from systematic thinking - from testing whether their perceived edge is statistically real, from understanding sample size requirements, and from recognizing the difference between a genuine pattern and a pattern they have imposed on random data.
The Components of a Complete Trading System:
| Component | Description | Common Mistakes |
|---|---|---|
| Entry Rules | Conditions that must be met to initiate a position | Too many conditions; over-optimization of entry timing |
| Exit Rules | Conditions for closing a position (profit target, stop loss, time exit, signal exit) | Neglecting exits in favor of entry optimization; using only fixed stops |
| Position Sizing | How much capital to allocate per trade | Ignoring position sizing entirely; sizing too aggressively |
| Risk Management | Maximum loss per trade, per day, per portfolio | No portfolio-level risk controls; correlated exposure |
| Universe Selection | Which markets or instruments the system trades | Selecting markets after seeing which ones the system works on |
| Timeframe | Bar size and holding period | Choosing timeframes that maximize backtest results |
The authors emphasize that a system is only as strong as its weakest component. An excellent entry rule paired with a poorly designed exit rule will underperform a mediocre entry with a well-designed exit. This is because exits control risk, and risk management is the primary determinant of long-term survival.
Chapter 2: The Philosophy of System Development
Before diving into technical methodology, Tomasini and Jaekle establish a philosophical framework for thinking about markets and trading systems. They argue that markets are not perfectly efficient but neither are they easy to exploit. The efficient market hypothesis (EMH) is wrong in its strong form - persistent anomalies exist - but it is correct in its implication that most apparent patterns are noise.
This has a profound implication for system development: the prior probability that any given trading idea works is low. You should approach backtesting with the assumption that your system does not have an edge, and require strong statistical evidence to overturn that assumption. This is the scientific method applied to trading: you start with a null hypothesis (the system has no edge) and try to reject it.
The authors introduce the concept of "degrees of freedom" in trading systems. Every parameter, filter, and condition you add to a system consumes a degree of freedom. A system with ten optimizable parameters has enough flexibility to fit almost any historical dataset, regardless of whether any genuine pattern exists. Conversely, a system with two or three parameters has far less room to overfit, and positive results from such a system carry much more statistical weight.
"Simplicity is the ultimate sophistication in trading system design. Each parameter you add is a potential source of overfitting."
This principle aligns directly with the experience of AMT practitioners. The core concepts of Market Profile - value area, point of control, initial balance range - are remarkably simple. Their power comes not from complexity but from the fundamental truth they capture about how markets organize price discovery. Similarly, the most robust mechanical systems tend to be based on simple, universal market behaviors: trend persistence, mean reversion, volatility expansion and contraction.
Part II: The Development Lifecycle
Chapter 3: Design - From Hypothesis to Testable Rules
The design phase is where most system developers go wrong, and the authors devote considerable attention to establishing proper methodology. The key principle is that system design must proceed from hypothesis to test, never from observation to post-hoc rationalization.
The Correct Design Process:
- Formulate a market hypothesis based on economic logic or observed market behavior
- Translate the hypothesis into precise, programmable rules
- Select appropriate markets and timeframes for testing
- Define all parameters before seeing any backtest results
- Run the backtest and evaluate the results against statistical benchmarks
The Incorrect (But Common) Design Process:
- Scan through indicators and parameter combinations on historical data
- Find a combination that shows great profits
- Construct a narrative about why this works
- Believe the backtest results represent future performance
The difference between these two approaches determines whether you are doing science or data mining. Both processes will produce backtests with positive returns. Only the first process has any probability of producing a system that works in live trading.
The authors introduce the concept of the "design space" - the total number of possible parameter combinations for a given system. If you have three parameters, each with 20 possible values, your design space contains 8,000 combinations. If you test all of them and select the best, you have not discovered a good system - you have selected the combination that happened to align most favorably with the noise in your historical data. The probability of this alignment persisting in the future is essentially zero.
Framework 1: The System Design Quality Assessment
| Design Quality Criterion | High Quality | Low Quality |
|---|---|---|
| Hypothesis basis | Clear economic or structural logic | "It tested well" |
| Number of parameters | 2-4 | 8+ |
| Parameter sensitivity | Smooth performance surface | Jagged, spike-dependent surface |
| Market coverage | Works across related markets | Works on one market only |
| Timeframe robustness | Works on adjacent timeframes | Fails on any other timeframe |
| Regime dependence | Specified and understood | Unknown or denied |
| Sample size | 300+ trades minimum | Under 100 trades |
| Out-of-sample testing | Walk-forward validated | In-sample only |
| Simplicity | Rules fit on one page | Complex conditional logic |
For AMT/Bookmap traders looking to systematize their approach, this framework provides immediate guidance. Consider a system based on the hypothesis that "price breaks out of the initial balance range in the direction of the other-timeframe auction." This is a clear, testable hypothesis rooted in auction theory. It has a small number of parameters (IB range definition, entry offset, stop placement). It should work across liquid markets where institutional activity creates meaningful initial balance ranges. This is a high-quality design starting point.
Contrast this with a system that uses seven indicators, each with its own lookback period, combined with day-of-week filters and a volatility threshold. Even if this system produces spectacular backtest results, the probability of overfitting is extremely high because the design space is enormous and the system lacks a coherent underlying hypothesis.
Chapter 4: Backtesting - Methodology and Pitfalls
Backtesting is the process of applying a set of trading rules to historical data to evaluate performance. It is simultaneously the most useful and most dangerous tool in the system developer's arsenal. Useful because it allows you to test hypotheses without risking capital. Dangerous because it creates a false sense of certainty about future performance.
The authors identify several categories of backtesting errors:
Data Quality Issues:
- Survivorship bias: Testing only on markets that still exist, missing those that were delisted or went bankrupt
- Look-ahead bias: Using data that would not have been available at the time of the trading decision
- Data errors: Bad ticks, missing data, incorrect adjustments for splits and dividends
- Time zone inconsistencies: Mixing data from different exchanges without accounting for time differences
Methodological Issues:
- In-sample bias: Selecting parameters based on the same data used for evaluation
- Selection bias: Testing many systems and reporting only the best performer
- Transaction cost assumptions: Underestimating slippage, commissions, and market impact
- Fill assumptions: Assuming fills at prices that would not have been achievable in live trading
For Bookmap users, the last two points deserve special attention. Order flow visualization reveals that the price you see on a chart and the price you would actually receive for a trade are often different - sometimes dramatically so during fast markets or at illiquid price levels. Backtests that assume fills at the close of a bar or at exact limit prices systematically overestimate performance. The authors recommend adding at least one tick of slippage per side for liquid futures markets and considerably more for less liquid instruments.
Key Performance Metrics:
| Metric | Formula/Definition | What It Reveals | Acceptable Minimum |
|---|---|---|---|
| Net Profit | Total profits minus total losses | Absolute profitability | Depends on capital and risk |
| Profit Factor | Gross profit / Gross loss | Efficiency of the system | > 1.3 |
| Sharpe Ratio | (Mean return - Risk-free rate) / Std deviation | Risk-adjusted return | > 1.0 annualized |
| Maximum Drawdown | Largest peak-to-trough decline in equity | Worst-case pain | < 25% of account |
| Recovery Factor | Net profit / Maximum drawdown | How quickly losses are recovered | > 3.0 |
| Win Rate | Winning trades / Total trades | Percentage of profitable trades | Depends on payoff ratio |
| Payoff Ratio | Average win / Average loss | Reward-to-risk per trade | > 1.5 for trend systems |
| Trade Count | Total number of trades | Statistical significance | > 300 |
| Average Trade | Net profit / Total trades | Expected value per trade | > 2x transaction costs |
The authors stress that no single metric should be used in isolation. A system with a high Sharpe ratio but only 30 trades lacks statistical significance. A system with high net profit but a 60% maximum drawdown is practically untradeable because no human being can psychologically survive watching 60% of their equity disappear, even if they "know" the system will recover.
"A backtest is not a prediction. It is a historical simulation that, at best, provides weak evidence about future performance."
Chapter 5: Optimization - The Double-Edged Sword
Optimization is the process of selecting parameter values that maximize some performance criterion (typically net profit, Sharpe ratio, or profit factor) on historical data. It is essential because systems need parameter values, and those values should not be chosen arbitrarily. It is dangerous because it is the primary mechanism through which overfitting occurs.
The authors present a sophisticated framework for understanding optimization:
Framework 2: The Optimization Spectrum
| Optimization Level | Description | Overfitting Risk | Appropriate Use |
|---|---|---|---|
| No Optimization | Fixed, arbitrary parameters | None | Only for pure hypothesis testing |
| Coarse Optimization | Parameters tested in wide steps across broad ranges | Low | Initial viability screening |
| Standard Optimization | Parameters tested in medium steps across reasonable ranges | Medium | In-sample parameter selection for walk-forward |
| Fine Optimization | Parameters tested in tiny steps across narrow ranges | High | Almost never appropriate |
| Exhaustive Optimization | Every possible combination tested | Very High | Data mining; not system development |
| Genetic Algorithm | Evolutionary search through parameter space | Variable | Large design spaces where exhaustive search is impractical |
The critical concept the authors introduce is the "parameter surface" or "parameter landscape." Instead of looking at the single best parameter combination, you should visualize performance across the entire parameter space. If the best parameters sit on a sharp peak surrounded by poor performance, the system is fragile and almost certainly overfitted. If the best parameters sit in a broad plateau where nearby combinations also perform well, the system is robust.
This concept maps directly onto how AMT/Bookmap traders think about value areas. A sharp peak in a parameter surface is analogous to a single-print spike in a Market Profile - it represents an extreme that is unlikely to be revisited. A broad plateau is analogous to a well-developed value area - it represents a consensus that is likely to persist.
The Parameter Surface Test:
- Select your optimal parameters from the backtest
- Vary each parameter by +/- 10%, 20%, and 30%
- Re-run the backtest for each variation
- If performance degrades gracefully (slow, smooth decline), the system is robust
- If performance collapses (sharp, discontinuous decline), the system is fragile
The authors also address the problem of multiple testing. If you test 1,000 parameter combinations, the best one will look impressive even if the underlying system has zero edge. This is the multiple comparisons problem from statistics, and it requires adjustment. The simplest adjustment is the Bonferroni correction: divide your significance threshold by the number of tests performed. If you require p < 0.05 normally and you tested 100 combinations, you should now require p < 0.0005.
"Optimization does not find the best system. Optimization finds the parameter combination that was best on the specific historical data you tested. These are very different things."
Part III: Validation and Robustness
Chapter 6: Walk-Forward Analysis - The Gold Standard
Walk-forward analysis (WFA) is the centerpiece of the book's methodology and represents the authors' most important contribution to practical system development. WFA is a rolling out-of-sample testing procedure that simulates how a system would actually be developed and deployed in real time.
How Walk-Forward Analysis Works:
- Divide your historical data into multiple segments
- Take the first N segments as your "in-sample" (IS) optimization window
- Optimize the system's parameters on the IS window
- Apply the optimized parameters to the next segment - the "out-of-sample" (OOS) window - without any further adjustment
- Record the OOS performance
- Slide the window forward by one segment
- Repeat the optimization-then-test cycle until all data is consumed
- Concatenate all OOS results to form the walk-forward equity curve
The walk-forward equity curve is the closest approximation to how the system would have actually performed if you had developed and deployed it in real time. It is always worse than the fully optimized in-sample equity curve, often dramatically so. This difference is the "overfitting tax" - the portion of backtest profits that exist only because of curve-fitting.
Walk-Forward Analysis Parameters:
| WFA Parameter | Description | Recommended Range |
|---|---|---|
| In-Sample Window | Period used for optimization | Long enough for 100+ trades |
| Out-of-Sample Window | Period used for validation | 20-30% of IS window length |
| Re-optimization Frequency | How often parameters are refreshed | Every OOS window (anchored or rolling) |
| Optimization Criterion | Metric being maximized during IS optimization | Sharpe ratio or profit factor preferred |
| Anchored vs. Rolling | Whether IS window start is fixed or slides forward | Rolling preferred for adaptive systems |
The Walk-Forward Efficiency Ratio:
The key metric from WFA is the Walk-Forward Efficiency (WFE) ratio:
WFE = (Annualized OOS return) / (Annualized IS return)
A WFE of 50% means the system retained half its in-sample performance when tested out-of-sample. The authors consider a WFE above 50% to be strong evidence of a genuine edge, while a WFE below 30% suggests substantial overfitting.
Interpreting Walk-Forward Results:
| WFE Range | Interpretation | Action |
|---|---|---|
| > 70% | Very robust; minimal overfitting | High confidence for deployment |
| 50-70% | Good robustness; some optimization benefit lost | Acceptable for deployment with monitoring |
| 30-50% | Moderate overfitting present | Requires investigation; may need simplification |
| 10-30% | Significant overfitting | Likely not a genuine edge; redesign needed |
| < 10% | Severe overfitting; in-sample performance is almost entirely fictional | Discard the system |
| Negative | System loses money out-of-sample | Strong evidence of no edge; definitely discard |
The authors provide extensive examples showing how systems with spectacular in-sample performance - 100%+ annual returns, Sharpe ratios above 3.0, smooth equity curves - can produce flat or negative walk-forward equity curves. These examples serve as sobering reminders that in-sample performance means essentially nothing without walk-forward validation.
For AMT/Bookmap traders, the walk-forward concept has a natural analogy. When you identify a pattern in order flow - say, large resting bid orders at the bottom of the value area that lead to bounces - you are implicitly performing in-sample observation. The pattern "worked" in the data you observed. Walk-forward thinking asks: will this pattern continue to work tomorrow, next week, next month? And more critically: how would you test this rigorously rather than relying on confirmation bias and selective memory?
Chapter 7: Monte Carlo Simulation
Walk-forward analysis tells you whether a system has an edge. Monte Carlo simulation tells you how that edge might manifest in the future. Specifically, it generates probability distributions of key performance metrics by resampling the system's historical trades.
The Monte Carlo Process:
- Take the system's historical trade results (the sequence of individual trade P&Ls)
- Randomly reshuffle the order of these trades
- Calculate the equity curve, maximum drawdown, final equity, etc. from the reshuffled sequence
- Repeat steps 2-3 thousands of times (typically 5,000-10,000 iterations)
- Analyze the distribution of results
The key insight is that the specific sequence of trades you experienced in your backtest is only one of many possible sequences. If you happened to have your largest winning trades early and your largest losing trades late, your maximum drawdown was understated. A different ordering might have produced the largest losses first, creating a much deeper drawdown.
What Monte Carlo Simulation Reveals:
| Output | Description | Why It Matters |
|---|---|---|
| Drawdown distribution | Range of possible maximum drawdowns at various confidence levels | The historical drawdown is the minimum you should expect; the 95th percentile is what you should plan for |
| Return distribution | Range of possible final equity values | Shows how wide the range of outcomes truly is |
| Ruin probability | Percentage of simulations where the system hit a predefined loss threshold | Quantifies the risk of catastrophic loss |
| Recovery time distribution | How long drawdown recovery takes | Critical for psychological and capital management |
| Confidence intervals | Upper and lower bounds on performance metrics | Replaces point estimates with realistic ranges |
The authors emphasize a critical point: the maximum drawdown shown in a backtest is almost certainly an underestimate of the maximum drawdown you will experience in live trading. Monte Carlo simulation typically reveals that the 95th percentile drawdown is 1.5x to 2.5x the historical maximum. This means that if your backtest shows a 20% maximum drawdown, you should plan for drawdowns of 30-50% with reasonable probability.
"Monte Carlo simulation replaces the dangerous illusion of a single backtest result with the sobering reality of a probability distribution."
This has direct implications for position sizing and risk management. If you size your positions based on the historical maximum drawdown, you will almost certainly experience a drawdown that exceeds your planning threshold. The authors recommend sizing positions based on the 95th or 99th percentile Monte Carlo drawdown estimate, which typically means trading at 40-60% of the size suggested by the historical backtest.
Framework 3: The Monte Carlo Risk Assessment Framework
| Risk Parameter | How to Calculate | Acceptable Threshold | AMT/Bookmap Application |
|---|---|---|---|
| 95th Percentile Drawdown | Monte Carlo 95th percentile MDD | < 30% of account equity | Sets position size ceiling for systematized order flow strategies |
| 99th Percentile Drawdown | Monte Carlo 99th percentile MDD | < 50% of account equity | Determines margin of safety for worst-case scenarios |
| Probability of Ruin | % of simulations hitting -50% | < 1% | Must be near-zero for any viable system |
| Expected Recovery Time | Median time to recover from 95th percentile DD | < 12 months | Psychological survival threshold |
| Return/Risk Ratio | Median return / 95th percentile drawdown | > 2.0 | Ensures positive asymmetry |
| Minimum Trade Count | Number of trades needed for stable Monte Carlo output | > 200 | Fewer trades make Monte Carlo unreliable |
Chapter 8: Additional Robustness Tests
Beyond walk-forward analysis and Monte Carlo simulation, the authors describe several supplementary robustness tests:
8.1 Market Robustness
A system that only works on one market is suspect. The authors recommend testing any system across a basket of related markets. A trend-following system for crude oil should also show positive results on natural gas, heating oil, and other energy futures. This does not mean the system must be profitable on every market, but the average performance across markets should be positive.
8.2 Timeframe Robustness
If a system works on 15-minute bars but fails completely on 10-minute or 20-minute bars, it is likely overfitted to the specific noise patterns of the 15-minute timeframe. Robust systems show positive performance across a range of timeframes, even if the optimal timeframe produces the best results.
8.3 Entry Point Robustness
One of the most elegant robustness tests is to introduce random delays or random entry points. If you delay the entry by a random number of bars (say 0-5 bars), how much does performance degrade? If the answer is "dramatically," the system's edge is entirely dependent on precise entry timing, which is unlikely to be replicable in live trading due to slippage and execution delays.
8.4 Direction Robustness
Does the system work on both the long and short sides? If a system only works long, it may be capturing the underlying upward drift of equity markets rather than any genuine pattern.
8.5 Regime Analysis
Markets cycle between trending and range-bound regimes, between high and low volatility, between risk-on and risk-off. A robust system should be tested against these different regimes to understand when it makes money and when it loses money. The authors argue that understanding your system's regime dependence is critical for capital allocation and risk management.
For AMT traders, regime analysis is particularly relevant. The balance-to-imbalance cycle that Market Profile identifies is precisely the kind of regime transition that determines whether trend-following or mean-reversion systems perform well. During balance (range) regimes, mean-reversion systems thrive and trend-following systems get chopped up. During imbalance (trending) regimes, the opposite is true. Understanding this relationship allows traders to allocate capital more intelligently.
Part IV: Trading System Examples
Chapter 9: Trend-Following Systems
The authors present several complete trend-following systems, beginning with simple moving average crossover strategies and progressing to channel breakout systems. The key lesson is not that any particular system is the "best" but rather how the development and validation methodology applies to each.
Moving Average Crossover System:
The simplest trend-following approach: go long when a short-term moving average crosses above a long-term moving average, and go short when the reverse occurs. The authors demonstrate that while this system generates positive returns on many markets over long time periods, its parameter surface is relatively sensitive. Small changes in the moving average lengths can produce dramatically different results, which is a warning sign for overfitting.
Channel Breakout System (Donchian-style):
Buy when price makes a new N-period high; sell when price makes a new M-period low. This system tends to have a much smoother parameter surface than moving average systems, which the authors attribute to its simpler underlying logic. The channel breakout system captures the fundamental trend-following premise - that markets tend to continue in the direction of their most recent extreme - without the lag problems inherent in moving averages.
Volatility Breakout System:
Enter when price moves a specified multiple of recent volatility (typically Average True Range) from the previous close. This system adapts to changing market conditions because the entry threshold scales with volatility. During quiet markets, the threshold is tight; during volatile markets, it widens. The authors show this to be one of the most robust trend-following approaches because it is self-adapting and has very few parameters.
Comparison of Trend-Following Approaches:
| System Type | Parameters | Robustness | Win Rate | Payoff Ratio | Best Market Regime |
|---|---|---|---|---|---|
| MA Crossover | 2 (fast/slow MA periods) | Moderate | 35-40% | 2.0-3.0 | Strong, sustained trends |
| Channel Breakout | 2 (entry/exit lookback) | High | 30-35% | 2.5-4.0 | Markets making new highs/lows |
| Volatility Breakout | 2 (ATR period/multiplier) | Very High | 40-45% | 1.5-2.5 | Volatility expansion from compression |
| Triple MA | 3 (three MA periods) | Low-Moderate | 35-40% | 2.0-2.5 | Steadily trending markets |
All trend-following systems share common characteristics: low win rates (typically 30-45%) compensated by high payoff ratios (average winners much larger than average losers), extended drawdown periods during range-bound markets, and strong performance during sustained directional moves.
Chapter 10: Mean-Reversion Systems
Mean-reversion systems operate on the opposite premise: when price has moved too far from its mean, it will tend to revert. The authors present systems based on RSI extremes, Bollinger Band touches, and distance from moving averages.
RSI Mean-Reversion System:
Buy when RSI drops below an oversold threshold; sell when RSI rises above an overbought threshold. The authors demonstrate that this approach works well in range-bound, high-liquidity markets but fails catastrophically during trends. The walk-forward analysis reveals that the system's performance is highly regime-dependent, reinforcing the need for understanding when to deploy mean-reversion versus trend-following strategies.
Bollinger Band Mean-Reversion:
Buy when price touches the lower Bollinger Band; sell when it touches the upper band (or reverts to the mean). Similar characteristics to the RSI system, with the advantage of incorporating volatility through the band width.
The authors make an important observation about mean-reversion versus trend-following: these are not competing approaches but complementary ones. A well-constructed portfolio combines both types, because the regimes where one loses money tend to be the regimes where the other makes money. This is portfolio-level thinking, which the authors develop fully in the final section of the book.
For Bookmap and order flow traders, mean-reversion has a natural interpretation in terms of absorption. When the tape shows large resting limit orders absorbing aggressive selling at a price level, this is a mean-reversion signal - the market is being prevented from continuing lower by responsive buyers. Conversely, when aggressive sellers are overwhelming the bid stack and the market is moving lower with increasing delta, this is a trend-continuation signal. The system developer's challenge is to quantify these observations in testable rules.
Chapter 11: Breakout Systems
Breakout systems sit at the intersection of trend-following and mean-reversion. They wait for price to break out of a defined range and then enter in the direction of the breakout, anticipating that the breakout represents the beginning of a new trend.
The authors present systems based on:
- Opening Range Breakout: Trading the breakout from the first N minutes of the session. This has direct relevance for AMT traders because the opening range is closely related to the Initial Balance (IB) concept from Market Profile.
- Volatility Compression Breakout: Waiting for a period of unusually low volatility (narrow Bollinger Bands, low ATR, narrow ranges) and then entering when volatility expands. This exploits the well-documented tendency of volatility to cluster and mean-revert.
- Pattern Breakout: Trading breakouts from specific chart patterns such as triangles, rectangles, and flags.
The key finding from the authors' testing is that opening range and volatility compression breakouts tend to be more robust than pattern breakouts. Pattern breakouts require more parameters to define the pattern, increasing overfitting risk. Volatility compression breakouts have few parameters and capture a fundamental market dynamic: periods of low volatility compress energy that is eventually released in a directional move.
"Breakout systems are conceptually simple and difficult to over-optimize, which is precisely why they tend to be robust."
Part V: Portfolio Construction and Optimisation
Chapter 12: The Portfolio of Systems Approach
The final major section of the book addresses what the authors consider the most important and most neglected aspect of systematic trading: portfolio construction. Most system developers focus all their energy on finding the single best system. Tomasini and Jaekle argue that this is suboptimal. A portfolio of mediocre-but-uncorrelated systems will consistently outperform a single "great" system, just as a diversified stock portfolio outperforms any single stock on a risk-adjusted basis.
Why Portfolio Thinking Matters:
- Drawdown reduction: When one system is in drawdown, uncorrelated systems are likely still performing, reducing the portfolio-level drawdown
- Smoother equity curve: The combined equity curve has less variance than any individual system
- Psychological sustainability: Smaller, shorter drawdowns are psychologically survivable; single-system 40% drawdowns are not
- System failure insurance: If one system stops working (edge decays), the portfolio survives on the remaining systems
- Capital efficiency: Capital can be shared across systems that are not simultaneously deployed
Correlation Analysis Between Systems:
The authors provide detailed methodology for measuring correlation between systems. The key metric is the equity curve correlation - the correlation between the daily (or trade-by-trade) returns of different systems. Systems with low or negative equity curve correlation provide the best diversification benefit.
| Correlation Range | Diversification Benefit | Portfolio Impact |
|---|---|---|
| -1.0 to -0.5 | Excellent (rare to find) | Dramatically reduces drawdowns; may reduce returns |
| -0.5 to 0.0 | Very Good | Significant drawdown reduction with minimal return sacrifice |
| 0.0 to 0.3 | Good | Meaningful diversification benefit |
| 0.3 to 0.6 | Moderate | Some diversification benefit but limited |
| 0.6 to 0.9 | Poor | Systems are too similar; acts almost like one system |
| 0.9 to 1.0 | None | Redundant systems; no diversification |
Sources of Low Correlation:
- Different market types: A system trading equity indices and one trading agricultural commodities will typically have low correlation
- Different system types: A trend-following system and a mean-reversion system on the same market will be negatively correlated during regime transitions
- Different timeframes: A short-term system (holding minutes to hours) and a long-term system (holding days to weeks) on the same market will have low correlation
- Different entry logic: Systems based on different premises (momentum vs. value, breakout vs. fade) generate different trade sequences
Chapter 13: Capital Allocation Across Systems
Once you have a portfolio of systems, you must decide how to allocate capital among them. The authors present several approaches:
Equal Allocation: Give each system the same capital. Simple and effective, avoids the overfitting risk inherent in optimized allocation.
Risk-Parity Allocation: Allocate capital inversely proportional to each system's volatility, so that each system contributes equally to portfolio risk. This prevents a single volatile system from dominating portfolio performance.
Performance-Weighted Allocation: Allocate more capital to systems with better recent risk-adjusted performance. The authors caution against this approach because it is a form of optimization that is subject to the same overfitting risks as parameter optimization.
Optimal f and Kelly Criterion: Mathematical frameworks for determining the optimal fraction of capital to risk per trade. The Kelly Criterion maximizes the geometric growth rate of capital. However, full Kelly is extremely aggressive and produces large drawdowns. The authors recommend using half-Kelly or quarter-Kelly to reduce risk.
Framework 4: Capital Allocation Decision Matrix
| Allocation Method | Complexity | Overfitting Risk | Drawdown Management | Best For |
|---|---|---|---|---|
| Equal Weight | Very Low | None | Moderate | Beginners; small portfolios |
| Risk Parity | Low | Minimal | Good | Most traders; default recommendation |
| Performance-Weighted | Medium | Moderate | Variable | Experienced practitioners with long track records |
| Mean-Variance (Markowitz) | High | High | Theoretically optimal | Academic research; rarely practical |
| Kelly Criterion | Medium | Low (formula-based) | Poor (aggressive) | Theoretical benchmark; use fractional Kelly |
| Fractional Kelly (1/4 to 1/2) | Medium | Low | Good | Practical capital sizing for live trading |
The authors strongly recommend starting with equal weight or risk parity and only moving to more sophisticated approaches after accumulating significant live trading experience with the portfolio. Optimizing capital allocation introduces the same dangers as optimizing system parameters: you are fitting to historical data and the optimal allocation in the past is unlikely to be optimal in the future.
Part VI: Implementation and Practical Considerations
Chapter 14: Going Live - From Backtest to Real Trading
The transition from backtesting to live trading is where most system traders encounter their first major crisis. The authors identify several reasons why live performance almost always underperforms backtested performance:
The Live Trading Degradation Factors:
- Slippage: Real fills are worse than backtest fills, especially in fast markets and at illiquid price levels
- Commission and fees: Often underestimated in backtests, especially for high-frequency systems
- Partial fills: Limit orders may not be completely filled; backtests assume full fills
- Timing delays: Signal generation, transmission, and execution all take time
- System failure: Technology breaks - internet outages, broker platform crashes, data feed interruptions
- Psychological interference: The trader overrides signals, sizes down after losses, sizes up after wins
- Market regime change: The market shifts to a regime where the system underperforms
- Edge decay: The market inefficiency the system exploits may diminish over time as other participants discover it
The authors recommend a structured approach to going live:
Phase 1: Paper Trading (2-4 weeks) Run the system on live data in simulation mode. Compare the simulated results to what the backtest would have predicted for this same period. Identify any discrepancies.
Phase 2: Small Size Trading (1-3 months) Trade with one contract or minimum position size. The goal is to verify execution quality, identify operational issues, and build psychological comfort with the system's drawdown profile.
Phase 3: Scaled Trading (ongoing) Gradually increase position size toward the target allocation. Monitor performance against Monte Carlo confidence intervals. If performance falls below the 5th percentile of Monte Carlo simulations, investigate whether the edge has decayed.
Chapter 15: System Monitoring and Maintenance
A deployed trading system is not a "set and forget" investment. The authors emphasize continuous monitoring:
When to Stop a System:
This is one of the hardest decisions in systematic trading. If you stop a system during a normal drawdown, you give up the recovery. If you continue running a system whose edge has decayed, you hemorrhage capital. The authors provide a framework based on Monte Carlo analysis:
- Use your Monte Carlo simulation to establish drawdown confidence intervals
- If the current live drawdown exceeds the 95th percentile Monte Carlo estimate, stop the system and investigate
- If investigation reveals no operational issues and market conditions appear consistent with the system's design, consider restarting with reduced size
- If the drawdown exceeds the 99th percentile estimate, stop the system entirely and treat the edge as potentially decayed
Re-optimization Schedule:
Walk-forward analysis implicitly determines when parameters should be re-optimized. The OOS window length from your WFA defines the re-optimization cycle. If your WFA used 6-month OOS windows, you should re-optimize every 6 months. The authors caution against re-optimizing too frequently, as this increases the risk of chasing noise.
Part VII: Critical Analysis and AMT/Bookmap Integration
The Book's Strengths
1. Intellectual Honesty The single greatest strength of "Trading Systems" is its refusal to oversell. The authors repeatedly state that most systems do not work, that backtests are unreliable, and that even validated systems will underperform expectations. This honesty is rare in trading literature and immensely valuable.
2. Statistical Rigor The walk-forward analysis methodology is presented with sufficient mathematical detail for implementation while remaining accessible to non-statisticians. The Monte Carlo framework is equally well-developed.
3. Portfolio-Level Thinking Most system-trading books stop at the single-system level. Tomasini and Jaekle's treatment of multi-system portfolios, correlation analysis, and capital allocation elevates the book beyond its peers.
4. Complete Code Examples The authors provide full TradeStation/EasyLanguage code for all systems discussed, making the book immediately practical.
The Book's Weaknesses
1. Software Dependence The code examples use TradeStation's EasyLanguage, which limits their utility for traders using other platforms. Python or pseudocode would have broader applicability.
2. Market Structure Evolution Published in 2009, the book predates many significant market structure changes: the rise of high-frequency trading, the crypto market, the proliferation of ETFs as trading instruments, and the shift to near-zero commissions. While the methodology is timeless, some specific system examples may need updating.
3. Limited Treatment of Execution For daytraders, execution quality is often the difference between a profitable and unprofitable system. The book's treatment of slippage and market impact is adequate but not detailed enough for intraday strategies where these factors dominate.
4. No Order Flow Discussion The book focuses entirely on price-based systems. There is no discussion of order flow, volume profile, delta, or other microstructure data that Bookmap and AMT traders rely on. This is a significant gap for the target audience of this platform.
5. Underemphasis on Transaction Costs for Short-Term Systems While the authors mention transaction costs, they do not fully explore how costs scale with trading frequency. For daytraders executing multiple round-trips per session, transaction costs (including slippage) can easily consume the entire theoretical edge.
Integration with AMT/Bookmap Trading
For traders using Auction Market Theory and Bookmap, this book's methodology provides a critical missing piece: the statistical validation framework. AMT gives you the conceptual framework for understanding market structure. Bookmap gives you the visualization tools for seeing order flow in real time. But neither tells you whether the patterns you observe constitute a statistically valid edge. Tomasini and Jaekle's methodology does.
Systematizing AMT Concepts:
| AMT Concept | Potential Systematic Rule | Testability |
|---|---|---|
| Value Area Rejection | Short when price enters prior VA and reverses within N bars | High - value area is precisely defined |
| Initial Balance Breakout | Buy/sell the breakout from the first 60-minute range | High - IB is precisely defined |
| Single Print Fill | Fade moves into single-print zones from prior profile | Medium - requires profile database |
| Poor High/Low | Fade poor highs/lows (lacking excess) | Medium - "poor" requires definition |
| Balance-to-Imbalance Transition | Enter when price exits N-day balance area | High - balance area measurable |
| Excess Identification | Use excess (long tails) as stop placement reference | Medium - excess quantification varies |
The challenge is that many AMT concepts resist precise quantification. "Other-timeframe buying" is a concept that experienced tape readers can identify but that is difficult to encode in a backtest. However, proxies exist: large delta divergences, absorption patterns at key levels, and volume profile shape characteristics can all be programmed and tested. The book's methodology then provides the framework for determining whether these proxies capture real alpha.
Key Quotes Collection
"The biggest enemy of the system trader is overfitting. If you do not rigorously guard against it, you are guaranteed to fail."
"A robust system should work across a range of parameters, not just the optimal ones."
"Walk-forward analysis is the closest thing we have to a crystal ball for trading systems."
"Simplicity is the ultimate sophistication in trading system design. Each parameter you add is a potential source of overfitting."
"The market does not care about your feelings. A well-designed system ensures that your feelings do not affect your trading."
"A backtest is not a prediction. It is a historical simulation that, at best, provides weak evidence about future performance."
"Optimization does not find the best system. Optimization finds the parameter combination that was best on the specific historical data you tested. These are very different things."
"Monte Carlo simulation replaces the dangerous illusion of a single backtest result with the sobering reality of a probability distribution."
"Breakout systems are conceptually simple and difficult to over-optimize, which is precisely why they tend to be robust."
"A portfolio of mediocre but uncorrelated systems will outperform a single 'great' system, just as a diversified stock portfolio outperforms any single stock on a risk-adjusted basis."
Trading Takeaways for AMT/Bookmap Daytraders
Immediate Action Items
-
Quantify your edge. If you trade based on order flow observations, keep a detailed log of every signal, every entry, every exit. After 100+ trades, analyze the results statistically. Is your win rate and payoff ratio combination actually profitable after transaction costs?
-
Test your hypotheses before trading them. If you believe that "price rejects the previous day's value area high" is a tradeable pattern, define the rules precisely and backtest them. Use walk-forward analysis to validate.
-
Size positions based on Monte Carlo worst-case estimates, not backtest averages. If your backtest shows a 15% maximum drawdown, plan for 25-35%.
-
Build a portfolio of approaches. Do not rely on a single strategy. Combine mean-reversion (fade extremes back to value) with trend-following (ride breakouts from balance) and breakout strategies (trade IB extensions).
-
Re-evaluate your systems periodically. Use the walk-forward re-optimization schedule. Do not change systems during drawdowns unless the drawdown exceeds your Monte Carlo confidence intervals.
Longer-Term Development Path
-
Learn to program. The ability to backtest rigorously requires programming skills. Python with libraries like Backtrader, Zipline, or QuantConnect is the modern standard.
-
Study statistics. Understanding hypothesis testing, confidence intervals, multiple comparisons, and probability distributions is essential for serious system development.
-
Keep a system journal. Document every system you develop, including the hypothesis, the parameters, the walk-forward results, and the reason for deployment or rejection. This prevents repeating mistakes.
-
Accept that most ideas will not work. The scientific method requires many failed experiments for each success. Do not become emotionally attached to any system idea.
-
Focus on risk management before return generation. The systems that survive are the ones that manage drawdowns, not the ones that maximize returns.
System Development Checklist
Use this checklist before deploying any mechanical trading system with real capital:
Design Phase
- The system is based on a clear, articulable market hypothesis
- The hypothesis has economic or structural logic (not just "it tested well")
- The system has 4 or fewer optimizable parameters
- Entry rules, exit rules, position sizing, and risk management are all fully specified
- The system can be described in plain language in under 60 seconds
Backtesting Phase
- Data quality has been verified (no survivorship bias, look-ahead bias, or bad ticks)
- Transaction costs include realistic slippage estimates (not just commissions)
- The backtest produces at least 300 trades for statistical significance
- The average trade profit exceeds 2x the total transaction cost per round trip
- The profit factor exceeds 1.3
- The maximum drawdown is psychologically and financially survivable
Optimization Phase
- The parameter surface is smooth (broad plateau, not sharp peak)
- Performance degrades gracefully when parameters are varied by +/- 20%
- The number of parameter combinations tested is documented
- Multiple comparison adjustments have been considered
Walk-Forward Validation Phase
- Walk-forward analysis has been performed with appropriate IS/OOS splits
- The Walk-Forward Efficiency ratio exceeds 50%
- The walk-forward equity curve is positive and relatively smooth
- Multiple walk-forward configurations (different window lengths) produce consistent results
Monte Carlo Phase
- Monte Carlo simulation has been run with at least 5,000 iterations
- The 95th percentile drawdown has been calculated and is acceptable
- The probability of ruin (50% drawdown) is below 1%
- Position sizing is based on Monte Carlo drawdown estimates, not backtest drawdown
Robustness Phase
- The system works on at least 2-3 related markets
- The system works on adjacent timeframes
- Random entry delay does not destroy performance
- The system's regime dependence is understood and documented
Portfolio Phase
- The system's correlation with existing portfolio systems has been measured
- The system adds diversification benefit (correlation < 0.5 with existing systems)
- Capital allocation follows risk parity or equal weight methodology
- Total portfolio risk (Monte Carlo 95th percentile drawdown) is acceptable
Deployment Phase
- Paper trading period of 2-4 weeks has been completed successfully
- Small-size live trading period of 1-3 months has been completed
- Drawdown monitoring thresholds based on Monte Carlo have been established
- Re-optimization schedule has been defined
- System shutdown criteria have been pre-committed
Comparison With Related Works
| Feature | Tomasini & Jaekle: Trading Systems | Perry Kaufman: Trading Systems and Methods | Robert Pardo: The Evaluation and Optimization of Trading Strategies | Kevin Davey: Building Winning Algorithmic Trading Systems |
|---|---|---|---|---|
| Primary Focus | Complete development lifecycle with statistical rigor | Encyclopedia of trading methods | Walk-forward analysis deep dive | Practical system building for retail traders |
| Statistical Depth | Very High | High | Very High | Moderate |
| Walk-Forward Coverage | Comprehensive | Mentioned | Definitive treatment | Good practical coverage |
| Monte Carlo Coverage | Strong | Limited | Good | Good |
| Portfolio Construction | Excellent | Good | Limited | Limited |
| Code Language | EasyLanguage | Multiple/Pseudocode | TradeStation | TradeStation and MultiCharts |
| Accessibility | Graduate level | Reference manual | Graduate level | Intermediate |
| Publication Year | 2009 | 2013 (5th ed.) | 2008 (2nd ed.) | 2014 |
| Best For | Rigorous developers wanting complete methodology | Researchers needing comprehensive indicator reference | Specialists in validation methodology | Beginners wanting step-by-step guide |
| AMT Relevance | Moderate - methodology applicable to any edge | Low - indicator-centric | Moderate - validation applies universally | Low - specific strategies not AMT-related |
Further Reading
Essential Companion Texts
-
"The Evaluation and Optimization of Trading Strategies" by Robert Pardo - The definitive deep dive on walk-forward analysis. Where Tomasini and Jaekle provide the complete lifecycle, Pardo goes into exhaustive detail on the validation phase specifically. Read this after mastering the walk-forward concepts in Trading Systems.
-
"Evidence-Based Technical Analysis" by David Aronson - Applies the scientific method to technical analysis with rigorous statistical testing. Complements Tomasini and Jaekle's approach with deeper coverage of hypothesis testing and data mining bias.
-
"Quantitative Trading" by Ernest Chan - Practical guide to implementing systematic strategies, with strong coverage of execution, capital management, and the practical realities of running automated systems.
-
"Building Winning Algorithmic Trading Systems" by Kevin Davey - A more accessible treatment of similar topics, useful as a practical companion to Tomasini and Jaekle's more theoretical approach. Davey won the World Cup Trading Championship and writes from direct competitive experience.
-
"Advances in Financial Machine Learning" by Marcos Lopez de Prado - Takes the overfitting problem to its logical extreme with advanced techniques for detecting and preventing it in machine learning contexts. The purged walk-forward cross-validation technique extends Tomasini and Jaekle's methodology.
For AMT/Bookmap Integration
-
"Markets in Profile" by James Dalton - The foundational AMT text that provides the market structure understanding to complement systematic validation methods.
-
"Mind Over Markets" by James Dalton - The precursor to Markets in Profile, establishing the Market Profile framework that can be systematized using Tomasini and Jaekle's methodology.
-
"Trading and Exchanges: Market Microstructure for Practitioners" by Larry Harris - Deep treatment of market microstructure, execution, and the ecosystem of market participants. Essential for understanding the execution challenges that systematic daytraders face.
For Statistical Foundations
-
"The Art of Statistics" by David Spiegelhalter - Accessible introduction to statistical thinking for traders without a quantitative background.
-
"Fooled by Randomness" by Nassim Taleb - The philosophical complement to Tomasini and Jaekle's statistical approach. Taleb's treatment of how humans systematically mistake noise for signal is the behavioral underpinning of everything in Trading Systems.
Conclusion
"Trading Systems: A New Approach to System Development and Portfolio Optimisation" is not a book of trading strategies. It is a book about how to determine whether a trading strategy actually works. This distinction makes it one of the most valuable - and most humbling - texts in the trading canon.
For AMT/Bookmap daytraders, the book fills a critical gap. Auction Market Theory provides a powerful framework for understanding market structure. Bookmap provides real-time visualization of order flow dynamics. But neither provides a rigorous method for testing whether the patterns you observe translate into a genuine, persistent, statistically significant edge. Tomasini and Jaekle's methodology - hypothesis-driven design, walk-forward validation, Monte Carlo stress testing, and portfolio diversification - provides exactly that.
The central message is one of intellectual discipline. Most patterns are noise. Most backtests are overfitted. Most systems fail. The tools presented in this book - walk-forward analysis, Monte Carlo simulation, robustness testing, portfolio construction - do not guarantee success. What they guarantee is that you will not fool yourself. And in a domain where self-deception is the default mode of failure, that guarantee is worth more than any trading signal.
The trader who internalizes this book's methodology will develop fewer systems, but the systems they develop will be more likely to survive contact with live markets. They will size positions more conservatively, drawdown less severely, and persist through difficult periods with greater confidence. They will think in terms of portfolios rather than individual bets, in terms of probability distributions rather than point estimates, and in terms of edge decay rather than permanent alpha. This is the mature, professional approach to systematic trading, and it begins with the rigorous methodology that Tomasini and Jaekle have laid out in this essential text.