**TL;DR**

- Even with favourable odds- e.g. a game where you win 50% on heads and lose only 40% on tails – you’re likely to eventually go bust. This is because of the path-dependency of time.
- Ergodicity is when an ensemble-average equals the time-average. For instance, if 10 people doing something once produces the same result as 1 person doing it 10 times. Essentially, ergodicity lets you conveniently convert time into a space equivalent.
- For non-ergodic systems, the aggregate’s performance is a misleading indicator of individual performance. As a whole, the average might look like it’s winning, despite majority of individuals losing.
- The Kelly criterion allows us to optimize our bet size to maximize long-term winnings.
- While most financial systems are non-ergodic, the foundations of classical economic theory rests on fallacious ergodic assumptions – from DCF valuations to the way we measure GDP growth.
- Putting these ergodicity concepts to practical use, I’ve now increased the weighting I give to non-recoverable risks; recognized the substantial role of luck but remain focused on what I can control; and (try to) look for Kelly bets when making important life decisions.

**Being less of a sucker**

The way we interpret data shapes the way we see and understand the world. Mistakes have a compounding effect and the stakes are high.

You’re probably familiar with these common mistakes in interpreting data:

- Looking at
**average**when you should be looking at**median**e.g. for personal income distributions - Looking at
**aggregates**rather than more**granular****segments**e.g. sales increased 10% this year, but if you break it down none of this growth came from the core business segments - Mistaking
**non-linear**outcomes to be**linear**ones e.g. cumulative number of COVID-19 cases

We’ll never get it right 100% of the time, as ‘right’ is context-dependent. But the more we know, the less likely we are to be a sucker.

One such mistake is the **ergodicity assumption**. Surprisingly common, but generally poorly understood and barely known. Much like Black Swans, antifragility, non-linearity, and optionality, it’s another concept that Nassim Taleb has raised his voice on, but this one has received far less attention.

This post:

**Going bust despite favourable odds**: A gambling example, ergodic vs non-ergodic, and the Kelly criterion**Ergodocity economics:**Prevalence and dangers of the ergodic assumption in modern portfolio theory, DCF models, and GDP**Living in bets:**How I’m personally applying ergodicity concepts to life

**1. Going bust despite favourable odds**

**1.1. Definition**

A process is said to be

ergodicif theensemble-average andtime-average isequivalent.

A more formal definition from Ole Peters from his Dec-19 Nature journal article:

The ergodic hypothesis is a key analytical device of equilibrium statistical mechanics. It underlies the assumption that the

timeaverageand theexpectationvalueof an observable are the same. Where it is valid, dynamical descriptions can often be replaced with much simpler probabilistic ones — time is essentially eliminated from the models.

A mouthful right? Let’s leave the definition here for now, and explore a more digestible example.

**1.2. Rigged coin flip game**

Flip a coin. If heads, your net wealth is **increased by 50%**. If tails, your net wealth is **decreased by 40%**.

Would you play? If so, how many times?

At first glance it seems like the game is rigged in your favour. We expect the number of heads and tails to be equal over a long run. And given you win more from each head than you lose on each tail, you should just play as many times as possible right? Besides, even if you end up on a losing streak, you don’t ever lose all your money, you only lose 40% of whatever you were on. You can win it back with the favourable payoffs.

Let’s run a simulation of what could happen if you played 50 times.

- Horizontal axis is time.
- Vertical axis is net worth, a function of time.
- N=1 means we’re looking at the outcome for 1 individual.

There doesn’t seem to be a trend here. We were expecting an upward gradient given the favourable odds. But we end up at a loss.

Let’s extend the simulation for this one individual into 1,000 rounds.

Looks like they’re stuck at zero. This is surprisingly given the favourable odds. We expected the math to offset the unlucky first 50 rounds. Is this guy just really unlucky?

Well here’s the same as above but with a logarithmic axis.

Remember they never go completely bust – since a loss is 40% of their current wealth. Regardless, the downward trend is rather apparent.

Back to our initial graph.

This was a simulation for 1 individual over 50 rounds.

Now, let’s run simulations for 100 individuals playing concurrently over 50 rounds, and show the average results (wealth) at each point in time in one image.

The green line shows the **average** y-values (wealth) for 100 players at each x-value (time). Again, no clear trend.

Simulate for 10,000 players (red line) and we start to see a trend. On average, this group of 10,000 players is sitting on a nice ~10x profit.

Simulate a million concurrent players, and the trend is more apparent.

Again, a logarithmic axis shows this more clearly.

The average wealth of a million people increases exponentially as they play this game. (Recall from all the March COVID-19 graphs that a linear increase on logarithmic axis means exponential growth. It’s a neat trick to distinguish exponential growth from mere very fast growth.)

This trend persists as we extend time to 1,000 rounds and so on.

But hang on, something’s not right.

Earlier, we showed that for an **individual** player, wealth trends to **zero **over time.

But for the **aggregate** of individuals, wealth trends **exponentially upwards** over time.

How can many things that trend to zero add up to exponential growth?

Simple answer: there’s a few really lucky bastards that distort the average.

A more technical answer: **ergodicity**. Or more specifically, a false assumption of ergodicity for this non-ergodic game.

Let’s look at some scenarios for an individual starting with $100.

**Round 1**: 2 scenarios

- H: $100 x 1.5 = $150 (up $50)
- T: $100 x 0.6 = $60 (down $40)
- Outcomes: 1 up, 1 down
**50% up.** - Average after 1 round:
**$105**

**Round 2**: 4 scenarios

- HH: $100 x 1.5 x 1.5 = $225 (up $125)
- HT: $100 x 1.5 x 0.6 = $90 (down $10)
- TH: $100 x 0.6 x 1.5 = $90 (down $10)
- TT: $100 x 0.6 x 0.6 = $36 (down $64)
- Outcomes: 1 up, 3 down.
**25% up.** - Average after 2 rounds:
**$110.3**

**Round 3**: 8 scenarios

- HHH: $100 x 1.5 x 1.5 x 1.5 = $337.5 (up $237.5)
- HTH: $100 x 1.5 x 0.6 x 1.5 = $135 (up $35)
- THH: $100 x 0.6 x 1.5 x 1.5 = $135 (up $35)
- TTH: $100 x 0.6 x 0.6 x 1.5 = $54 (down $66)
- HHT: $100 x 1.5 x 1.5 x 0.6 = $135 (up $35)
- HTT: $100 x 1.5 x 0.6 x 0.6 = $54 (down $66)
- THT: $100 x 0.6 x 1.5 x 0.6 = $54 (down $66)
- TTT: $100 x 0.6 x 0.6 x 0.6 = $21.6 (down $78.4)
- Outcomes: 4 up, 4 down.
**50% up.** - Average after 3 rounds:
**$115.8**

Similarly, we repeat this logic for even more rounds with the help of binomial probability distributions in Excel.

**Round 10**: 1,024 scenarios

- H 10 times: $100 x 1.5
^{10 }= $5,767 - H 5 times, T 5 times (order doesn’t matter here): $100 x 1.5
^{5 }x 0.6^{5 }= $59 - T 10 times: $100 x 0.6
^{10 }= $0.6 - Outcomes: 386 up, 638 down.
**38% up.** - Average after 10 rounds:
**$163 ($100 x 1.05**^{10}).

**Round 100**: 2^{100 }scenarios

- H 100 times: $100 x 1.5
^{100 }= $4.1 x 10^{19} - H 50 times, T 50 times (order doesn’t matter here): $100 x 1.5
^{50 }x 0.6^{5 }= $0.5 - T 100 times: $100 x 0.6
^{100 }= $6.5 x 10^{-21} - Outcomes: 1.7 x 10
^{29 }up, 1.1 x 10^{30 }down.**14% up.** - Average after 100 rounds:
**$13,150 ($100 x 1.05**^{100})

Some observations.

(i) In aggregate, payout is favourable to the players since average wealth is consistently above starting value of $100. This is the ensemble-average.

(ii) But this asymmetric payoff is only favourable on the **ensemble dimension** (taking the aggregate and dividing it by number of players). The pay-off structure is not favourable to an individual on the **temporal **(time) **dimension**. On an individual level, people are actually more likely to lose the longer they play. The gambler’s ruin. Thus, looking at ensemble returns is a poor indicator of individual returns.

So as the game goes on: the house loses money as they’re paying out more than they receive, while more people become losers, and just a lucky few get obscenely rich. (For context, the winnings from 100 consecutive heads, $4.1 x 10^{19} is taking the global GDP of ~$100t and multiplying it by 400k years. So all of the wealth we’re creating today multiplied by longer than Homo sapiens’ existence. The probability of 1 in 2^{100 }is even more ludicrous.)

A variation to the example we explored is a triple-or-nothing coin flip. We get similar results except instead of decaying to zero, people go completely bust and end the game there, while some extremely lucky get fanatically rich.

**1.3. Back to ergodicity**

Recall the definition from earlier.

A process is said to be

ergodicif theensemble-average andtime-average isequivalent.

It’s pretty clear that simulating an individual playing 10,000 rounds gives a drastically different result to simulating 10,000 individuals playing 1 round.

The trajectory of an individual over a long time (time-average) gives a different result to the aggregate of many individual’s trajectories at an instant in time (ensemble-average, aka. space-average). (Note: we can substitute ‘average’ in this statement with expectation value/outcome/probability/pay-off.)

What accounts for this difference is the **path-dependent property of time**. This distinction between the ensemble dimension and temporal dimension is crucial.

In our rigged coin-flip game, most players experience losing streaks. But some extremely lucky ones emerge and push the ensemble average higher, despite most individuals being losers. This game is a typical case a non-ergodic process/outcome/game/system.

A **common** **mistake** is **assuming** **ergodicity.** **Conflating** the ensemble and time average. Assuming non-ergodic systems to be ergodic (similar to mistaking non-linear systems to be linear).

“No probability without ergodicity.” – Skin in the Game (2018), Nassim Taleb

**1.4. More examples**

**Ergodic**

(i) **Brownian motion **is random motion caused by collision of particles with each other, and/or with the walls of a container. For gas molecules in Brownian motion, the average amount of time spent in each region in the container is proportional to the volume of that region. That is, over a long time frame, the time a given molecule spends in one half of the container is equivalent to the time spent in the other half.

(ii) Hypothetical example. Consider **wealth** status / social class in the US population. If individual wealth status was perfectly ergodic, and suppose we lived forever, then we’d spend the same proportion of our lifetime according to the frequency of each of the wealth conditions in the representative sample. So for a given century, about 20 years in blue-collar class, 60 years in the lower-middle class, 10 years in upper-middle, 9 years in upper, and 1 year in the one percent. More on this example later.

(iii) **Colour** perception. Our brains make visual measurements every 20 milliseconds. So if we switch colours any faster than that, the brain will just perceive a time-average which converges with the ensemble-average of the two colours physically mixed. Source: Ergodicity, jail, and time scales; Ergodicity Economics

**Non-ergodic:**

(i) **Mythical man-month**. “If it takes 2 engineering FTE to complete the project in 1 month (i.e. 2 man-months), let’s throw 8 engineers onto it and finish in 1 week.”

(ii) “If it takes 1 pregnant woman 9 months to give birth, then 9 pregnant women should only take 1 month.”

(iii) “Recipe says bake for for 30 minutes at 180 degrees, but I’ll just bake for 60 minutes at 90 degrees.”

**1.5. Another way of looking at it**

You’ll find alternative definitions of ergodicity depending on its disciplinary origin: mathematics, physics, and economics. But they all converge to similar central points.

Mathematics: ergodicity is a property of a (discrete or continuous) dynamical system which expresses a form of

irreducibilityof the system, from a measure-theoretic viewpoint.

That is, ability to substitute a time-derived thing for an ensemble-derived thing and vice-versa.

Econometrics: a process is ergodic if its statistical properties can be

deducedfrom asingle,sufficientlylong, random sampleof the process.

In some cases, you may have access to a large sample but not enough time. In others, you may only be able to get a small sample but you have data on that for a long time period. So long as what you’re measuring is ergodic, you can convenient extrapolate either

Physics: the ergodic hypothesis

^{}says that, over long periods of time, thetimespentby a system in some region of the phase space of microstates with the same energy isproportionalto thevolumeof this region, i.e., that all accessible microstates are equiprobable over a long period of time.

Again, this is linking time and space (ensemble) – if ergodic, you can transform data you have on one into the other.

General dictionary. Ergodic: of or relating to a process in which every sequence or sizable sample is equally

representativeof the whole (as in regard to a statistical parameter)

I think you get it now.

**1.6. Absorption: opposite of ergodicity**

The exact opposite of perfect ergodicity is an **absorbing state**.

Back to the Brownian motion example. What we saw earlier were perfectly elastic walls (no energy lost on bounce). But imagine if the walls were perfectly plastic (absorbs all energy upon contact). That is, sticky. The molecules would stick upon first contact, and its time in a given region of the box would no longer be proportional to the volume of that region. Such a system would be **strongly non-ergodic**. High irreversibility. (There’s ergodic, weak non-ergodic, and strong non-ergodic. Won’t elaborate more on this here but check out this video if interested in going more technical.)

Back to the hypothetical wealth example. In reality, when someone enters the 1%, they’re quite likely to remain in the 1%. And if someone was born in the bottom 20%, it’s more likely they’ll stay there. Social mobility has come a long way but there’s still a long way to go.

“Dynamic equality is what restores ergodicity, making time and ensemble probabilties substitutable.” – Skin in the Game (2018), Nassim Taleb

**1.7. Kelly criterion: adjust bet size dynamically**

Take the rigged coin flip game we saw earlier, but instead of being forced to bet your entire net wealth, you could adjust your bet size on each round. We saw that going all-in every time is a dumb idea – you’ll eventually decay to zero. But given the favourable odds on an ensemble basis, is it then best to just walk away?

There actually exists a mathematically proven **optimal** **bet** **size** depending on the how favourable the odds (or pay-outs) are. This is the Kelly criterion. (Note that pay-off and odds are interchangeable here. “*$1.50 pay-off on heads and $0.60 on tails on regular coin*” is the same as “*$1.00 pay-off on heads and $1.00 pay-off on a heads-biased coin*“. )

**Kelly criterion: **the optimal bet size (as a portion) to maximize long-term pay-off.

Bet too small and you miss exploiting a favourable bet. Bet too big and you’ll eventually go bust. A Kelly bet is just the right size. There’s varying versions of the formula, and I find this to be the simplest:

- f* = fraction of bankroll to bet
- p = probability of winning
- q = probability of losing (= 1 – p)
- a = fraction you could lose (i.e. if lose, go from 1 to 1 – a)
- b = fraction you could win (i.e. if win, go from 1 to 1 + b)

In our rigged coin flip game, optimal bet size is thus = 0.5/0.4 – 0.5/0.5 = 0.75. We should put 75% of our money in regardless of how much money we have. (From the formula we can see that it’s possible for f* to be negative, in which case the optimal strategy is to take the bet on the other side.)

We can also visualize this.

- Vertical axis is pay-off
- Horizontal axis is bet size
- Coloured curves show expected long-term pay-offs for a given bet-size for a given odd. Each coloured curve represents different odds. From top, blue line 0 is 0% chance of losing. From bottom, red curve is 50% chance of losing.
- Red dot is the Kelly bet that maximizes pay-off for a given odd in the long-run.

Clearly, if your losing odd is 0%, the optimal bet size is 100%. Go all-in if you can’t lose. The more favourable the bias, the more aggressive you should be. Conversely, if your losing odds is more than 50%, the optimal bet size is 0% of your stack. Why bet anything in an unfavourable game?

Note that even if your odds of losing is just 5%, if you gamble too consistently, you’ll get caught in an unrecoverable losing spiral. This is shown to the right of the red dot on the chart.

There are two methods to consider in a risky strategy…

The first is to know all parameters about the future and engage in optimized portfolio construction, a lunacy unless one has a god-like knowledge of the future. …one needs to know the entire joint probability distribution of all assets for the entire future, plus the exact utility function for wealth at all future times. And without errors!

[Second], Kelly’s method… requires no joint distribution or utility function. It is very robust. …estimate the ratio of expected profit to worst- case return– dynamically adjusted to avoid ruin. …worst case is guaranteed (leave 80% or so of your money in reserves)… So, assuming one has the edge (as a sole central piece of information), engage in a dynamic strategy of variable betting, getting more conservative after losses (“cut your losses”) and more aggressive “with the house’s money”.

The first strategy was only embraced by academic financial economists –empty suits without skin in the game — because you can make an academic career writing BS papers with method 1 much better than with method 2. On the other hand EVERY SURVIVING speculator uses explicitly or implicitly method 2 (evidence: Ray Dalio, Paul Tudor Jones, Renaissance, even Goldman Sachs!) For the first method, think of LTCM and the banking failure.– Nassim Taleb’s Amazon book review of Kelly Capital Growth Investment Criterion

**2. Ergodicity economics**

While most financial systems are non-ergodic, the foundations of classical economic theory rests on fallacious ergodic assumptions. Equilibrium is not the natural state of the economy. Many models remove time from their equations by reducing it into an ensemble equivalent. This results in a false sense of security and optimism in most aggregate measurements, and risk management approaches. The importance of a quantity’s path through time is underestimated.

**2.1. Modern Portfolio Theory**

Modern portfolio theory stipulates that there exists some optimal risk-to-return profile.

But the theory rests on the ergodic hypothesis by conflating an ensemble aggregate return with the individual’s path-dependent return through time.

Modern Portfolio Theory utilizes ensemble or average returns to calculate a portfolio’s expected return. However, an individual portfolio manager is

not interested in the ensemble returnbut in theindividual portfolio’s return through time(i.e., the path-dependent return). The catch is that ensemble and time average returns are NOT equal because the distribution of returns is not normal. This is referred to as non-ergodic. – Complexity Investing whitepaper, NZS Capital

**2.2. DCF models**

The DCF (discounted cash flow) model is the textbook method for valuing a company. Basically it involves summing all future cash flows while valuing future dollars that are further out from now as less than dollars today. This model assumes ergodicity because it reduces the time dimension by converting it into an ensemble expectation value – the net present value. While it’s useful in some instances, as with all models, we need to be well informed of its assumptions and limitations.

First, the discount factor applied is constant over time – an assumption that doesn’t hold up particularly well. Second, this constant discount rate typically comes from the weighted average cost of capital (WACC), which in turn is based on the contentious notion of equity risk premium and beta – both of which are easy to manipulate.

Quick detour on modern portfolio theory for those unfamiliar

- y = α + β+ ε
- y, gamma = return of security
- α, alpha = idiosyncratic factors specific to this security
- β, beta = co-movement with relevant market indices
- ε, epsilon = everything else
- Equity risk premium is the additional risk on top of beta, that is, on top of the general market risk

All of these concepts in the DCF model derives from the Capital Asset Pricing Model, which derives from Expected Utility Theory, which derives from a 1738 Bernoulli paper, which had a critical math error that was only pointed out recently (by Ole Peters). Here’s the TED talk on it.

**2.3. From GDP to DDP**

GDP (gross domestic product): a measure popularized during the production era amidst and post-WWII to dimension get a quantitative grasp on economic activity, and a proxy for measuring national prosperity. Much of the criticisms center on what it excludes: environmental sustainability, its emphasis of having a trade surplus, social/cultural/emotional factors such as happiness, and other well-justified formula-imperfection-picking. Meanwhile, less often discussed is recognizing that GDP is an ensemble value. It assumes ergodicity.

Remember, for something to be ergodic, what happens to the aggregate (ensemble-average) needs to be the same as what happens to the individual over time (time-average). When we say GDP increased by 2.3% and people applaud, that’s a reference to an aggregate increase of 2.3%. As we saw in the non-ergodic distribution in our rigged coin-flip game, this average outcome is far from a representative one for most individuals. Growth in GDP is a poor indicator for wealth growth at an individual level. Because a few really rich people pull up the average.

Ole Peters writes about a democratic domestic product (DDP). In a hypothetical country, 100 have an income of $50k/year. Suppose in the next year, one person enslaves everyone and steals all of their income ($5m = $50k x 100). The GDP would show as the same even though majority of people are now worse off. If this one person somehow grows his income from $5m to $5.5m, the country would show a 10% growth. Essentially, under this plutocratic system, GDP is weighted by the dollar, not by the individual.

Alternatively, what if we computed the aggregate growth not as a dollar-weighted average growth (ensemble), but as a time-average for each of the individuals. That is, we take the income growth rate for each individual, and then average those averages. DDP gives equal weight to the individual rather than to the dollar. In this scenario, this DDP growth would show as a big fat negative number as majority of people saw reductions.

**3. Living in bets**

So how am I practically applying ergodicity in everyday life?

In 3 ways.

**3.1. Avoid doing really stupid shit**

Our lives are non-ergodic. It’s path-dependent on time. Put simply, **we only live once**.

There’s 2 types of risks. One is recoverable, the other is not.

For an individual, unrecoverable risks include death, permanent disability, severe mental illness, unrecoverable personal brand damage (committing severe crime) etc. Once we have brain-machine interfaces, we can add data deletion to this list too. I think I take pretty standard measures to minimize these probabilities: caution on road, don’t get into fights etc. Then there’s less obvious but much more prevalent things that kill you slowly: excessive stress, poor quality sleep, bad nutrition etc. (Note: Cardiovascular disease is leading cause of death. Sleeping less than 6 hours a day increases risk of cardiovascular disease by 20%. Pareto-optimize health outcomes: focus on the 20% of actions that translate to 80% of the results. For both physical and mental health.)

Personally, I know I’m more exposed to unrecoverable risks from my travels. My obsession with under-rated destinations has taken me to places like: Mongolia, Siberian part of Russia, Myanmar, Uzbekistan, Kyrgyzstan, Ethiopia, and Uganda. With trips like these you’re bound to run into some kind of trouble. Seeing crocodiles while white-water rafting, fending off a thief breaking into room while asleep, driving mountain roads on a motorbike in winter while my fingers were frozen, getting stranded in a desert for several hours amidst a violent tribal conflict, riding a horse at full speed without proper training or equipment, and being in taxi speeding down a country highway at night no working headlights.

My intention is not to glorify dangerous travel. When I look back at moments like this, there’s 3 types:

- (i) “That was fun and I’d do it again. Although with more caution.”
- (ii) “That was fun, but some things are only fun the first time. I wouldn’t do it again.”
- (iii) “That was so fucking stupid. You could have died. Narrowly missed out on a Darwin Award. I’ll never put myself in that position again.”

So now, I’m more cautious and do more research for type (i) and (ii). And do more to mitigate type (iii) from occurring – and pay more to for safer transport arrangements, secure accommodation, reliable tour operators etc.

I still want to visit some more wild/chaotic places, but I won’t take unnecessarily high risk of death just to ease my travel bug. Understanding ergodicity has shifted my balancing of pragmatism with adventure. More emphasis on **min-max** (minimize the maximum downside).

My friend coloured this beautifully:

- Me: “Bali? Oh, hire a scooter and just drive around the island yourself. More fun. More freedom.”
- Him: “But I can’t drive a scooter.”
- Me: “It’s easy. You can learn. Only took me 5 minutes and I was off.”
- Him: “Nah man. It’s so cheap to just pay someone to drive me around.
**Life’s too good to take that unnecessary risk.**“ - Wow. Well if you put it that way. Agree. Life’s too good.

“One may be risk loving yet completely averse to ruin… In a strategy that entails ruin, benefits never offset risks of ruin… Every single risk you take adds up to reduce your life expectancy… Rationality is avoidance of systemic ruin.” – Skin in the Game (2018), Nassim Taleb

“…three distinct sides of risk:

The odds you will get hit.

The average consequences of getting hit.

The tail-end consequences of getting hit.

…The first two are easy to grasp. It’s the third that’s hardest to learn, and can often only be learned through experience… Tail-end events are all that matter.” – The Three Sides of Risk, Collaborative Fund

**3.2. Responding to luck: focus on what’s in control**

To borrow from the modern portfolio theory equation we saw earlier, I conceptualize life ‘success’ as something like this:

- y = f (α, β, ε) = a function of α, β, ε.
- y, gamma = ‘success’ in life. Can be defined in many ways.
- α, alpha = a pinch of genetic endowment, but mostly grit, growth-mindset, skills, experiences etc
- β, beta = your environment, your network
- ε, epsilon = luck, and other unexpected factors e.g. if world suddenly changes such that the niche skills you had suddenly becomes in hot-demand

As I learn more about how the world works, it seems that many conventionally success people continue to under-play the role of luck. Yes they all worked hard, but they all had luck on their side – in some shape or form. We saw the magnitude of of absurd unequal distributions that a simple coin flip game can yield. Even in a world with favourable ensemble odds, you need a healthy amount of luck on your side to come out ahead (without applying Kelly criterion that is). I’m not talking about net wealth, but rather a higher level dimension. If you want to be rich, why? Then why again? Then why again? Eventually you’ll hit an end-point. I ended up with something like **fulfillment**. Others might call it happiness. Doesn’t matter. Let’s not get pedantic with the semantics. That’s the highest level dimension that you’re optimizing for. That’s the thing that I’m talking about now.

Now I’m not saying luck is the only path to ‘success’ in a defeatist manner. I’m saying that luck seems to have a bigger than what I previously thought. I still value alpha (grit, growth mindset etc), but I’ve also incorporated the effect of Black Swans, luck, and ergodicity into my life model (epsilon).

Based on this belief – the disproportionately high-value of luck – what do I do?

My answer: just focus on what’s in control. The alpha, and partially, beta.

If luck comes, great. If it doesn’t, all good. Either way, I’m fulfilled. I really did my best – what more could I ask for.

If it turns out that luck really can be controlled to some extent – work the subconscious, energy vibrations, law of attraction, hard work attracts luck etc – then great, I’ve got pretty good exposure to a lot of that stuff too. If not, all good. I’ve still derived a happiness utiility from the positive thoughts. It’s a set-up for large uncapped exposure with a capped downsize risk – a strategy I discussed in detail when articulating antifragility and optionality.

Although, sometimes I wonder if this optimistic outlook is a cowardly concession to accommodate an unacknowledged fear of failure. Fear of mediocrity. Fear of living a boring life. A way of reassuring myself that no matter what outcome, I’ve done my best. What if I get lazy and blame that on being unlucky. What if this I-can’t-lose mindset actually hinders me from taking more bold actions to live the life that I want? I’m pretty sure it’s not. But I’m not 100% sure. So I continue to explore this possibility from time to time. Though, I’m also aware of the cost to thinking about this – there’s a good chance it may hinder rather than fix.

So for now, I’m confident living by the principle, **focus on what’s in control**. Mental energy spent on the uncontrollable factors is a sunk opportunity cost.

**3.3. Make Kelly bets in major life decisions**

Life is an accumulation of bets we’ve made (as well as bets made by our parents, and the society we grew up in etc). The future is also just a series of bets. A non-ergodic path-dependent sequence.

Some bets have higher stakes than others: changing jobs, re-locating, picking a lifelong partner etc. And other bets seem like lower stakes, but due to their frequency, accumulate into high stakes – e.g. lifestyle choices. Each bet is associated with a pay-off/return, potential loss/downside, and a probability. The currency of the downside, the thing we’re gambling, is typically our **time**.

So it’s worth putting in more conscious effort into assessing the favourability of a bet. Especially for higher stakes. Let the beauty of asymmetry compound. Set yourself up so that time is more of an ally than an enemy. Sure, you could still go bust by getting hit by a bus tomorrow. But what can you do. Getting as close as possible to favourable bets and making lots of them is an optimal life strategy.

Extremely conceptual (even by my standards), but I ask myself: “Is there some Kelly criterion that’s applicable here to adjust the bet size to maximize the pay-off?”

In practice, we don’t know the exact odds. Even if we did, it’ll change over time. It’s a Bayesian probability. We don’t know the pay-out. And even if we did, we wouldn’t have the computing power, or time to conduct thorough analysis. We’d be at the mercy of analysis paralysis.

But we can still try. Keep calibrating the guide in your head that tells you what’s worthwhile, and what’s important. That’s what really counts.

P.S. Did you notice the title of this post also makes an ergodic assumption?