Skip to content

When should we raise the stakes?

The Kelly Criterion

A version of this article appears in TWISTED LOGIC: Puzzles, Paradoxes, and Big Questions. By Leighton Vaughan Williams. Chapman & Hall/CRC Press. 2024.

TAKING ADVANTAGE OF THE ODDS

One of the most critical aspects of any betting strategy is determining the size of the bet when we believe the odds are in our favour. The answer to this pressing question was formalised by John L. Kelly, Jr., an engineer at Bell Labs by profession, as well as a hobbyist daredevil pilot and recreational gunslinger. His methodology, known as the Kelly Criterion, is a mathematical formula designed to establish the optimal bet size when we have an advantage, that is, when the odds favour us.

However, having the advantage doesn’t guarantee a successful outcome. Irrespective of our edge, excessive betting can lead to large losses and, in worst-case scenarios, to bankruptcy. This is where the Kelly Criterion comes into play. It takes account of both the size of our edge and the potentially damaging impact of volatility.

UNDERSTANDING THE KELLY CRITERION

The Kelly Criterion is essentially a mathematical formula that calculates the optimal amount to bet or invest when the odds are in our favour. The fundamental principle underpinning the Kelly Criterion is that the amount of capital wagered should be related to our advantage at the available odds. It emphasises the relationship between the size of our bet and our perceived edge.

Consider this simple illustration: Suppose you have a coin where the probability of getting a head (winning) is expected to be equal to the probability of a tail (losing). Now, suppose you have secret information that the next coin toss will certainly be heads. In this case, you have a 100% edge. According to the strict Kelly Criterion, you should bet your entire capital because you’re guaranteed to win. In real-life applications, even with very high confidence, betting one’s entire capital is risky, however, due to the possibility of unforeseen factors. This is more a thought experiment than a practical recommendation.

In any case, in most scenarios the outcomes are not binary, and the probability of winning is rarely 100%, even in theory.

Let’s consider a different situation: You’re still betting on the coin toss, but this time your secret information gives you a 60% chance of landing heads and a 40% chance of tails. Your edge is now 20%, and a very basic Kelly strategy is to stake 20% of your capital.

This example reflects the core concept of the Kelly Criterion. It’s not only about gauging when you have the advantage—it’s also important to understand precisely how much to stake when you do. In theory, this sounds simple, but in practice, accurately identifying that advantage can be complex.

On a broader scale, the Kelly Criterion can be employed in various fields beyond betting, such as investing and trading, to determine the optimal size of a series of bets or investments. Its aim is to maximise the exponential growth of the bettor’s or investor’s wealth over the long term.

A strength of the Kelly Criterion is its flexibility. It allows you to adjust the proportion of the capital that you bet based on how strong your edge is.

It’s important to note, however, that the Kelly Criterion assumes that the bettor can reinvest their winnings. This is crucial for the ‘compounding’ aspect of the strategy, which is what allows the wealth to grow faster than it would with other systems. This compound growth strategy is what differentiates Kelly betting from more static strategies, but also introduces higher volatility in the short term.

Ultimately, the Kelly Criterion offers a robust methodology for managing risk and maximising returns when the odds are in our favour. However, as with any strategy, understanding the core principles is just the beginning—it’s the accurate identification of the edge and the consistent application of the strategy that’s critical for long-term success. Misestimations can lead to over-betting and significant losses.

APPLYING THE KELLY CRITERION

The application of the Kelly Criterion can have profound implications for various fields beyond gambling, such as investing and trading. The crucial component is to understand that the Kelly Criterion isn’t just about betting when we have an edge; it’s about calculating the precise amount to bet to maximise compounded return over time.

This is where the Kelly formula can come into play:

F = Pw − (Pl/W)

where

F is the Kelly criterion fraction of capital to bet,

W is the amount won per amount wagered (i.e. win size, net of the stake, divided by loss size),

Pw is the probability of winning, and

Pl is the probability of losing.

When we apply this formula, we calculate the optimal fraction of our capital to bet, given our probability of winning (Pw), our probability of losing (Pl), and our potential return (W).

Consider a simple example: Suppose we have an even-money bet, i.e. the amount you stand to win, net of the stake, is the same as the amount you risk. In this scenario, the value of W is 1. If our chance of winning is 60% and our chance of losing is 40%, substituting these values into the simplified Kelly formula (F = Pw − Pl) gives us F = 0.60 − 0.40 = 0.20 or 20%. This means that in order to maximise our long-term return, we should bet 20% of our capital.

Let’s consider a slightly more complicated scenario: Suppose we have a bet where we stand to win double the amount we risk, i.e. W = 2, and the probability of winning and losing is both 50%. Substituting these values into the original Kelly formula gives us F = 0.50 − (0.50/2) = 0.50 − 0.25 = 0.25 or 25%. This means we should bet 25% of our working capital to maximise our long-term return.

The Kelly Criterion is designed to ensure that you never go bankrupt because the recommended bet size decreases as your capital decreases. However, this doesn’t mean you can’t lose money. The Kelly Criterion maximises long-term growth rather than short-term returns. This means that there will be times when you lose money, but over the long run, you should come out ahead.

It’s crucial to remember that the Kelly Criterion assumes you know the true probabilities of the outcomes, which is often not the case. In practice, we’re often working with estimated probabilities, which means there’s a risk that we could, for example, overestimate our edge and bet too much. Therefore, many investors and bettors use a fraction of the Kelly Criterion (betting a fixed fraction of the amount recommended by Kelly) to reduce their risk.

Lastly, while the Kelly Criterion offers a mathematical approach to betting and investing, it doesn’t account for the emotional aspect of risking money. Remember that the goal is not just to maximise returns, but also to sleep well at night.

POTENTIAL RISKS AND LIMITATIONS

While the Kelly Criterion can be an effective strategy for maximising the growth of capital in the long run, it is not without its potential risks and limitations. These should be understood and considered before applying the formula.

Estimation Errors

The effectiveness of the Kelly Criterion hinges on the accuracy of the probabilities used in the calculation. An overestimation of the probability of winning (Pw) can lead to excessive bet sizes and the risks associated with over-betting.

Minimum Bet Size

The Kelly Criterion presupposes that there is no minimum bet size, which is rarely the case in real-world scenarios, especially in investing and trading. In situations where a minimum bet size exists, the possibility of losing all of the capital becomes a reality if the amount falls below this threshold.

Risk Tolerance

The Kelly Criterion determines bet sizes purely based on mathematical calculations to maximise long-term growth. It does not take into account the individual bettor’s or investor’s risk tolerance. An aggressive bet size recommended by the Kelly Criterion may not be psychologically comfortable for some, causing stress and potentially leading to sub-optimal decisions.

Given these potential risks and limitations, it is common for many investors and bettors to use a fractional Kelly strategy, betting a fraction (like half or a third) of the amount recommended by the Kelly formula. This approach can help mitigate the risks associated with over-betting and inaccuracies in probability estimation while still providing the benefits of proportional betting and capital growth. However, even a fractional Kelly strategy should be tailored to individual circumstances, including risk tolerance and the ability to withstand potential losses.

CONCLUSION: TAKING ADVANTAGE OF OUR EDGE

The Kelly Criterion, devised by John L. Kelly, Jr., is a unique betting strategy that uses probability and potential payout to determine the optimal bet size when the odds are in our favour. The mathematical formula suggests betting a fraction of capital equivalent to the size of one’s advantage. However, it’s crucial to account for potential errors and uncertainties that can affect the real-world implementation of this strategy.

Uncertainty in the size of any actual edge at the odds and the potential for a bumpy ride due to volatility mean that we should always exercise caution. As a result, unless we’re prepared for potential high volatility and have unwavering confidence in our judgment, adopting a fractional Kelly strategy might be the most prudent approach. This strategy allows us to stake a defined fraction of the recommended Kelly amount, reducing risk while still taking advantage of our edge.

Does seeing a white tennis shoe make it more likely that all flamingos are pink?

Introducing the Accessibility Principle

UNDERSTANDING HEMPEL’S PARADOX

In the mid-20th century, philosopher Carl Gustav Hempel introduced a paradox that came to be known as ‘Hempel’s Paradox’ or the ‘Raven Paradox’. The paradox begins with a seemingly simple and clear premise: If the hypothesis is that ‘all ravens are black’, then any observation of a black raven should help to support the hypothesis.

However, Hempel pointed out that this statement is logically equivalent to the statement: ‘Everything that is not black is not a raven’. Hence, any observation of a non-black, non-raven object, such as a white tennis shoe, should also help to support the hypothesis.

Yet it feels strange to believe that seeing a white tennis shoe should serve to increase our belief that all ravens are black.

HEMPEL’S PARADOX AND THE COLOUR OF FLAMINGOS

Now, let’s apply this principle to another statement: ‘All flamingos are pink’. This proposition is logically equivalent to: ‘Everything that is not pink is not a flamingo’. By Hempel’s argument, observing an object that is not pink and not a flamingo, such as a white tennis shoe, would provide evidence in support of the hypothesis that all flamingos are pink.

From a formal logic perspective, this argument makes sense. However, our intuition may still find this hard to accept, mirroring the original conflict inherent in Hempel’s Paradox.

TESTING THE HYPOTHESIS

In conventional hypothesis testing, we would go out and find some flamingos, verifying if they are indeed pink. But the Raven Paradox suggests that we could conduct meaningful research by simply looking at random non-pink things and checking if they are flamingos. As we collect data, we increasingly lend support to the hypothesis that all non-pink things are non-flamingos, equivalently that all flamingos are pink.

While this approach holds up logically, it does have its limitations. Considering the vast number of non-pink things in the world compared to the population of flamingos, the hypothesis can be much more confidently validated by sampling flamingos directly. Hence, although Hempel’s Paradox does not contain a logical flaw, it is not an efficient or practical method for testing the hypothesis.

THE ACCESSIBILITY PRINCIPLE (OR OBSERVATIONAL LIKELIHOOD PRINCIPLE)

Suppose we have two hypothetical species—one is a type of bird that frequents populated areas (Species A), and the other is a rare kind of lizard that lives in remote, inaccessible areas (Species B). If both these species are unobserved, it’s more likely that Species B exists rather than Species A, because Species B is less likely to be observed due to its habitat even if it exists. In contrast, Species A should have been observed if it were real due to its frequent presence in populated areas. I term this the ‘Accessibility Principle’, or alternatively the ‘Observational Likelihood Principle’. These terms suggest that the likelihood of an entity’s existence depends on its observability. This aligns with real-world scientific practices, where the absence of evidence is not always evidence of absence, particularly when dealing with hard-to-observe phenomena.

So, let’s take the propositions in the thought experiment in turn. Proposition 1: All flamingos are pink. Proposition 2 (logically equivalent to Proposition 1): Everything that is not pink is not a flamingo. Proposition 3 (the ‘Accessibility Principle’): If something might or might not exist but is difficult to observe, it is more likely to exist than something which can be easily observed but is not observed.

Following from these propositions, when I see two white tennis shoes, I am ever so slightly more confident that all flamingos are pink than before. This is especially so if any non-pink flamingos that might be out there would be easy to spot. And I’d still be wrong, but for all the right reasons.

CONCLUSION: THE OBSERVATION PARADOX

In summary, Hempel’s Paradox is an intriguing clash between intuitive reasoning and formal logic. It forces us to confront the subtleties of hypothesis testing and belief formation. In this example, the paradox implies that we may gain a tiny bit more confidence in the hypothesis that all flamingos are pink if we observe a white tennis shoe. However, such indirect evidence should be considered in its appropriate context, not as a substitute for direct evidence. The key point of the paradox is instead to challenge our understanding of the meaning of evidence and to provide valuable insights into the nature of logical reasoning. Essentially, any hypothesis is always susceptible to new evidence that can strengthen support for it. In the case of the pink flamingo hypothesis, this applies whether it comes from observing a flock of pink flamingos or (to a much lesser degree) a pair of white tennis shoes. Until you see an orange flamingo, then you know otherwise!

When should we accept the simplest explanation?

Exploring Occam’s Razor

THE PRINCIPLE OF SIMPLICITY

In this section we explore Occam’s Razor. William of Occam (also spelled William of Ockham) was a prominent 14th-century philosopher and theologian known for his emphasis on simplicity in philosophical and theological matters. His philosophical contributions, particularly the principle of simplicity, have had a lasting impact on various fields of knowledge. Occam’s Razor, derived from his philosophy, has become synonymous with the method of eliminating unnecessary hypotheses and choosing the simplest explanation consistent with the evidence.

OCCAM’S RAZOR: PRINCIPLE AND EXPLANATION

At the heart of Occam’s philosophy, therefore, is the principle of simplicity, which later became known as Occam’s Razor. The razor can be summarised as follows: ‘Entities should not be multiplied without necessity’. In other words, when faced with competing explanations or hypotheses, the simplest one that adequately explains the available evidence should be preferred.

Occam’s Razor guides our thinking by encouraging us to avoid unnecessary assumptions and complexities. It suggests that we should prefer explanations that require fewer additional elements or entities. By choosing simplicity over complexity, Occam’s Razor helps us navigate knowledge acquisition and hypothesis formation.

To be clear, it’s important to note that simplicity does not mean ‘easier to understand’ but rather ‘involving fewer assumptions or conjectures’. Complexity should only be considered when simplicity fails to adequately explain the phenomenon.

OCCAM’S RAZOR: A CRUCIAL HEURISTIC

Occam’s Razor serves as a crucial heuristic in problem-solving and theory formulation. It proposes that among competing hypotheses, the one with the fewest assumptions should be selected, provided it adequately explains the phenomenon in question.

Occam’s Razor does not just simplify our thinking processes; it actively steers us away from the allure of unnecessary complexities and conjectures. By advocating for simplicity, it aids in refining our approach to knowledge acquisition and hypothesis development, ensuring that complexity is introduced only when absolutely necessary to explain the data adequately.

THE ROLE OF OCCAM’S RAZOR IN SCIENCE: TOWARDS ELEGANT EXPLANATIONS

The implications of Occam’s Razor extend significantly into scientific inquiry. It underpins the scientific method, where explanations for observed phenomena are sought and hypotheses are developed. By favouring parsimonious explanations, the principle encourages scientists to construct theories that are not only elegant but also more comprehensible and testable. This preference for simplicity has facilitated remarkable advancements in our understanding of the world, emphasising that the most profound explanations often emerge from the most straightforward assumptions.

OCCAM’S RAZOR AND OVERFITTING: COMPLEXITY AND GENERALISATION

Occam’s Razor finds empirical support in the phenomenon of overfitting, particularly in the field of statistics and machine learning. Overfitting occurs when a model becomes overly complex and fits the noise or random variations in the data instead of capturing the true underlying patterns.

By adhering to Occam’s Razor, researchers can avoid the pitfall of overfitting, ensuring that their models capture the essential features of the data while remaining parsimonious.

OCCAM’S LEPRECHAUN: AVOIDING AD HOC HYPOTHESES

In the pursuit of explanations, it is common to encounter situations where additional assumptions are introduced to save a theory from being falsified. These ad hoc hypotheses act as patches to compensate for anomalies that were not anticipated by the original theory. Occam’s Razor plays an essential role in evaluating such situations.

Imagine a situation, for example, where someone claims that a mischievous leprechaun is responsible for breaking a vase. There is likely to be serious scepticism about this claim. However, the person who made the claim introduces a series of ad hoc explanations to counter potential falsification.

For instance, when a visitor to the scene sees no leprechaun, the claimant asserts that the leprechaun is invisible. To test this, the visitor suggests spreading flour on the ground to detect footprints. In response, the claimant states that the leprechaun can float, thus leaving no footprints. The visitor then proposes asking the leprechaun to speak, but the claimant asserts that the leprechaun has no voice. In this way, the claimant keeps introducing additional explanations to prevent the hypothesis of the leprechaun’s existence from being falsified.

The example of Occam’s Leprechaun illustrates how additional assumptions can be added in an ad hoc manner to preserve a theory from being disproven. These ‘saving hypotheses’ create a flow of additional explanations that make the theory less able to be falsified. Occam’s Razor encourages us to be sceptical of such ad hoc hypotheses and instead favours simpler explanations that adequately account for the evidence.

OCCAM’S RAZOR AND PREDICTIVE POWER: PARSIMONY AND EFFICIENCY

Another aspect of Occam’s Razor is its association with the predictive power of theories. A theory that can accurately predict future events or observations based on fewer assumptions is considered more efficient.

By favouring simplicity, Occam’s Razor guides scientists to develop theories that provide explanatory power and predictive efficiency. The simplest theory that adequately explains the available data and accurately predicts future outcomes is often preferred.

This emphasis on predictive power aligns with the pragmatic approach that Occam advocated, where theories are judged not only by their ability to explain past observations but also by their ability to make successful predictions.

CONCLUSION: THE PURSUIT OF KNOWLEDGE

In sum, Occam’s Razor, a principle deeply rooted in the philosophy of William of Occam, remains a fundamental tool in the pursuit of knowledge. It encourages simplicity, efficiency, and parsimony in our explanations and theories. By guiding us towards the simplest explanations that remain consistent with the available evidence, Occam’s Razor plays a crucial role in scientific methodology, theory development, and everyday reasoning. Its continued relevance underscores the timeless appeal of simplicity in our quest to understand and explain the world around us.

Do We Live in a Simulation?

And Does it Matter?

A version of this article appears in TWISTED LOGIC: Puzzles, Paradoxes, and Big Questions. By Leighton Vaughan Williams. Chapman & Hall/CRC Press. 2024.

THE SIMULATION HYPOTHESIS

Is our reality a simulation crafted by a more advanced civilisation? This provocative question, central to the Simulation Hypothesis popularised by philosopher Nick Bostrom, challenges our understanding and perceptions of existence and reality. Bostrom’s concept of “ancestor simulations” proposes that a sufficiently advanced civilisation could simulate consciousness, allowing simulated beings to experience life as we know it. If such civilisations exist and choose to run these simulations, the likelihood that we’re in one rather than ‘base reality’ increases significantly.

The creators could be located at any stage in the universe’s timeline, even billions of years into the future.

It’s a sort of digital time travel, allowing them to witness and potentially even interact with their own past. Bostrom argues that if any civilisation reaches a high enough technological level to be able to run these sorts of simulations and is interested in doing so, we are more likely to be in one of these simulations rather than in ‘base reality’.

THE FOUNDATION OF BOSTROM’S SIMULATION ARGUMENT

Bostrom argues that at least one of three possibilities must be true:

  1. Civilisations at our level of development almost invariably fail to reach a technologically super-advanced stage. This failure is marked by their extinction or incapacity to develop the technological means necessary to create highly detailed simulations of reality including simulated minds.
  2. Among civilisations that do reach a super-advanced stage, possessing the ability to create highly detailed simulations of their ancestors or historical periods, there is an overwhelming lack of interest in actually conducting such simulations.
  3. We are almost certainly existing within a simulation ourselves. This follows from the assumption that if super-advanced civilisations have both the interest and capability to run numerous simulations, the number of simulated consciousnesses would vastly outnumber the number of “real” consciousnesses.

Bostrom’s argument invites us to consider the implications of our technological trajectory and the nature of consciousness. It proposes a framework where the advancement towards and the potential capabilities of super-advanced civilisations lead to a significant probability that our perceived reality might not be the base reality. This philosophical inquiry not only challenges our understanding of existence but also highlights the profound implications of future technological capabilities on our perception of reality and consciousness.

PROBING THE DEPTHS OF THE SIMULATION ARGUMENT

To fully grapple with these propositions, we must examine each statement individually. For the first proposition to be false, a civilisation would need to exhibit the capability to survive potentially catastrophic phases, whether they are caused intentionally, accidentally, or through ignorance, without succumbing to complete annihilation.

The second proposition is highly dependent on factors we can hardly predict, such as the ethical frameworks of advanced civilisations, their curiosity, and their respect for the integrity of intelligent consciousness. Even so, it might seem implausible that almost no civilisations with the capacity to create such simulations would choose to do so.

Unless civilisations either fail to reach the stage at which they can create such simulations or choose not to do so, then we must face a startling conclusion: we are very probably living in a simulation.

NAVIGATING THE PROBABILITY LANDSCAPE

Summarising the argument, a ‘technologically mature’ civilisation would have the capability to create simulated minds. Hence, one of the following must hold:

The fraction of civilisations reaching ‘technological maturity’ is close to zero or zero.

The fraction of these advanced civilisations willing to run these simulations is close to zero or zero.

We are almost sure to be living in a simulation.

If the first proposition holds true, our civilisation will almost certainly not reach ‘technological maturity’, which introduces a sense of urgency and uncertainty regarding our collective future. If the second proposition is true, then almost no advanced civilisations are interested in creating simulations, raising questions about the nature and motivations of advanced civilisations. If the third proposition is true, then we should challenge our entire perception of reality.

In the face of such profound uncertainty, we might find it pragmatic to assign equal weight to each of these propositions.

THE SINGULAR CIVILISATION HYPOTHESIS

But what if ours is the only civilisation that will ever reach our stage of development? This concept fundamentally changes the dynamics of the simulation argument. In correspondence with me, Professor Bostrom sheds some light on this question:

‘If we are the only civilization at our stage there will ever have been, then the equation remains true, although some of the possible implications become less striking … the probability that we are not in a simulation is increased if ours is the only civilization that will have ever existed throughout the multiverse’ (Nick Bostrom e-mail, 10 February 2021).

This assertion reinforces the complexity of the simulation argument and the profound effect that our assumptions about the universe have on our interpretations of existence.

CONCLUSION: THE PARADOX OF CREATION

The Simulation Argument presents a curious paradox. The closer we get to the point of being capable of creating our own simulations, the greater the probability that we are living in a simulation ourselves. As we stand on the precipice of creating our virtual realities, we would be faced with the startling possibility that we are simulated beings about to create a simulation.

By abstaining from creating these simulations, we could perhaps decrease the likelihood of us being simulated, indicating that at least one civilisation capable of such feats decided against it. But the moment we dive into creating simulated realities, we would be compelled to accept that we are almost certainly doing so as simulations ourselves.

This paradox inevitably leads to the obvious question: Who created the first simulation? Might that really be us? Such questions punctuate our exploration of the possibility of a simulated reality. The answers may reshape our very understanding of existence itself. But would it ultimately change anything?

Should We Accept Pascal’s Wager?

Introducing the Guardian Principle

A version of this article appears in TWISTED LOGIC: Puzzles, Paradoxes, and Big Questions. By Leighton Vaughan Williams. Chapman & Hall/CRC Press. 2024.

THE ORIGINS OF PASCAL’S WAGER

To understand the significance of Pascal’s Wager in decision-making processes, we must first trace its roots. Blaise Pascal is known for his immense contributions to mathematics and probability theory. One of his notable contributions to philosophy and decision theory, however, was his articulation of what has come to be known as Pascal’s Wager.

PASCAL’S WAGER: THE CRUX OF THE ARGUMENT

The wager posed by Pascal is simple yet profound. It can be paraphrased as follows: If God exists and you wager otherwise, the repercussions can be enormous. On the contrary, if God does not exist and you wager that he does, the implications are trivial in relative terms. Essentially, believing in God could lead to infinite rewards (eternal life in heaven), while the downside if he does not is comparatively inconsequential. Thus, Pascal urges you always to lean to the side of believing in God.

ADDRESSING THE ‘MANY GODS’ OBJECTION

The argument often raised against Pascal’s Wager is the ‘many gods’ objection. Detractors argue that numerous characterisations of God are conceivable, including those that punish believers. However, this counterpoint presumes that all representations of a god are equally plausible, which is an assumption that may not hold.

For instance, the existence of a deity described by a major established religion with millions or even billions of adherents and millennia of theological development and intellectual underpinning could be perceived as vastly more plausible than a fledgling religion with relatively few adherents or consistent theology.

THE ROLE OF HUMAN BIASES AND FUTURE REWARDS

The ability to appreciate uncertainty and the value of future rewards too often gets overshadowed by human biases. Humans are predisposed to discount the future, focusing on immediate rewards and overlooking long-term consequences. This cognitive bias makes people prone to underestimate future risks and rewards. Pascal’s Wager prompts us to consider future implications more seriously, offering a framework to factor in future gains or losses in decision-making.

PASCAL’S WAGER IN CONTEMPORARY CONTEXT: CLIMATE CHANGE AND NOAH’S LAW

The relevance of the thinking behind Pascal’s Wager isn’t confined to theological considerations. A parallel can be drawn, for example, between Pascal’s Wager and the urgency to act against climate change. Even if there were only a slim chance of catastrophic climate disaster, the consequences of inaction, considering the potential existential harm, would be too high to ignore.

This approach to climate change action has been dubbed ‘Noah’s Law’. It reflects the sentiment of Pascal’s Wager: if there’s a chance an ark may be essential for survival, it’s prudent to start building it now, regardless of how sunny the day might seem.

THE GUARDIAN PRINCIPLE

Building upon these concepts, I propose the introduction of a new ethical and operational guideline, which I call the ‘Guardian Principle’. This principle extends the foundational ideas of Noah’s Law and Pascal’s Wager into a broader, more encompassing approach. It advocates for a stance of proactive stewardship over our planet and society, emphasising the importance of pre-emptive action in the face of potential existential threats, not limited to climate change but extending to all manner of such risks.

The Guardian Principle calls for an ethos of precaution and responsibility, urging humanity to act as guardians of its own future and the future of our shared environment. It suggests that in situations of significant uncertainty but potentially devastating outcomes, we should err on the side of caution and engage in preventative measures against a wide array of existential risks. In this way, we fulfil a collective duty to safeguard the well-being of current and future generations against all forms of irreversible harm.

By integrating the Guardian Principle into our global ethos, we expand the narrative from merely avoiding disaster to actively cultivating a safe, sustainable future. It’s a call to not only build arks against impending floods but to seek to prevent the floods themselves, and to act more broadly as vigilant guardians against potential threats. It encourages us not just to react but to anticipate, mitigate, and ideally avert existential risks through foresight, innovation, and collective action.

In this light, the Guardian Principle does not just complement the logic behind Pascal’s Wager and Noah’s Law; it amplifies it. It reinforces the argument that inaction in the face of existential uncertainty is not an option. Instead, we are urged to embrace a more vigilant, proactive approach, turning existential anxiety into a catalyst for holistic and forward-thinking action. In this way, it is a call for a shift in perspective – from reactive measures to a stance that actively seeks to prevent, mitigate, and anticipate risks before they manifest. It’s about building a legacy of sustainability, resilience, and foresight. It is a call to action that resonates with Pascal’s Wager.

PASCAL’S MUGGING: A MODERN SPIN

A modern spin on Pascal’s Wager, Pascal’s Mugging, presents a scenario where a stranger promises a life-changing return on a relatively small sum, or to wield an existentially negative impact if they don’t receive this sum. Even if the chance of the claim being true is infinitesimal, it might seem rational to hand over the sum, such is the scale of the reward or consequence compared to the outlay. It is a dilemma that underscores the need for a pragmatic balance between scepticism and action in the face of uncertainty.

CONCLUSION: PASCAL’S LIGHTHOUSE


Pascal’s Wager has assumed a new critical relevance in our times. With the stakes being higher than ever in terms of global existential risks, the urgency to revisit and appreciate the wager’s lessons has heightened.
While it might seem counterintuitive to expend resources and energy to avert what may be perceived by at least some as small risks, Pascal’s Wager prompts us to think otherwise. As the wager illuminates, the potential stakes of inaction—be it eternal damnation in a theological context or irreversible climate disaster in a worldly sense—may far outweigh the cost of preventive measures. As we steer through a world beset with systemic risks and uncertainties, Pascal’s Wager, and the Guardian Principle it inspires, can serve as a lighthouse, guiding us away from the rocks and towards prudence, long-term thinking, and existential risk management.

US Election 2024: What are the forecasts saying?

Every four years, as the U.S. Presidential election draws near, data enthusiasts eagerly dive into the world of forecasts and predictions, while the broader public often faces a mix of excitement and anxiety. With the election date of November 5th rapidly approaching, everyone is asking the same question: Who is most likely to win? And there is no shortage of forecasters ready to provide their answers.

The Common Ground Among Election Forecasters

At first glance, many election models seem to rely on similar data: state-level polls, national polls, economic indicators, and approval ratings. While these factors offer valuable insights, the U.S. Presidential election ultimately hinges on winning a majority of the 538 electoral votes allocated among the 50 states. As a result, most forecasters focus more heavily on state-level polling, using simulations to estimate each candidate’s probability of winning in individual states, which are then aggregated into a national forecast. This is particularly the case as the election draws nearer, with the broader signals receding into the background. We are not yet at that stage, however.

Key Decisions in Polling Models

As election day nears, forecasters face two critical decisions:

  1. Aggregating State-Level Data: Combining state-level polls and other relevant data to estimate the likelihood of each candidate winning the electoral votes assigned to each state.
  2. Formulating a National Forecast: Deciding how to aggregate these state-level probabilities into a single national outcome, especially when accounting for the potential for correlated polling errors across states.

The weight given to these decisions can vary depending on how close the race is. In a tightly contested race, small adjustments in state-level probabilities can significantly impact the overall forecast. Conversely, in a race where one candidate has a clear lead, a particular challenge is to accurately factor in the possibility of correlated errors, which can heavily influence the probability assigned to less likely outcomes.

Diverse Approaches to Election Forecasting

Currently, several major models are providing live forecasts for the 2024 election, each with its unique methodology:

  • FiveThirtyEight: A well-known model now under new leadership, which has recently refined its methodology and actively adjusts polling inputs based on new data.
  • The Economist: Applies an evolved version of the Votamatic model, which has shifted to a more closed-source approach but still provides a general overview of its methods.
  • Princeton Election Consortium (PEC): Focuses heavily on state-level polling but has been criticized for not adequately addressing the correlation of polling errors across states, an issue that significantly affected its 2016 predictions.
  • PollyVote: Founded in 2004, PollyVote emphasises combining multiple forecasting methods, such as polls, prediction markets, expert judgment, and econometric models, to enhance accuracy. Over time, it has added new components, such as citizen forecasts, continuously refining its approach to align with evidence-based principles.
  • Decision Desk HQ: This model employs a range of machine-learning techniques to predict election outcomes. Although the precise details of its approach are not fully transparent, it ultimately uses a simulation-based forecast similar to other models. Decision Desk HQ is known for its fast reporting on election night, combining its sophisticated modelling with a robust data-gathering network.
  • Data Diary: Data Diary adapts The Economist’s 2020 model into a more fully Bayesian framework, allowing it to handle uncertainty and new information more dynamically. The model uses complex statistical methods to integrate multiple data sources, giving it the flexibility to adjust to shifts in public opinion and other election dynamics.

Three other notable platforms that provide unique insights into election outcomes are Betfair, Polymarket, and PredictIt:

  • Betfair: One of the largest and most established form of prediction market, Betfair allows participants to bet on a wide range of political outcomes. As a major player in the online betting world, Betfair aggregates the collective sentiment of users on numerous markets, including U.S. elections. The market odds fluctuate in real-time based on the volume and direction of bets, providing a “wisdom of the crowd” perspective that reflects public sentiment and market confidence in a candidate’s chances.
  • Polymarket: A blockchain-based prediction market that also offers betting on political events, Polymarket enables users to buy and sell shares in various election outcomes using cryptocurrency as the medium of exchange. Because it is decentralised and leverages blockchain technology, Polymarket tends to attract a tech-savvy audience that may offer different insights from traditional prediction markets. The market’s odds shift based on trading activity, providing a dynamic view of public opinion.
  • PredictIt: An online prediction market launched in 2014, PredictIt allows participants to buy and sell shares on the outcome of political events, using a continuous double auction format. It is very influential, with over 160 academic data-sharing partners.

Current Predictions and Points of Disagreement

Most models agree on a some key points, for example tha: Kamala Harris has improved upon Joe Biden’s polling numbers, and that a very small number of key states are crucial to the overall outcome. However, they diverge on who currently holds the lead. FiveThirtyEight, The Economist, PollyVote, Decision Desk HQ, and the Princeton Election Consortium, currently lean towards Harris, the latter two albeit very marginally. The Silver Bulletin and Data Diary give Trump the edge, albeit only Nate Silver’s Bulletin by more than a whisker. Meanwhile, prediction markets like Betfair and PredictIt currently tilt to Harris, in line with major bookmakers, while Polymarket also leans to Harris, but only by the finest of margins. The much-heralded presidential debate between Harris and Trump, which the Vice President is widely acknowledged to have won decisively, is only just beginning to filter into the polls. It will be fascinating to see what immediate difference, if any, that makes.

Looking Ahead

In the coming weeks, it will be interesting to examine each of these models more closely, assessing their strengths and weaknesses, their historical accuracy, and the reasons behind their different predictions. Until the actual results come in after November 5th, though, that is all they are – forecasts – and some will prove rather more prescient than others. I have my own thoughts on that, but I’ll just say that I tend, based on my own published research, to trust at any point in time the market aggregate, if I’m pushed to choose, over the outlier. 

Harris v Trump: The First Presidential Debate

The first US presidential debate of this latest election cycle proved to be a high-stakes battle that may have significantly altered the landscape of the race. From the moment Vice President Kamala Harris, the Democratic nominee, confidently approached Donald Trump and extended her hand, the dynamics of the evening began to shift dramatically. The betting markets, which initially favoured Trump, soon started to move in Harris’s direction. But it wasn’t until Trump made the unsupported claim that immigrants in Ohio were “eating dogs and cats” that the betting markets saw a full crossover, favouring Harris as the likely winner of November’s election.

Harris Takes Control

Harris knew a lot was riding on this debate. With a tightening race and her momentum stalling in the days leading up to the event, she needed to make a strong impression on the many Americans who still felt they didn’t know her well. Presidential debates are known for providing a unique platform for relatively less exposed candidates to boost their visibility, and Harris seized this opportunity with a well-executed strategy.

In her prosecutorial style, Harris managed to control the debate’s tempo from the start, putting Trump on the defensive. Her pointed questions and direct eye contact were reminiscent of her days as a prosecutor. Early on, she achieved a key objective: getting under Trump’s skin. By highlighting his dwindling rally attendance and suggesting his speeches had become dull and boring, she struck a nerve. From that moment on, Trump struggled to stay focused, frequently veering off into bizarre claims and tangents.

Trump’s Performance: Unhinged and Unstable?

Trump’s performance quickly turned chaotic. His outlandish statements—such as the claim about immigrants eating pets in Ohio and his unsupported assertions about post-birth killings in Democratic states—only served to further unravel his argument. Each time Harris poked him, he took the bait, spiralling deeper into his grievances about the 2020 election and other well-worn talking points. For viewers at home, the contrast was stark: a calm and composed Harris against a visibly agitated Trump.

Even Trump’s attempt to pivot to his foreign policy record backfired. Harris pointedly mocked his admiration for Vladimir Putin and highlighted the potential consequences of a Trump presidency for Ukraine and Eastern Europe. When Trump refused to clearly state his position on Ukraine winning the war, it only seemed to add to a general sense that he was on shaky ground.

The Betting Markets Shift Decisively

The impact of the debate on the betting markets was immediate and decisive. As Trump continued to dig himself into a hole with every rambling answer, Harris’s position strengthened. By the halfway point of the debate, the markets were giving her a 97% chance of being declared the winner in post-debate polling. Her dominant hold only solidified further as the debate moved on to its final conclusion, leaving her as the clear election favourite going forward.

A Second Debate?

Now, speculation turns to whether there will be a second debate. Harris is reportedly keen for another round, while Trump appears less enthusiastic. The question looms: who stands to gain more from another showdown? With expectations now so low for Trump, even a modest improvement could help him. On the other hand, another performance similar to the first could be potentially fatal for his campaign. 

So will there be a second debate? And if so, can Trump change the narrative, or will Harris deliver another blow to his campaign? 

What Lies Ahead?

While Harris emerged from the debate as the clear winner, there are still challenges ahead. Some undecided voters felt she wasn’t specific enough on policy positions, choosing instead to focus on making Trump implode and outlining her broader vision for the country. However, she did manage to effectively argue against Trump’s suitability for the presidency, a critical move as voters in key states prepare to cast their ballots.

Debates often have a short-lived impact, as seen in Trump’s 2016 first debate against Hillary Clinton, where he was widely perceived to have lost but went on to win the election. Nevertheless, Harris’s strong performance may have come at a crucial moment. The race is far from over, but this debate could mark a pivotal turning point in her campaign’s favour.

Can we trust political betting markets?

A version of this article was first published in The Conversation UK

Election Betting

Records of the betting on US presidential elections can be traced back to 1868. Since then, no clear favourite for the White House had lost before 2016, except in 1948, when the 8 to 1 longshot and sitting president, Harry S. Truman, famously defeated his Republican rival, Thomas E. Dewey.

In 2016, the exception was repeated when Hillary Clinton, trading at 7 to 2 on (equivalent to a win probability of about 78%) as polls opened, lost in the electoral college to Donald Trump. In so doing, Trump defied not just the polls and the experts, but the “wisdom of the crowd” as displayed in the betting markets.

Trump achieved this by converting a near 3 million vote loss in the popular vote into a victory by 77 votes in the electoral college. In a larger sense, it might be said that crowd wisdom was trumped by the arcane US electoral system.

There was a similar consensus in the run-up to the 2020 election that Trump would lose – but the degree of confidence displayed by the markets and the models diverged markedly. To illustrate, Sporting Index, the spread betting company, announced it thought Joe Biden would win with between 305 and 311 electoral votes as the polls opened on election day, with Trump trailing on 227 to 233 electoral votes.

Taking the mid-points of these spreads, this equated to a Biden triumph by 308 votes to 230 in the electoral college – a majority of 78. Similar estimates were contained or implicit in the odds offered by other bookmakers, betting exchanges and prediction markets.

Forecasting Models

Meanwhile, other major forecasting models were much more bullish about Biden’s prospects. Based on 40,000 simulations, the midpoint estimate of the model provided by Nate Silver and FiveThirtyEight put Biden ahead by 348 electoral college votes to 190 for Trump, a margin of 158. The New Stateman model made it 339 votes to 199 in favour of Biden. The Economist’s model was even more lopsided in favour of Biden, estimating that he would prevail by 356 electoral votes to 182. Taking the unweighted mean of all three forecasting models, Biden was projected to win 348 votes in the electoral college to 190 for Trump.

The other go-to place for expert opinion with a long track record of solid performance (except in 2016) is Sabato’s Crystal Ball based at the University of Virginia’s Center for Politics. It was projecting Biden to win the electoral college by 321 votes to 217. The PollyVote project goes a step further, combining information contained in betting markets with forecasting models, experts and beyond. It forecast a Biden victory by 329 electoral votes to 209.

Last bets please 

When the dust had finally settled, literally and figuratively, Biden ended up with 306 votes in the electoral college to 232 for Trump. As such, the betting spreads were almost spot on. In fact, both these numbers were within the spreads offered on election day.

What this tells us is that the betting and prediction markets, which respond to the weight of money traded on each candidate, and are informed by considerable professional insight, recovered in 2020 a reputation dating back to at least 1868, and in the case of the Papal betting markets as far back as 1503.

Interestingly, a couple of weeks after declaration of all election results, Trump still merited a 7.8% chance of clinging on to office, according to the betting exchange trading. This factored in all the ways he might seek to reverse the declared results. January 6th was still a little while off. 

I asked at the time whether it was likely that he would prevail over all established custom and evidence. Not at all, I ventured. Was it possible? Yes, I thought it was. In the event, 7.8% was, at the time, probably about right.

Can we believe Nate Silver?

A Balanced View

As a long-time follower of political forecasts, and creator of some, I’ve often found Nate Silver’s insights both intriguing and valuable. Silver, the founder of FiveThirtyEight and now running his own platform, The Silver Bulletin, has built a reputation as one of the most influential data analysts in politics. But with the 2024 U.S. presidential election approaching, I find myself asking: Can we still trust Nate Silver’s predictions?

Understanding Why Polling Aggregators Differ

To answer this question, it’s important to understand why different polling aggregators might come to different conclusions, even when using similar data. Most aggregators, such as FiveThirtyEight, The Economist, and The Silver Bulletin, rely on state-level polls, national polls, approval ratings, and economic fundamentals to make their predictions.

However, the differences arise in two critical areas of judgment:

  1. How state-level polls are aggregated to determine each candidate’s chances of winning a state’s electoral votes. In a tight race like 2024, small adjustments in key states can significantly affect the overall probability.
  2. How these state-level probabilities are aggregated into an overall probability of winning the election, especially considering correlated polling errors across states. This was crucial in 2016 when correlated polling errors across states led to a surprise outcome.

Silver’s approach has drawn attention this cycle because of how it handles these judgments, and the resulting forecasts sometimes differ significantly from other models.

The “Convention Bounce” Adjustment: Reasonable or Overdone?

One of the most debated aspects of Silver’s current methodology is his use of a “convention bounce” adjustment. This adjustment accounts for the temporary bump candidates often receive in the polls following their party conventions. Silver argues that polls taken just after the Democratic National Convention (DNC) might temporarily overstate Kamala Harris’s support, so his model includes an adjustment to correct for this.

On the surface, this seems like a logical adjustment. However, it has sparked some debate. For example, Silver’s model recently moved more than three percentage points against Harris during a period when no significant new data emerged—no major events, no new battleground state polls, no economic changes. While other models, like The Economist’s or Decision Desk HQ’s, remained stable during the same period, reflecting the lack of new information, Silver’s swung noticeably.

This has led some observers to question whether the “convention bounce” adjustment is appropriately calibrated. Silver defends the adjustment as “highly defensible” and notes that it will phase out gradually, but this episode does raise questions about how such adjustments impact the overall forecast.

Is Silver’s Model Overly Reactive to Limited Information?

Another critique is that Silver’s model appears unusually reactive, responding to minor changes in data with relatively large movements. In a situation where there’s no significant new information, most models would show minimal movement. Yet, Silver’s has shifted markedly at times.

Silver’s supporters might argue that his model is designed to be highly responsive to any new data, reflecting the inherent uncertainty and volatility of the electoral landscape. However, this reactivity has led to speculation that the model might be more prone to showing movement, potentially to maintain engagement, even when the underlying data doesn’t fully justify such shifts.

The underlying question remains: Is this reactivity a strength, showing that Silver’s model adapts quickly to any new information, or a weakness, suggesting it might overemphasise minor fluctuations?

The Electoral College Focus: An Essential Consideration or Overplayed?

Silver also places considerable emphasis on the challenges posed by the Electoral College. He points out that even if Harris wins the popular vote by a few points, she could still lose the presidency due to the distribution of votes across key states.

This focus on the Electoral College is certainly valid; history has shown that a candidate can win the popular vote but lose the presidency. However, some critics question whether Silver’s emphasis on this factor might be overstating its impact compared to other models. Every major aggregator considers the Electoral College in their forecasts, but Silver’s model arguably amplifies its importance relative to the others.

Here, too, there are two sides. Silver’s supporters argue that his focus on the Electoral College is a necessary reminder of how U.S. elections actually work, ensuring we don’t overlook its impact. On the other hand, some might see it as a way to make his model appear more unique or insightful than it truly is.

Conclusion: Should We Trust Nate Silver’s Forecast in 2024?

So, where does that leave us? Can we trust Nate Silver’s model for the 2024 election? On the one hand, Silver’s experience, his focus on “tail risks” (scenarios where less likely but still possible outcomes could occur) and correlated polling errors, make his forecasts a valuable tool in understanding the electoral landscape. On the other hand, some of his adjustments and the model’s reactivity have raised questions about whether his approach this cycle might be about engagement as much as precision.

Ultimately, it may be wise to consider multiple perspectives. Nate Silver’s forecasts should certainly be a part of that consideration, but not the only voice in the conversation. Given the complexities of this election, looking at what other models and aggregators are saying—and understanding their different methodologies and assumptions—might offer a more balanced view.

Looking Ahead

As we get closer to Election Day, it will be interesting to see how Silver’s model and others evolve with new data. Staying informed through multiple sources and understanding their approaches will be key to navigating this race. Ultimately, no single model is likely to have all the answers, or even to be asking all the right questions.

When Should We Flip a Card?

Exploring the Four Card Problem

A version of this article appears in TWISTED LOGIC: Puzzles, Paradoxes, and Big Questions. By Leighton Vaughan Williams. Chapman & Hall/CRC Press. 2024.

The Four Card Problem

The Four Card Problem, also known as the Wason selection task, is a captivating puzzle that tests our logical reasoning abilities. Invented by Peter Cathcart Wason, this task challenges us to determine the minimum number of cards required to verify or falsify a given statement. Let’s look deeper into this intriguing problem.

The Scenario: Card Setup

Imagine being presented with four cards, each displaying either a letter or a number. These cards lay the foundation for the puzzle, providing the information necessary to reach a conclusion. Let’s examine an example:

The face-up sides of the cards show: 23; 28; R; B

Each card has a letter on one side and a number on the other side.

Alongside these cards, you are given a statement: ‘Every card with 28 on one side has R on the other side’.

Determining the Minimum Number of Cards

Now, the crucial question arises: How many cards must you turn over to determine the truthfulness of the given statement? And which specific cards should you investigate?

Common Misconceptions

At first glance, the task might appear deceptively simple. Many individuals are inclined to turn over the R card, assuming it holds the key to verifying the statement. However, this line of thinking is misguided. Regardless of what is on the other side of the R card, it does not contribute to determining whether every card with 28 on one side has R on the other.

Similarly, the inclination to turn over the 23 card is also misleading. Even if the 23 card reveals an R on its other side, it does not provide any insight into the truthfulness of the statement. The existence of R on the opposite side of the 23 card merely confirms that the statement ‘Every card with 23 on one side has B on the other side’ is false. It does not shed light on the validity of the statement regarding the 28 card and R.

The Key to Solving the Puzzle: Logical Analysis

To arrive at the correct solution, we must identify the cards that have the potential to disprove the given statement. The crucial observation lies in recognising that only a card displaying 28 on one side and something other than R on the other side can invalidate the statement.

In this scenario, the cards we need to focus on are the 28 card and the B card. Let’s explore the reasoning behind this.

The Correct Solution: Minimum Number of Cards

The Card with 28 on Its Face-Up Side: This is the most direct test of the statement. If the other side is not R, the statement is false.

The Card with B on Its Face-Up Side: This card needs to be checked because if the other side is 28, it would contradict the statement. The statement only mentions what is on the other side of 28, not what is on the other side of R.

The cards with 23 and R on their face-up sides do not need to be checked. The card with 23 is irrelevant to the rule, which only concerns 28. The card with R does not need to be checked because the rule does not specify what should be on the other side of R.

So, you only need to turn over two cards: the one showing 28 and the one showing B.

Conclusion: Thinking beyond Initial Assumptions

The Wason selection task, or the Four Card Problem, immerses us in the intricacies of logical analysis and conditional reasoning. By identifying the two necessary cards to flip, the 28 and the B, we confront the task’s real challenge, and learn the importance of testing for falsification rather than confirmation.

The puzzle serves as a powerful reminder of the complexities that lie beneath seemingly simple tasks and the importance of careful analysis when engaging in logical problem-solving. It challenges us to think beyond initial assumptions and consider the logical implications hidden within the given information. As such, it is a clear reminder of the complexities hidden within seemingly straightforward problems and the value of meticulous analysis in navigating the world of logic.