Skip to content

The Three Caskets Problem

A Shakespearean Puzzler

A version of this article appears in TWISTED LOGIC: Puzzles, Paradoxes, and Big Questions, by Leighton Vaughan Williams. Chapman & Hall/CRC Press. 2024.

THE THREE CASKETS PROBLEM

The narrative of William Shakespeare’s ‘Merchant of Venice’ contains intrigue around the character of the young heiress Portia. Amid the various plot developments, one of the more fascinating elements of the story lies in a puzzle set for anyone seeking her hand in marriage. Three caskets made of gold, silver, and lead each contain a different item. Only one holds the prize, a miniature portrait of Portia which symbolises the route to her heart. Portia alone knows that the portrait’s true location is in the lead casket.

SUITORS AND THE CRYPTIC CASKETS: UNRAVELLING THE PECULIAR TEST

As the story unfolds, we learn that to claim Portia in holy wedlock a suitor must choose the casket that houses her portrait. Each casket comes engraved with a cryptic inscription, adding a layer of interest and sophistication to the task.

THE ALLURING GOLD: THE FIRST SUITOR’S TEST

The Prince of Morocco steps forward to face this intriguing test. He is confronted with the inscriptions on the caskets, each one at least as cryptic as the others. Drawn by the promise of desire inscribed on the gold casket, ‘Who chooseth me shall gain what many men desire’, he chooses it, hoping to find ‘an angel in a golden bed’. His dreams are shattered when he finds a skull and a cryptic scroll, instead of the image of Portia. The message on the scroll serves as a harsh reminder, ‘All that glisters is not gold’. With a heavy heart, he retreats, leaving Portia with a sigh of relief, uttering, ‘A gentle riddance’.

SILVER’S DECEPTION: THE SECOND SUITOR’S TURN

Emboldened by his self-worth, and unaware of which casket his predecessor had chosen, the Prince of Arragon interprets the inscription on the silver casket, ‘Who chooseth me shall get as much as he deserves’, as a validation of his worthiness. He selects this casket.

ADDING COMPLEXITY: THE PUZZLE TAKES A TWIST

Now let’s indulge in a thought experiment by introducing an intriguing layer to this complex puzzle. Suppose that, after Arragon’s selection of the silver casket, Portia must open one of the remaining caskets without revealing the portrait’s location. She must, therefore, open the gold casket, which she knows does not contain her likeness. This presents Arragon with the opportunity to hold to his initial choice, the silver casket, or switch to the remaining, unopened casket made of lead.

WEIGHING THE ODDS: ARRAGON’S PROBABILITY PARADOX

If Arragon believes that Portia’s knowledge of the caskets is equal to his, should he stick with his initial choice or take a chance on the unopened lead casket? His decision is far from straightforward, hinging on his interpretation of the cryptic inscriptions, his understanding of the shifting probabilities, and his perception of Portia’s actions.

THE PROBABILITY PUZZLE: DECIPHERING THE GAME OF CHANCE

To understand the implications of the new development, we must first delve into the realm of probability. At the outset, Arragon’s initial choice, the silver casket, had a one-third chance of being correct, assuming he has no other information. There is, therefore, a two-thirds probability that the portrait lay in one of the other two caskets.

Portia’s revelation that the gold casket doesn’t contain the portrait effectively shifts these odds if we can assume that she knows which of the caskets contains her portrait, and must not reveal it. The two-thirds chance, which was initially split between the gold and lead caskets, now converges entirely on the lead casket. Consequently, if Arragon changes his choice from the silver casket to the lead one, his probability of finding Portia’s portrait doubles from one-third to two-thirds, other things being equal.

FATEFUL DECISION: TO SWITCH OR NOT TO SWITCH

If he dismisses the inscriptions as mere distractions and recognises the probability shift in favour of the lead casket, then switching seems like the most rational move. However, if he believes that he has deciphered the true meaning of the inscriptions, he might decide to stick with his original choice.

ARAGON’S DECISION: TO OPEN THE SILVER CASKET

Arragon is either unaware of the true probabilities or else is swayed by the cryptic clues. He chooses the silver casket. However, it only harbours disappointment. Instead of Portia’s portrait, he discovers an image of a fool and a note mocking his decision, ‘With one fool’s head I came to woo, But I go away with two’. His self-confidence leads to his downfall, leaving him more foolish than when he first arrived.

THE POWER OF THE INSCRIPTIONS: GUIDE OR DISTRACTION?

The inscriptions on the caskets add an extra layer of uncertainty and complexity to Arragon’s decision-making process. They could be seen as guides leading the suitors to the correct choice, or they could be deceptive distractions meant to confuse and mislead. The inscription on the lead casket, ‘Who chooseth me must give and hazard all he hath’, could be perceived as a warning of the risks involved or as a subtle hint about the potential rewards of choosing what appears to be the least valuable casket.

CONCLUSION: THE POWER OF INFORMATION

In this thought experiment, the key element is the new information introduced by Portia when she opens the gold casket. After all, she knows where the portrait is. This single action has the potential to increase significantly Arragon’s chance of success. If Arragon understands and acts upon this new information, he can potentially improve his chances of selecting the correct casket from one in three to two in three. However, this seemingly simple shift in probability is complicated by the presence of other potentially influential factors, such as the cryptic inscriptions on the caskets. This makes the problem different from the basic Monty Hall decision.

He might also believe that Portia has no idea which casket contains the portrait. In that case, by opening the gold casket, she would be adding no information to what Arragon already has. He may as well be guided by any additional information he thinks he might pick up from the cryptic inscriptions. Either way, he faces a lonely but life-altering decision.

The Death Row Problem

When Should We Expect Mercy?

A version of this article appears in TWISTED LOGIC: Puzzles, Paradoxes, and Big Questions. By Leighton Vaughan Williams. Chapman & Hall/CRC Press. 2024.

THE SETTING

The setting is a prison where three inmates—Amos, Bertie, and Casper—are awaiting the hangman’s noose. The warden, in an act of clemency to celebrate the King’s birthday, will grant clemency to just one of the three prisoners. The choice of which of these inmates to pardon is made randomly, with each name placed in a hat and drawn out. The warden now knows whom he will pardon, but the men on death row do not.

THE REQUEST

Amos makes a request to the warden. He asks the warden to name a prisoner who will NOT be pardoned, without revealing his (Amos’s) own fate. If the warden has chosen Bertie to be granted clemency, he should name Casper as one of the doomed. If it’s Casper who has been pardoned, the warden should name Bertie to be executed. If Amos himself is to be pardoned, the warden should simply toss a coin and name either Bertie or Casper as one of the doomed.

Amos’s Request: It’s essential to note here that Amos’s request is based on the assumption that the warden will not reveal if Amos himself is the pardoned prisoner.

THE WARDEN’S RESPONSE

The warden agrees to the request from Amos and reveals that Casper is not the pardoned prisoner.

WHAT DOES THIS MEAN FOR AMOS AND BERTIE?

With this new information to hand, each of the prisoners can re-evaluate their chances. Initially, Amos believes that his chance of a pardon is 1/3, but with Casper out of the running, he believes that his odds of clemency have risen to 1/2. But is he right in this belief?

RE-EVALUATING THE ODDS

Initially, the odds are 1/3 for each prisoner because only one of the three prisoners is chosen at random to be pardoned. However, when the warden reveals that Casper will not be pardoned, Amos gains new information but not about his own fate. There’s no new information regarding his own fate, so his chances remain as they were, at 1/3. Meanwhile, Bertie’s odds of being pardoned have now increased to 2/3.

WHY ARE THE ODDS DIFFERENT?

This difference in the odds between Amos and Bertie might seem to be counterintuitive. How can they both receive the same information, yet have different survival odds? The answer lies in the warden’s selection process. The warden would not have revealed Amos as the condemned prisoner due to Amos’s unique request, but he might have revealed Bertie as such, instead of Casper. The fact that he doesn’t name Bertie when he might have done so indicates that Bertie’s chances of being pardoned have increased, while nothing has changed for Amos. Amos’s belief that his odds have increased to 1/2 is a misconception.

A LARGER SCENARIO

If this still seems puzzling, consider a larger group of 26 prisoners. If Amos asks the warden to name 24 condemned prisoners in random order without revealing his own fate on any occasion, each prisoner initially has a 1/26 chance of being pardoned. But every time a doomed prisoner is named, the chance that each of the remaining prisoners (except for Amos) will be pardoned increases.

Once every prisoner but Bertie has been named as condemned, Amos’s chances of survival remain at 1/26. However, Bertie’s odds of being pardoned have now increased to 25/26, even though only two prisoners remain unnamed by the warden.

So, even though it might seem like Amos has a very good chance of being pardoned, the reality is that his odds have not changed and remain at 25/1, representing a chance that Amos will escape the noose of 1/26.

CONCLUSION: THE KEY TAKEAWAY

The Three Prisoners Problem highlights the importance of understanding the method by which we obtain information and its impact on the probabilities. It’s a fascinating exploration of conditional probability that shows how the same piece of information can affect the chances of two individuals differently, based on the process by which that information was revealed. As such, it is a classic example of how counterintuitive probability can be, especially in situations where information is revealed in a conditional manner.

The Monty Hall Problem

When Should We Change Our Mind?

A version of this article appears in TWISTED LOGIC: Puzzles, Paradoxes, and Big Questions. By Leighton Vaughan Williams, Chapman & Hall/CRC Press. 2024.

THE GENESIS OF THE MONTY HALL PROBLEM

The Monty Hall Problem was named after the original host of the American game show, ‘Let’s Make a Deal’. It became a topic of popular debate because of the answer provided to a question quoted in a column in Parade magazine.

Thanks for reading Twisted Logic! Subscribe for free to receive new posts and support my work.

Subscribed

The concept is that contestants are given a choice of three doors. Behind one door lies a highly desirable prize like a car, while behind the other two doors were much less desirable prizes like goats. The car is placed randomly behind one of the doors, preventing contestants from predicting its location based on prior observations or information.

THE PUZZLE UNVEILED

Imagine yourself on this game show. You are asked to choose one of three doors (let’s call them Doors 1, 2, and 3). After making your choice (let’s say you choose Door 1), the host, who knows what’s behind each door, opens another door (for instance, Door 3) to reveal a goat.

He then offers you a choice. You can stick with your original decision (Door 1 in this case), or you can switch to the remaining unopened door (Door 2). You should note that the host always opens a door that you didn’t choose and that hides a goat, increasing the suspense and making the game more interesting.

The question that the Monty Hall problem asks is: Should you stick with your original choice, or should you switch to the other unopened door?

THE COUNTERINTUITIVE ANSWER

At first glance, it might seem like your odds of winning the car are the same whether you stick to your original choice or switch. After all, there are only two doors left unopened, so isn’t there a 50% chance that the car is behind each of them?

In her column, Marilyn Vos Savant argued that the chance is not 50% either way, but that you have a higher chance of winning the car if you decide to switch doors. Despite receiving numerous objections from readers, including some leading academics, her answer holds up under scrutiny. Here’s why.

When you first choose a door, there is a 1 in 3 chance that it hides the car. This means that there’s a 2 in 3 chance that the car is behind one of the other two doors. Even after the host opens a door to reveal a goat, these probabilities do not change. Monty is simply providing more information about where the car is not.

So, if you stick with your original choice, your chances of winning the car remain at 1 in 3. However, if you switch, your chances increase to 2 in 3. Switching doors effectively allows you to select both of the other doors, doubling your odds of finding the car.

A CLOSER LOOK AT THE PROBABILITIES

Let’s examine the situation more closely to understand how this works.

If the car is behind Door 1, and you choose it and stick with your choice, you win the car. The chance of this happening is 1/3.

If the car is behind Door 2 and you initially choose Door 1, the host will open Door 3 (since it conceals a goat). If you switch to Door 2, you win the car. The chance of this happening is 2/3.

If the car is behind Door 3, and you initially choose Door 1, the host will open Door 2 (since it conceals a goat). If you switch to Door 3, you win the car. The chance of this happening is also 2/3.

From the above, you can see that you have a 2/3 chance of winning if you switch to whichever door Monty has not opened, and a 1/3 chance of winning if you stick to your initial choice.

THE ROLE OF THE HOST

It’s crucial to note that the host’s knowledge and actions play a pivotal role in these probabilities. If the host didn’t know what was behind each door or randomly chose a door to open, then the odds would indeed be 50–50, as he might have inadvertently opened a door to reveal the car. However, because the host always opens a door you didn’t choose and always reveals a goat, the odds shift in favour of switching doors.

To expand upon this, consider a version of the problem with 52 cards. This time, you’re invited to choose one card from a deck of 52. The objective is to select the Ace of Spades from a deck of cards lying face down on the table.

If you initially choose the Ace of Spades and stick with your choice, you win the game. The chance of this happening is 1/52, since there’s only one Ace of Spades in a 52-card deck.

However, if you initially choose any card other than the Ace of Spades (which has a 51/52 chance), the host, knowing where the Ace of Spades is, will begin to turn cards over one at a time, always leaving the Ace of Spades and your initial card choice in the remaining face-down deck. The host will continue to do this until only your card and one other card remain. One of these two cards will be the Ace of Spades.

At this point, there is still a 1/52 chance that your original card is the Ace of Spades. If you switch your choice to the remaining card, the chance that it will be the Ace of Spades is therefore 51/52, which is a much higher probability than if you stick with your initial choice.

This works because the host each time deliberately turns over a card that is not the Ace of Spades. So the other card left face down at the end is either the Ace of Spades, with a chance of 51/52, or else your original choice is the Ace of Spades, with a probability of 1/52.

If the host doesn’t know where the Ace of Spades is located, he might inadvertently reveal it every time he turns a card over, so he would be providing no new information about the location of the Ace of Spades by exposing a card.

This shows how the Monty Hall problem can scale to larger numbers. The initial odds of choosing the Ace of Spades are 1/52, but if you switch your choice after the host takes away all but one of the other cards, your odds improve dramatically to 51/52. This is a counterintuitive result, but it follows from the fact that the host’s actions (because he knows where the Ace of Spades is) give you additional information about where the Ace of Spades is not.

OVERCOMING INTUITION WITH LOGIC

The Monty Hall problem can be difficult to grasp because it seems to contradict our intuition. The human brain tends to simplify complex situations, and when there are two unopened doors, it’s easy to fall into the trap of assuming there’s a 50% chance of winning either way. However, the Monty Hall problem highlights how understanding probability requires careful thought and a logical analysis of the situation.

EXPLORING THE MONTY HALL PROBLEM WITH SIMULATIONS

If you’re still having trouble grasping the Monty Hall problem, you might find it helpful to see it in action. Numerous online simulators let you play the Monty Hall game repeatedly, and over time, you’ll see that switching doors indeed wins about 2/3 of the time.

THE MONTY HALL PROBLEM IN POPULAR CULTURE

The Monty Hall problem has seeped into popular culture, appearing in films, television series, and even songs. It serves as a reminder that intuition and probability sometimes have a complicated relationship. The logical and statistical reasoning involved in this puzzle, as well as its seemingly paradoxical result, have made it a favourite topic in probability and statistics classes across the world.

CONCLUSION: PROBABILITY AND INTUITION

The Monty Hall problem is a captivating illustration of how probability can sometimes be counterintuitive. Although it’s been debated, analysed, and confirmed many times over, it continues to intrigue and perplex. It provides a clear lesson: intuition isn’t always reliable when it comes to probability.

When Should We Double Up?

Exploring the Martingale Betting Strategy

A version of this article appears in TWISTED LOGIC: Puzzles, Paradoxes, and Big Questions. By Leighton Vaughan Williams. Chapman and Hall/CRC Press. 2024.

The Martingale betting strategy is based on the principle of chasing losses through progressive increase in bet size. To illustrate this strategy, let’s consider an example: A gambler starts with a £2 bet on Heads, with an even money payout. If the coin lands Heads, the gambler wins £2, and if it lands Tails, they lose £2.

In the event of a loss, the Martingale strategy dictates that the next bet should be doubled (£4). The objective is to recover the previous losses and achieve a net profit equal to the initial stake (£2). This doubling process continues until a win is obtained. For instance, if Tails appears again, resulting in a cumulative loss of £6, the next bet would be £8. If a subsequent Heads occurs, the gambler would win £8, and after subtracting the previous losses (£6), they would be left with a net profit of £2. This pattern can be extended to any number of bets, with the net profit always equal to the initial stake (£2) whenever a win occurs.

CHASING LOSSES AND THE LIMITATIONS

While the Martingale strategy may appear promising in theory, it is important to recognise its limitations and the inherent risks involved. The strategy involves chasing losses in the hope of recovering them and generating a profit. However, it’s crucial to understand that the expected value of the strategy remains zero or even negative.

The main reason behind this lies in the presence of a small probability of incurring a significant loss. In a game with a house edge, such as in a casino, the odds contain an edge against the player. The house edge ensures that, over time, the expected value of the bets is negative. Therefore, even with the Martingale strategy, which aims to recover losses, the expected value of the bets remains unfavourable.

Moreover, in a casino setting, there are structural limitations that impede the effectiveness of the Martingale strategy. Most casinos impose limits on bet size. These limits prevent gamblers from doubling their bets indefinitely, even if they have boundless resources and time, thereby constraining the strategy’s potential for recovery.

THE DEVIL’S SHOOTING ROOM PARADOX

A parallel thought experiment known as the Devil’s Shooting Room Paradox adds an intriguing twist. In this scenario, a group of people enters a room where the Devil threatens to shoot everyone if he rolls a double-six. The Devil further states that over 90% of those who enter the room will be shot. Paradoxically, both statements can be true. Although the chance of any particular group being shot is only 1 in 36, the size of each subsequent group in this thought experiment is over ten times larger than the previous one. Thus, when considering the cumulative probability of being shot across multiple groups, it surpasses 90%.

Essentially, the Devil’s ability to continually usher in larger groups, each with a small probability of being shot, ultimately results in the majority of all the people entering the room being shot.

A key assumption underlying the Devil’s Shooting Room Paradox is the existence of an infinite supply of people. This assumption aligns with the concept of infinite wealth and resources often associated with Martingale-related paradoxes. Without a boundless supply of individuals to fill the room, the cumulative probability of over 90% cannot be definitively achieved.

The Devil’s Shooting Room Paradox serves in this way as another illustration of how probabilities and cumulative effects can lead to counterintuitive outcomes.

CONCLUSION: THE LIMITS OF A MARTINGALE STRATEGY

The Martingale strategy is based on chasing losses, but its expected value remains zero or negative due to the house edge. The strategy’s viability is further diminished by limitations on bet size in real-world casino scenarios. As such, the Martingale system cannot be considered a winning strategy in practical gambling situations. The Devil’s Shooting Room Paradox further demonstrates the complexities and counterintuitive outcomes that can arise when infinite numbers are assumed. Ultimately, a comprehensive understanding of these paradoxes provides valuable insights into the rationality of betting strategies and decision-making in the realm of gambling.

When Should We Expect to be Kicked by a Horse?

Exploring the Poisson Distribution

A version of this article appears in TWISTED LOGIC: Puzzles, Paradoxes, and Big Questions. By Leighton Vaughan Williams. Chapman & Hall/CRC Press. 2024.

A STATISTICAL TOOL

Thanks for reading Twisted Logic! Subscribe for free to receive new posts and support my work.

Subscribed

The Poisson distribution, inspired by the work of Siméon Denis Poisson, is a statistical concept that is particularly useful for helping us understand events that occur infrequently. It indicates the number of such events we can expect to occur in a fixed interval if we know the average rate at which they arrive. In simpler terms, if you want to predict how often something will happen over a certain period, and this event is infrequent, the Poisson distribution can be your go-to method for making this prediction.

This distribution finds practical applications in various fields, ranging from studying historical events to analysing everyday situations and even sports.

UNDERLYING ASSUMPTIONS OF THE POISSON DISTRIBUTION

The accuracy and applicability of the Poisson distribution hinge on several key assumptions:

Independence of Events: Each event must occur independently of the others. This means the occurrence of one event does not affect the probability of another event occurring.

Constant Average Rate: The events are expected to occur at a constant average rate. In other words, the average number of events per unit of time or space remains consistent throughout the period being considered.

Random Occurrence: The events occur randomly, without any predictable pattern or structure. This randomness is crucial for the Poisson model to provide accurate predictions.

Discrete Events: The events are distinct and countable. For instance, the number of emails received per day or the number of accidents at a particular intersection per month.

Understanding these assumptions is vital for correctly applying the Poisson distribution. It is most effective in situations where these conditions are met, such as modelling the number of meteor showers observed in a year, counting the number of times a rare bird is spotted in a forest, or predicting the number of cars passing through a toll booth in an hour.

It’s also very useful in predicting how likely you are to be kicked by a horse next week! The next section explains.

PREDICTING RARE EVENTS: PRUSSIAN CAVALRY OFFICER DEATHS

Let’s travel back in time to the 19th century, when the Poisson distribution was used to study a particular historical event. During this period, researchers were interested in understanding the number of Prussian cavalry officers who were kicked to death by horses in different Army regiments over a span of 20 years. This unfortunate occurrence was relatively rare, but was it random, or were there some underlying factors influencing their occurrence?

Enter Ladislaus Bortkiewicz, an economist and statistician. Bortkiewicz collected data from 14 corps over 20 years, which resulted in observations of yearly numbers of deaths per corps. Using the formula associated with the Poisson distribution, he was able to predict the number of such deaths in specific time intervals. These fitted quite closely to the observed data, indicating that the deaths were indeed random events, and nothing more mysterious or sinister.

This application of the Poisson distribution became a textbook example of real-world events that can be modelled as Poisson processes, which include radioactive decay, arrival of emails, number of phone calls received by a call centre, etc. The deaths of Prussian cavalry officers are an early example of a statistical study in the field of survival analysis.

WORLD WAR II BOMBING RAIDS

During the Second World War, a British statistician named R.D. Clarke used this method to study where the new V-1 ‘flying bombs’ were falling in London. He wanted to figure out if the German military was successfully targeting specific areas or if the bombs were falling randomly. This was strategically important information. It was clear that the V-1s sometimes fell in clusters. The question was whether this could be expected from random chance or whether precision guidance was at play.

To find out, Clarke divided London into thousands of small, equal-sized areas. He assumed to start with that each area had the same small chance of being hit by a bomb. This situation was similar to playing a game many times where you ‘win’ only infrequently. Clarke’s calculations showed that the number of bomb hits in each area matched what the Poisson distribution predicted for random hits. This meant that where the bombs fell seemed to be a product of chance, not because specific areas were targeted.

FROM HISTORY TO FOOTBALL: PREDICTING GOAL SCORING

In football, goals are a relatively infrequent event within the setting of a match, and so are suitable for the application of the Poisson distribution. This provides a simple and effective tool to examine and predict the likely incidence of goals in a match, based on historical data and average goal rates.

Consider, say, a match between two teams, one with an average goal rate of 1.6 goals per game and the other with an average goal rate of 1.2 goals per game. The Poisson distribution allows us to calculate the probabilities of various goal-scoring outcomes for this specific match.

For example, by examining the historical data and applying the Poisson distribution, analysts can estimate the probability of a goalless draw, a 1-1 draw, a win for either team, or any other scoreline based on the average goal rates of the teams involved.

More generally, the Poisson formula allows us to calculate the chance of observing a specific number of events of this kind when we know how often they usually occur on average. It considers the average rate and calculates the probability of obtaining the specific number we’re interested in.

REAL-WORLD APPLICATIONS

The practical applications of the Poisson distribution extend far beyond historical events and sports analytics. This versatile statistical concept finds relevance in a wide range of modern real-world scenarios, helping us understand and analyse various phenomena. Let’s explore some of its notable applications.

Homes Sold and Business Planning

Imagine you are a local estate agent. Understanding the number of homes you are likely to sell in a given time period is crucial for business planning and forecasting. The Poisson distribution provides a framework for estimating the probability of selling a specific number of homes per day, week, or any other timeframe based on historical data and average sales rates. This information helps in making informed decisions about marketing strategies, staffing, and resource allocation.

Disease Spread and Epidemiology

In the field of epidemiology, the Poisson distribution plays a vital role in understanding the spread of infectious diseases. By analysing historical data and considering the average rate of infection, researchers can utilise the Poisson distribution to estimate the likelihood of disease outbreaks and their progression.

Telecommunications and Network Traffic

The Poisson distribution finds application in the analysis of telecommunications systems and network traffic. By studying the arrival patterns of these events using the Poisson distribution, companies can anticipate network demand, allocate resources effectively, and ensure smooth and reliable communication services.

Quality Control and Manufacturing Processes

The Poisson distribution is also used in quality control, particularly in manufacturing settings. By analysing the number of defective products using the Poisson distribution, manufacturers can estimate the probability of observing a specific number of defects. This information helps them identify areas for improvement and enhance overall product quality.

Traffic Accidents and Road Safety

Another area where the Poisson distribution finds application is in analysing traffic accidents and road safety. By examining historical data on accidents, researchers can use the Poisson distribution to model accident rates based on factors such as location, time of day, and road conditions. This understanding helps in the development of targeted interventions to reduce accidents and improve road safety.

CONCLUSION: A POWERFUL TOOL FOR INFREQUENT EVENTS

The Poisson distribution is a valuable statistical tool that helps us understand and analyse events that happen infrequently but have an average rate of occurrence. It may seem complicated at first, but it allows us to make predictions and informed decisions based on probabilities. By using the principles of the Poisson distribution, we can gain insights into rare events and use that knowledge to improve various aspects of our lives.

When should we roll the dice?

Exploring Games of Chance

A version of this article appears in TWISTED LOGIC: Puzzles, Paradoxes, and Big Questions. By Leighton Vaughan Williams. Chapman & Hall/CRC Press. 2024.

UNDERSTANDING THE CHEVALIER’S DICE PROBLEM

Probability is the science of uncertainty, providing a way to measure the likelihood of events occurring. It can be viewed as a measure of relative frequency or as a degree of belief. In the context of gambling, understanding probability is crucial for making informed decisions and avoiding common pitfalls.

A famous problem, known as the Chevalier’s Dice Problem, sheds light on the some of the intricacies of probability.

To understand the problem, it is essential to grasp some fundamental concepts of probability. Consider a single die roll—each outcome represents a possible event, such as rolling a 1, 2, 3, 4, 5, or 6. When rolling two dice, there are 36 possible outcomes (six outcomes for the first die multiplied by six outcomes for the second die).

THE FLAWED REASONING OF THE CHEVALIER

The Chevalier’s Dice Problem originated from a gambling challenge offered by the Chevalier de Méré, a 17th-century French gambler. The Chevalier offered even money odds that he could roll at least one six in four rolls of a fair die.

The Chevalier’s reasoning was based on the assumption that since the chance of rolling a six in a single die roll is 1/6, the probability of rolling a six in four rolls would be 4/6 or 2/3. However, this reasoning can be shown to lead to inconsistent results when extrapolated to more rolls.

The correct approach involves considering the independent nature of each throw of the die. The probability of a six in one go is 1/6, so the probability of not getting a six on that go is 5/6. To calculate the probability of not rolling a six in four throws, we multiply the probabilities: (5/6) × (5/6) × (5/6) × (5/6) = 625/1296.

Therefore, the probability of at least one six in four attempts is obtained by subtracting the probability of not rolling a six in any of those four attempts from 1: 1 − (625/1,296) = 671/1,296 ≈ 0.5177, which is greater than 0.5.

Despite his faulty reasoning, the Chevalier still had an edge in this game by offering even money odds on an event with a probability of 51.77%.

THE CHEVALIER’S MISSTEP WITH THE MODIFIED GAME

Encouraged by his initial success, the Chevalier expanded the game to 24 rolls of a pair of dice, betting on the occurrence of at least one double-six. His reasoning followed the same flawed pattern: since the chance of rolling a double-six with two dice is 1/36, he believed the probability of at least one double-six in 24 rolls would be 24/36 or 2/3.

The correct probability calculation involved considering the independent nature of each dice roll. The probability of no double-six in one roll is 35/36. Therefore, the probability of no double-six in 24 rolls is (35/36) raised to the power of 24, which is approximately 0.5086.

Subtracting this value from 1 yields the probability of at least one double-six in 24 rolls: 1 − 0.5086 = 0.4914, which is less than 0.5. Hence, the Chevalier’s edge in this modified game was negative: 49.14% − 50.86% = −1.72%.

This outcome demonstrated that even if the odds seem favourable, incorrect reasoning can lead to erroneous conclusions. The Chevalier’s faulty understanding of probability caused him to lose over time.

THE IMPORTANCE OF CORRECT PROBABILITY CALCULATION

These examples underscore the critical nature of accurate probability calculations in games of chance. While intuitive reasoning may seem convincing, it often leads to incorrect conclusions, as demonstrated by the Chevalier’s bets. Understanding the true probability of events is essential for informed decision-making in gambling and many other contexts where risk and uncertainty play significant roles.

THE GAMBLER’S RUIN AND UNDERSTANDING FINITE EDGES

The Gambler’s Ruin problem raises the complementary question of whether, in a gambling game, a player will eventually go bankrupt if playing for an extended period against an opponent with infinite funds, even if the player has an edge.

For instance, imagine a fair game where you and your opponent flip a coin, and the loser pays the winner £1. If you start with £20 and your opponent has £40, the probabilities of you and your opponent ending up with all the money can be calculated using the following formulas:

P1 = n1/(n1 + n2); P2 = n2/(n1 + n2)

Here, n1 represents the initial amount of money for player 1 (you) and n2 represents the initial amount for player 2 (your opponent). In this case, you have a 1/3 chance of winning the £60 (20/60), while your opponent has a 2/3 chance. However, even if you win this game, playing it repeatedly against various opponents or the same one with borrowed money will eventually lead to the loss of your betting bank. This holds true even when the odds are in your favour. This is an important lesson in risk management, emphasising the importance of not only the odds but also the size of one’s bankroll relative to the stake sizes.

The Gambler’s Ruin problem, as explored by Blaise Pascal, Pierre Fermat, and later mathematicians like Jacob Bernoulli, reveals the inherent risks of prolonged gambling, even with favourable odds.

PILOT ERROR: MISUNDERSTANDING CUMULATIVE PROBABILITY

In Len Deighton’s novel ‘Bomber’, a statistical claim suggests that a World War II pilot with a 2% chance of being shot down on each mission is ‘mathematically certain’ to be shot down after 50 missions. This assertion is a classic example of misinterpreting cumulative probability. In reality, if a pilot has a 98% chance of surviving each mission, their probability of not being shot down after 50 missions is 0.98 to the power of 50 (0.9850)which is approximately 0.36, or 36%. Thus, their chance of being shot down over these 50 missions is 64% (1 − 0.36), not 100%.

SURVIVORSHIP BIAS: THE CASE OF BULLET-RIDDEN PLANES

The concept of survivorship bias is vividly illustrated in the case of analysing planes returning from missions during World War II. Upon examining these planes for bullet holes, it was observed that most hits were on the wings, tail, and the body of the plane, with few on the engine. The initial, intuitive response might be to reinforce the areas with the most bullet holes. However, this would be a misinterpretation of the data.

The key realisation, identified by statistician Abraham Wald, was that the planes being analysed were those that survived and returned to base. The areas with fewer bullet holes, such as the engines, were likely critical to survival. Planes hit in these areas probably didn’t make it back, hence the lack of data for these hits. This understanding exemplifies survivorship bias—focusing on survivors (or what’s visible) can lead to incorrect conclusions about the whole population.

Wald’s insight led to the reinforcement of seemingly less-hit areas like engines, contributing significantly to the survival of many pilots. His work in operational research during the war provided a critical perspective on interpreting data and making decisions under uncertainty.

CONCLUSION: DICE, ODDS, AND RUIN

The Chevalier’s Dice Problem illustrates the importance of understanding probability in gambling scenarios. Probability theory, as developed through famed correspondence between Pascal and Fermat, has contributed to modern probability concepts and the understanding of risk involved in gambling.

The Gambler’s Ruin is a kind of warning from the world of probability, telling us that in gambling, a slight edge is no guarantee of success. Imagine two gamblers, one with an edge over the other but with much less money to play with. Even if the first player is more likely to win each round, their thinner wallet means they could run out of money after a few bad games. In contrast, the player with the deep pockets can keep playing longer, until (given enough money) luck swings their way. This underlines the importance and impact of losing streaks in games of chance.

The wartime examples highlight the real-world importance of understanding probability and statistical concepts accurately. They serve as a reminder that intuition can often lead us astray. Correctly interpreting data, especially in high-stakes situations, can have life-saving implications.

When should we stake it all?

The Gambler’s Dilemma

A version of this article appears in TWISTED LOGIC: Puzzles, Paradoxes, and Big Questions. By Leighton Vaughan Williams. Chapman & Hall/CRC Press. 2024.

THE DILEMMA

When the stakes are high and time is not a luxury, finding a solution can be like gambling with fate. This was the scenario for Mike, needing £216 to settle an urgent debt, with only £108 in hand. The roulette wheel beckoned as a potential salvation, but what was the most effective strategy to double his money?

UNDERSTANDING THE ODDS IN ROULETTE

To fully grasp the situation that Mike finds himself in, it’s crucial to examine the mechanics and probabilities of the game he’s chosen as his lifeline: roulette. Specifically, we are considering a single-zero roulette wheel, a version of the game commonly found in European casinos.

Roulette consists of a spinning wheel and a small ball. The wheel is divided into 37 compartments or ‘slots’: numbers from 1 to 36 (randomly assigned as red or black) and a single zero slot. Bets can be placed on a single number, colour, or various combinations thereof.

In a single-zero roulette wheel, the player has a 1 in 37 chance of correctly predicting the outcome. This is because there are 37 slots in total: 36 numbers and the zero. So if you bet on a single number, the odds of the ball landing on that number are 1 in 37, or 36/1. The payout for such a bet, however, is 35/1. This discrepancy between the actual odds (36/1) and the payout odds (35/1) is where the house gains its edge. Every time a player wins, the house pays out less than the actual odds would dictate. In this way, the house earns a profit over time.

The ‘house edge’ is approximately 2.7%, a figure derived from the ratio of the single zero slot to the total number of slots (1/37). This constant advantage in favour of the casino is what makes the game fundamentally a game of negative expectation for players.

To understand the house edge in another way, consider this: if you were to place a £1 bet on each of the 37 slots, totalling £37, your return would be £36 (the £35 returned on the winning number plus the stake of £1). So for every £37 wagered, you would lose £1 using this strategy, which is approximately a 2.7% loss—exactly the house edge.

In conclusion, roulette, like all casino games, is a game of probabilities. And these probabilities, owing to the discrepancy between the actual odds and the payout odds, are slightly skewed in favour of the house. This fundamental understanding of the game’s odds is pivotal when contemplating betting strategies, as we will see with the employment of ‘bold’ and ‘timid’ approaches.

THE BOLD STRATEGY: STAKING IT ALL

Mike’s precarious situation leads him to contemplate a high-risk, high-reward approach known as the ‘bold’ strategy, which involves wagering all his available money at once. In this instance, he considers staking his entire £108 on the colour Red, a bet with almost a 50-50 chance, as the roulette wheel has 18 red slots out of 37 total slots.

To fully appreciate the audaciousness of this approach, it’s essential to understand the mathematics behind it. When betting on a colour, there’s a near-even split of potential outcomes: 18 red slots, 18 black slots, and the zero slot. Thus, the likelihood of the ball landing on a red slot is 18 out of 37, or roughly 48.6%. Consequently, with this single bet, he has about a 48.6% chance of doubling his money and obtaining the £216 he urgently needs.

However, it’s important to note that this is a single-round probability. Unlike a ‘timid’ strategy, where multiple rounds are played, the bold strategy is a one-off scenario. Therefore, the 48.6% chance of winning must be interpreted as his overall chance of achieving his target sum. There are no second chances or opportunities to recoup losses; it’s an all-or-nothing situation.

By putting all his money on one bet, he is maximising his return if that bet is successful. This is in contrast to a timid strategy, where the payout would be spread over multiple smaller bets, with the likelihood of achieving the target sum being significantly less.

But the bold strategy also comes with the highest level of risk. If the ball doesn’t land on Red, Mike loses everything. His entire available funds are at stake, making the potential loss just as significant as the potential gain.

In conclusion, the bold strategy is a high-stakes, high-reward approach. It encapsulates the old saying, ‘Who dares, wins’, and, in this case, provides him the best chance of reaching his £216 target. Why is this so?

TIMID APPROACH: MULTIPLE SMALL BETS

As opposed to the bold strategy, he could consider dividing his available £108 into 18 separate bets of £6 each. These small, successive bets would be placed on a single number until he either depletes his funds or hits the winning number, which would yield a payout of 35 to 1, giving him the £216 he needs.

To fully understand the implications of this strategy, we need to analyse it in detail. The probability of winning a single number bet in roulette is 1 in 37, as there are 36 numbers and one zero. Hence, for each individual bet, John has a 1 in 37 chance of winning, or approximately 2.7%.

However, the timid strategy involves making multiple small bets, and so we must calculate the probability of these successive bets all losing. Since each individual bet has a 36 in 37 chance of losing, the probability that all 18 bets lose would be calculated as (36/37) to the power of 18, which equates to around 0.61, or 61%.

As such, the probability of him winning at least once using this timid strategy would be equal to 1 minus the losing probability. Hence, the chance of hitting the target £216 is 1 − 0.61, or 39%.

Interestingly, the timid strategy, although appearing less risky, significantly reduces Mike’s chances of achieving his target sum compared to the bold approach. By spreading out his available funds across multiple bets, he lowers his exposure to loss in each individual game, but also decreases the likelihood of achieving his overall goal.

This strategy extends the length of play and the suspense, providing more instances of potential winning and losing. However, each bet also exposes Mike to the house edge, and therefore the risk of losses incrementally increases.

In this way, the timid approach offers more sustained engagement with the game but sacrifices the higher winning potential found in the bold approach.

THE POWER OF BOLD PLAY: TAKING A CALCULATED RISK

To look at it another way, consider a scenario where equal amounts are bet on red and black in each round. In most cases, the outcome will lead to breaking even, specifically 36 out of 37 times. However, when the ball lands on the single zero slot, the entire bank is lost. The more games played, the greater the chance of this happening.

By limiting the game to a single spin, the bold strategy minimises the number of times the house edge comes into play. Hence, playing fewer rounds decreases the likelihood of the house edge depleting the funds before reaching the target.

This strategy is not just about boldness in the face of risk, but more about understanding and working around the inherent disadvantage players face in casino games. By playing fewer games, you reduce the opportunities for the house edge to work against you.

CONCLUSION: THE INTUITION BEHIND BOLD PLAY

The intuition behind bold play in unfavourable games is grounded in a nuanced understanding of the mechanics of casino games and their built-in house edge. Bold play aims at striking hard and fast, capitalising on the relatively high chance of achieving the target sum in a single round, instead of facing the progressively increasing exposure associated with multiple rounds. In this sense, it’s a calculated and strategic form of boldness.

When should we raise the stakes?

The Kelly Criterion

A version of this article appears in TWISTED LOGIC: Puzzles, Paradoxes, and Big Questions. By Leighton Vaughan Williams. Chapman & Hall/CRC Press. 2024.

TAKING ADVANTAGE OF THE ODDS

One of the most critical aspects of any betting strategy is determining the size of the bet when we believe the odds are in our favour. The answer to this pressing question was formalised by John L. Kelly, Jr., an engineer at Bell Labs by profession, as well as a hobbyist daredevil pilot and recreational gunslinger. His methodology, known as the Kelly Criterion, is a mathematical formula designed to establish the optimal bet size when we have an advantage, that is, when the odds favour us.

However, having the advantage doesn’t guarantee a successful outcome. Irrespective of our edge, excessive betting can lead to large losses and, in worst-case scenarios, to bankruptcy. This is where the Kelly Criterion comes into play. It takes account of both the size of our edge and the potentially damaging impact of volatility.

UNDERSTANDING THE KELLY CRITERION

The Kelly Criterion is essentially a mathematical formula that calculates the optimal amount to bet or invest when the odds are in our favour. The fundamental principle underpinning the Kelly Criterion is that the amount of capital wagered should be related to our advantage at the available odds. It emphasises the relationship between the size of our bet and our perceived edge.

Consider this simple illustration: Suppose you have a coin where the probability of getting a head (winning) is expected to be equal to the probability of a tail (losing). Now, suppose you have secret information that the next coin toss will certainly be heads. In this case, you have a 100% edge. According to the strict Kelly Criterion, you should bet your entire capital because you’re guaranteed to win. In real-life applications, even with very high confidence, betting one’s entire capital is risky, however, due to the possibility of unforeseen factors. This is more a thought experiment than a practical recommendation.

In any case, in most scenarios the outcomes are not binary, and the probability of winning is rarely 100%, even in theory.

Let’s consider a different situation: You’re still betting on the coin toss, but this time your secret information gives you a 60% chance of landing heads and a 40% chance of tails. Your edge is now 20%, and a very basic Kelly strategy is to stake 20% of your capital.

This example reflects the core concept of the Kelly Criterion. It’s not only about gauging when you have the advantage—it’s also important to understand precisely how much to stake when you do. In theory, this sounds simple, but in practice, accurately identifying that advantage can be complex.

On a broader scale, the Kelly Criterion can be employed in various fields beyond betting, such as investing and trading, to determine the optimal size of a series of bets or investments. Its aim is to maximise the exponential growth of the bettor’s or investor’s wealth over the long term.

A strength of the Kelly Criterion is its flexibility. It allows you to adjust the proportion of the capital that you bet based on how strong your edge is.

It’s important to note, however, that the Kelly Criterion assumes that the bettor can reinvest their winnings. This is crucial for the ‘compounding’ aspect of the strategy, which is what allows the wealth to grow faster than it would with other systems. This compound growth strategy is what differentiates Kelly betting from more static strategies, but also introduces higher volatility in the short term.

Ultimately, the Kelly Criterion offers a robust methodology for managing risk and maximising returns when the odds are in our favour. However, as with any strategy, understanding the core principles is just the beginning—it’s the accurate identification of the edge and the consistent application of the strategy that’s critical for long-term success. Misestimations can lead to over-betting and significant losses.

APPLYING THE KELLY CRITERION

The application of the Kelly Criterion can have profound implications for various fields beyond gambling, such as investing and trading. The crucial component is to understand that the Kelly Criterion isn’t just about betting when we have an edge; it’s about calculating the precise amount to bet to maximise compounded return over time.

This is where the Kelly formula can come into play:

F = Pw − (Pl/W)

where

F is the Kelly criterion fraction of capital to bet,

W is the amount won per amount wagered (i.e. win size, net of the stake, divided by loss size),

Pw is the probability of winning, and

Pl is the probability of losing.

When we apply this formula, we calculate the optimal fraction of our capital to bet, given our probability of winning (Pw), our probability of losing (Pl), and our potential return (W).

Consider a simple example: Suppose we have an even-money bet, i.e. the amount you stand to win, net of the stake, is the same as the amount you risk. In this scenario, the value of W is 1. If our chance of winning is 60% and our chance of losing is 40%, substituting these values into the simplified Kelly formula (F = Pw − Pl) gives us F = 0.60 − 0.40 = 0.20 or 20%. This means that in order to maximise our long-term return, we should bet 20% of our capital.

Let’s consider a slightly more complicated scenario: Suppose we have a bet where we stand to win double the amount we risk, i.e. W = 2, and the probability of winning and losing is both 50%. Substituting these values into the original Kelly formula gives us F = 0.50 − (0.50/2) = 0.50 − 0.25 = 0.25 or 25%. This means we should bet 25% of our working capital to maximise our long-term return.

The Kelly Criterion is designed to ensure that you never go bankrupt because the recommended bet size decreases as your capital decreases. However, this doesn’t mean you can’t lose money. The Kelly Criterion maximises long-term growth rather than short-term returns. This means that there will be times when you lose money, but over the long run, you should come out ahead.

It’s crucial to remember that the Kelly Criterion assumes you know the true probabilities of the outcomes, which is often not the case. In practice, we’re often working with estimated probabilities, which means there’s a risk that we could, for example, overestimate our edge and bet too much. Therefore, many investors and bettors use a fraction of the Kelly Criterion (betting a fixed fraction of the amount recommended by Kelly) to reduce their risk.

Lastly, while the Kelly Criterion offers a mathematical approach to betting and investing, it doesn’t account for the emotional aspect of risking money. Remember that the goal is not just to maximise returns, but also to sleep well at night.

POTENTIAL RISKS AND LIMITATIONS

While the Kelly Criterion can be an effective strategy for maximising the growth of capital in the long run, it is not without its potential risks and limitations. These should be understood and considered before applying the formula.

Estimation Errors

The effectiveness of the Kelly Criterion hinges on the accuracy of the probabilities used in the calculation. An overestimation of the probability of winning (Pw) can lead to excessive bet sizes and the risks associated with over-betting.

Minimum Bet Size

The Kelly Criterion presupposes that there is no minimum bet size, which is rarely the case in real-world scenarios, especially in investing and trading. In situations where a minimum bet size exists, the possibility of losing all of the capital becomes a reality if the amount falls below this threshold.

Risk Tolerance

The Kelly Criterion determines bet sizes purely based on mathematical calculations to maximise long-term growth. It does not take into account the individual bettor’s or investor’s risk tolerance. An aggressive bet size recommended by the Kelly Criterion may not be psychologically comfortable for some, causing stress and potentially leading to sub-optimal decisions.

Given these potential risks and limitations, it is common for many investors and bettors to use a fractional Kelly strategy, betting a fraction (like half or a third) of the amount recommended by the Kelly formula. This approach can help mitigate the risks associated with over-betting and inaccuracies in probability estimation while still providing the benefits of proportional betting and capital growth. However, even a fractional Kelly strategy should be tailored to individual circumstances, including risk tolerance and the ability to withstand potential losses.

CONCLUSION: TAKING ADVANTAGE OF OUR EDGE

The Kelly Criterion, devised by John L. Kelly, Jr., is a unique betting strategy that uses probability and potential payout to determine the optimal bet size when the odds are in our favour. The mathematical formula suggests betting a fraction of capital equivalent to the size of one’s advantage. However, it’s crucial to account for potential errors and uncertainties that can affect the real-world implementation of this strategy.

Uncertainty in the size of any actual edge at the odds and the potential for a bumpy ride due to volatility mean that we should always exercise caution. As a result, unless we’re prepared for potential high volatility and have unwavering confidence in our judgment, adopting a fractional Kelly strategy might be the most prudent approach. This strategy allows us to stake a defined fraction of the recommended Kelly amount, reducing risk while still taking advantage of our edge.

Does seeing a white tennis shoe make it more likely that all flamingos are pink?

Introducing the Accessibility Principle

UNDERSTANDING HEMPEL’S PARADOX

In the mid-20th century, philosopher Carl Gustav Hempel introduced a paradox that came to be known as ‘Hempel’s Paradox’ or the ‘Raven Paradox’. The paradox begins with a seemingly simple and clear premise: If the hypothesis is that ‘all ravens are black’, then any observation of a black raven should help to support the hypothesis.

However, Hempel pointed out that this statement is logically equivalent to the statement: ‘Everything that is not black is not a raven’. Hence, any observation of a non-black, non-raven object, such as a white tennis shoe, should also help to support the hypothesis.

Yet it feels strange to believe that seeing a white tennis shoe should serve to increase our belief that all ravens are black.

HEMPEL’S PARADOX AND THE COLOUR OF FLAMINGOS

Now, let’s apply this principle to another statement: ‘All flamingos are pink’. This proposition is logically equivalent to: ‘Everything that is not pink is not a flamingo’. By Hempel’s argument, observing an object that is not pink and not a flamingo, such as a white tennis shoe, would provide evidence in support of the hypothesis that all flamingos are pink.

From a formal logic perspective, this argument makes sense. However, our intuition may still find this hard to accept, mirroring the original conflict inherent in Hempel’s Paradox.

TESTING THE HYPOTHESIS

In conventional hypothesis testing, we would go out and find some flamingos, verifying if they are indeed pink. But the Raven Paradox suggests that we could conduct meaningful research by simply looking at random non-pink things and checking if they are flamingos. As we collect data, we increasingly lend support to the hypothesis that all non-pink things are non-flamingos, equivalently that all flamingos are pink.

While this approach holds up logically, it does have its limitations. Considering the vast number of non-pink things in the world compared to the population of flamingos, the hypothesis can be much more confidently validated by sampling flamingos directly. Hence, although Hempel’s Paradox does not contain a logical flaw, it is not an efficient or practical method for testing the hypothesis.

THE ACCESSIBILITY PRINCIPLE (OR OBSERVATIONAL LIKELIHOOD PRINCIPLE)

Suppose we have two hypothetical species—one is a type of bird that frequents populated areas (Species A), and the other is a rare kind of lizard that lives in remote, inaccessible areas (Species B). If both these species are unobserved, it’s more likely that Species B exists rather than Species A, because Species B is less likely to be observed due to its habitat even if it exists. In contrast, Species A should have been observed if it were real due to its frequent presence in populated areas. I term this the ‘Accessibility Principle’, or alternatively the ‘Observational Likelihood Principle’. These terms suggest that the likelihood of an entity’s existence depends on its observability. This aligns with real-world scientific practices, where the absence of evidence is not always evidence of absence, particularly when dealing with hard-to-observe phenomena.

So, let’s take the propositions in the thought experiment in turn. Proposition 1: All flamingos are pink. Proposition 2 (logically equivalent to Proposition 1): Everything that is not pink is not a flamingo. Proposition 3 (the ‘Accessibility Principle’): If something might or might not exist but is difficult to observe, it is more likely to exist than something which can be easily observed but is not observed.

Following from these propositions, when I see two white tennis shoes, I am ever so slightly more confident that all flamingos are pink than before. This is especially so if any non-pink flamingos that might be out there would be easy to spot. And I’d still be wrong, but for all the right reasons.

CONCLUSION: THE OBSERVATION PARADOX

In summary, Hempel’s Paradox is an intriguing clash between intuitive reasoning and formal logic. It forces us to confront the subtleties of hypothesis testing and belief formation. In this example, the paradox implies that we may gain a tiny bit more confidence in the hypothesis that all flamingos are pink if we observe a white tennis shoe. However, such indirect evidence should be considered in its appropriate context, not as a substitute for direct evidence. The key point of the paradox is instead to challenge our understanding of the meaning of evidence and to provide valuable insights into the nature of logical reasoning. Essentially, any hypothesis is always susceptible to new evidence that can strengthen support for it. In the case of the pink flamingo hypothesis, this applies whether it comes from observing a flock of pink flamingos or (to a much lesser degree) a pair of white tennis shoes. Until you see an orange flamingo, then you know otherwise!

When should we accept the simplest explanation?

Exploring Occam’s Razor

THE PRINCIPLE OF SIMPLICITY

In this section we explore Occam’s Razor. William of Occam (also spelled William of Ockham) was a prominent 14th-century philosopher and theologian known for his emphasis on simplicity in philosophical and theological matters. His philosophical contributions, particularly the principle of simplicity, have had a lasting impact on various fields of knowledge. Occam’s Razor, derived from his philosophy, has become synonymous with the method of eliminating unnecessary hypotheses and choosing the simplest explanation consistent with the evidence.

OCCAM’S RAZOR: PRINCIPLE AND EXPLANATION

At the heart of Occam’s philosophy, therefore, is the principle of simplicity, which later became known as Occam’s Razor. The razor can be summarised as follows: ‘Entities should not be multiplied without necessity’. In other words, when faced with competing explanations or hypotheses, the simplest one that adequately explains the available evidence should be preferred.

Occam’s Razor guides our thinking by encouraging us to avoid unnecessary assumptions and complexities. It suggests that we should prefer explanations that require fewer additional elements or entities. By choosing simplicity over complexity, Occam’s Razor helps us navigate knowledge acquisition and hypothesis formation.

To be clear, it’s important to note that simplicity does not mean ‘easier to understand’ but rather ‘involving fewer assumptions or conjectures’. Complexity should only be considered when simplicity fails to adequately explain the phenomenon.

OCCAM’S RAZOR: A CRUCIAL HEURISTIC

Occam’s Razor serves as a crucial heuristic in problem-solving and theory formulation. It proposes that among competing hypotheses, the one with the fewest assumptions should be selected, provided it adequately explains the phenomenon in question.

Occam’s Razor does not just simplify our thinking processes; it actively steers us away from the allure of unnecessary complexities and conjectures. By advocating for simplicity, it aids in refining our approach to knowledge acquisition and hypothesis development, ensuring that complexity is introduced only when absolutely necessary to explain the data adequately.

THE ROLE OF OCCAM’S RAZOR IN SCIENCE: TOWARDS ELEGANT EXPLANATIONS

The implications of Occam’s Razor extend significantly into scientific inquiry. It underpins the scientific method, where explanations for observed phenomena are sought and hypotheses are developed. By favouring parsimonious explanations, the principle encourages scientists to construct theories that are not only elegant but also more comprehensible and testable. This preference for simplicity has facilitated remarkable advancements in our understanding of the world, emphasising that the most profound explanations often emerge from the most straightforward assumptions.

OCCAM’S RAZOR AND OVERFITTING: COMPLEXITY AND GENERALISATION

Occam’s Razor finds empirical support in the phenomenon of overfitting, particularly in the field of statistics and machine learning. Overfitting occurs when a model becomes overly complex and fits the noise or random variations in the data instead of capturing the true underlying patterns.

By adhering to Occam’s Razor, researchers can avoid the pitfall of overfitting, ensuring that their models capture the essential features of the data while remaining parsimonious.

OCCAM’S LEPRECHAUN: AVOIDING AD HOC HYPOTHESES

In the pursuit of explanations, it is common to encounter situations where additional assumptions are introduced to save a theory from being falsified. These ad hoc hypotheses act as patches to compensate for anomalies that were not anticipated by the original theory. Occam’s Razor plays an essential role in evaluating such situations.

Imagine a situation, for example, where someone claims that a mischievous leprechaun is responsible for breaking a vase. There is likely to be serious scepticism about this claim. However, the person who made the claim introduces a series of ad hoc explanations to counter potential falsification.

For instance, when a visitor to the scene sees no leprechaun, the claimant asserts that the leprechaun is invisible. To test this, the visitor suggests spreading flour on the ground to detect footprints. In response, the claimant states that the leprechaun can float, thus leaving no footprints. The visitor then proposes asking the leprechaun to speak, but the claimant asserts that the leprechaun has no voice. In this way, the claimant keeps introducing additional explanations to prevent the hypothesis of the leprechaun’s existence from being falsified.

The example of Occam’s Leprechaun illustrates how additional assumptions can be added in an ad hoc manner to preserve a theory from being disproven. These ‘saving hypotheses’ create a flow of additional explanations that make the theory less able to be falsified. Occam’s Razor encourages us to be sceptical of such ad hoc hypotheses and instead favours simpler explanations that adequately account for the evidence.

OCCAM’S RAZOR AND PREDICTIVE POWER: PARSIMONY AND EFFICIENCY

Another aspect of Occam’s Razor is its association with the predictive power of theories. A theory that can accurately predict future events or observations based on fewer assumptions is considered more efficient.

By favouring simplicity, Occam’s Razor guides scientists to develop theories that provide explanatory power and predictive efficiency. The simplest theory that adequately explains the available data and accurately predicts future outcomes is often preferred.

This emphasis on predictive power aligns with the pragmatic approach that Occam advocated, where theories are judged not only by their ability to explain past observations but also by their ability to make successful predictions.

CONCLUSION: THE PURSUIT OF KNOWLEDGE

In sum, Occam’s Razor, a principle deeply rooted in the philosophy of William of Occam, remains a fundamental tool in the pursuit of knowledge. It encourages simplicity, efficiency, and parsimony in our explanations and theories. By guiding us towards the simplest explanations that remain consistent with the available evidence, Occam’s Razor plays a crucial role in scientific methodology, theory development, and everyday reasoning. Its continued relevance underscores the timeless appeal of simplicity in our quest to understand and explain the world around us.