Rigorous Reasoning

Decision And Rational Choice

Expected Value and Expected Utility

Teaches the mechanics of computing expected value across outcomes, introduces utility functions that capture risk aversion through diminishing marginal utility, and walks through the St Petersburg paradox as motivation for why utility must sometimes replace money in the calculation.

Read for structure, not just vocabulary. The goal is to learn how natural-language claims are converted into a cleaner formal shape.

IntegratedFormalizationLesson 2 of 50% progress

Start Here

What this lesson is helping you do

Teaches the mechanics of computing expected value across outcomes, introduces utility functions that capture risk aversion through diminishing marginal utility, and walks through the St Petersburg paradox as motivation for why utility must sometimes replace money in the calculation. The practice in this lesson depends on understanding Expected Value, Utility, Preference Ordering, and Diminishing Marginal Utility and applying tools such as Maximize Expected Utility and Transitivity of Preferences correctly.

How to approach it

Read for structure, not just vocabulary. The goal is to learn how natural-language claims are converted into a cleaner formal shape.

What the practice is building

You will put the explanation to work through formalization practice, quiz, analysis practice, classification practice, guided problem solving, rapid identification, evaluation practice, argument building, and diagnosis practice activities, so the goal is not just to recognize the idea but to use it under your own control.

What success should let you do

Correctly compute expected value and expected utility for 6 decision problems and explain when and why the two calculations produce different recommendations.

Reading Path

Move through the lesson in this order

The page is designed to teach before it tests. Use this sequence to keep the reading, examples, and practice in the right relationship.

Read

Build the mental model

Move through the guided explanation first so the central distinction and purpose are clear before you evaluate your own work.

Study

Watch the move in context

Use the worked examples to see how the reasoning behaves when someone else performs it carefully.

Do

Practice with a standard

Only then move into the activities, using the pause-and-check prompts as a final checkpoint before you submit.

Guided Explanation

Read this before you try the activity

These sections give the learner a usable mental model first, so the practice feels like application rather than guesswork.

Core formula

Expected value is a probability-weighted average

Expected value is the workhorse of decision theory under risk. For an option with possible outcomes x1, x2, up to xn and corresponding probabilities p1, p2, up to pn, the expected value is the sum of each probability times its payoff: EV = p1 times x1 plus p2 times x2 plus ... plus pn times xn. The probabilities must sum to 1, which is a reminder that you have not forgotten any possible outcome. The expected value is a weighted average, not a specific outcome: it is the value you would receive on average if you could replay the decision many independent times.

This averaging interpretation is important. Expected value does not predict what will happen on any single trial. A lottery ticket with a 1-in-10-million chance of winning 5 million dollars has an expected value of 50 cents, but you will almost certainly get nothing rather than 50 cents on any single purchase. The expected value tells you the long-run average, and that is the right quantity when you can afford to play enough times for the averaging to work, or when you are committed in advance to a policy of repeated similar choices.

What to look for

  • Write down every possible outcome, including the ones that seem trivially small or trivially large.
  • Assign probabilities that sum to exactly 1.
  • Multiply each probability by its payoff and sum the products.
  • Interpret the result as a long-run average, not as a forecast of what will happen next.
Expected value is a probability-weighted average of outcomes, not a prediction of any particular trial.

Concrete demonstration

Walking through an expected value calculation

Consider a simple gamble: pay 10 dollars to roll a fair six-sided die, and win 30 dollars if the result is a 6 and nothing otherwise. The expected value of the gamble's payoff is (1/6) times 30 plus (5/6) times 0 = 5 dollars. After subtracting the 10-dollar cost of playing, the net expected value is negative 5 dollars per roll. A rational risk-neutral agent would decline this bet because the expected value is negative.

Now consider a better gamble: pay 10 dollars to roll the same die, and win 60 dollars on a 6 instead of 30. The expected payoff is now (1/6) times 60 = 10 dollars, exactly equal to the cost. The net expected value is 0, a fair bet. If we raised the winning amount to 90 dollars on a 6, the expected payoff would be 15 dollars, the net expected value would be 5 dollars per roll, and a rational risk-neutral agent would take the bet — provided they could afford to play enough rolls to let the averaging work and were not ruined by the variance in the meantime.

What to look for

  • Compute gross expected value from probabilities and payoffs.
  • Subtract any certain costs (the cost of playing) to get net expected value.
  • Compare the net expected value against zero and against the expected value of alternative uses of the same resources.
Expected value is easy to compute once you have listed outcomes, probabilities, and payoffs — the discipline is in listing them honestly.

Motivation for utility

The St Petersburg paradox and the birth of utility

In 1713 Nicholas Bernoulli proposed a game: a fair coin is flipped repeatedly until it lands heads. If heads occurs on the first flip, you win 2 dollars. If on the second, 4 dollars. If on the third, 8 dollars. In general, if heads first occurs on flip number n, you win 2 to the n dollars. The expected value of the game is (1/2)(2) + (1/4)(4) + (1/8)(8) + ... = 1 + 1 + 1 + ... which sums to infinity. The expected value calculation therefore says you should pay any finite amount — a million dollars, a billion dollars — to play this game once. But almost no one is willing to pay more than a modest amount, and that intuition is not irrational.

Daniel Bernoulli, Nicholas's cousin, resolved the paradox in 1738 by arguing that what matters is not the expected value of the money but the expected utility of the outcomes. If utility grows more slowly than wealth — for example, as the logarithm or square root of wealth — then the utility contributions of very large payoffs are much smaller than the money amounts suggest. A gain of 2 to the 30 dollars (about a billion dollars) is worth much less than thirty times as much as a gain of 2 dollars, because the marginal utility of wealth diminishes as wealth grows. Under a logarithmic utility function, the expected utility of the St Petersburg game is finite and modest, matching the intuition that no reasonable person should pay a fortune to play.

What to look for

  • Recognize that expected value over money can produce absurd recommendations when payoffs are very large relative to the agent's situation.
  • Switch to expected utility when money is not a good proxy for what the agent cares about.
  • Understand diminishing marginal utility as the reason why large gains matter less in utility than they do in dollars.
The St Petersburg paradox is why rational decision theory uses expected utility rather than expected money — utility captures diminishing marginal value in a way that money does not.

Linking risk preference to utility

Risk aversion is encoded in the shape of the utility function

A utility function that is concave — curving upward but flattening as wealth grows — automatically produces risk-averse behavior. Consider a choice between a sure 50 dollars and a 50/50 gamble between 0 and 100 dollars. Both have the same expected money value (50 dollars), so a purely money-maximizing agent would be indifferent. But under a concave utility function like U(x) = square root of x, the sure option has utility square root of 50, approximately 7.07, while the gamble has expected utility 0.5 times square root of 0 plus 0.5 times square root of 100 = 0.5 times 0 plus 0.5 times 10 = 5. The sure thing has higher expected utility. A square-root-utility agent prefers the sure 50 dollars over the gamble, even though the money expected values are equal.

This is the mathematical content of risk aversion: a concave utility function means the utility loss from a downside outcome is larger than the utility gain from an equal-sized upside outcome, so the agent prefers certainty to variance when expected money is the same. The more concave the utility function, the more risk-averse the agent. A linear utility function corresponds to risk neutrality (the agent cares only about expected money), and a convex utility function corresponds to risk seeking (the agent prefers variance and would pay to gamble even at a fair expected value). Most ordinary human preferences are risk-averse for substantial sums and roughly risk-neutral for small ones.

What to look for

  • Remember that concave utility produces risk aversion, linear utility produces risk neutrality, and convex utility produces risk seeking.
  • Use square root or logarithm as default concave utility functions when you need a working example.
  • Expect your recommendations under expected utility to differ from those under expected money whenever the stakes are large enough that diminishing marginal utility matters.
Risk aversion is not a psychological quirk — it is the mathematical consequence of a concave utility function, which is itself the mathematical expression of diminishing marginal utility.

Core Ideas

The main concepts to keep in view

Use these as anchors while you read the example and draft your response. If the concepts blur together, the practice usually blurs too.

Expected Value

The probability-weighted average of an action's possible outcomes, computed as EV = sum over outcomes of probability times payoff.

Why it matters: Expected value is the starting point for evaluating any choice whose outcome depends on chance, and it generalizes straightforwardly to expected utility.

Utility

A numerical measure of how much an outcome is worth to a particular agent, calibrated so that higher numbers always correspond to more preferred outcomes.

Why it matters: Utility converts money, time, health, and other goods into a single scale that captures what actually matters to the decision maker, including attitudes toward risk.

Preference Ordering

A ranking of alternatives that a decision maker holds, ideally satisfying completeness (every pair is comparable) and transitivity (if A is preferred to B and B to C, then A is preferred to C).

Why it matters: A coherent preference ordering is the foundation of rational choice; inconsistent preferences make any decision tool unreliable and expose the agent to money pumps.

Diminishing Marginal Utility

The property that successive units of a good produce smaller and smaller increases in utility; the tenth dollar gained matters less to you than the first.

Why it matters: Diminishing marginal utility is what makes concave utility functions and risk aversion rational rather than timid.

Reference

Open these only when you need the extra structure

How the lesson is meant to unfold

Concept Intro

The core idea is defined and separated from nearby confusions.

Formalization Demo

The lesson shows how the same reasoning looks once its structure is made explicit.

Worked Example

A complete example demonstrates what correct reasoning looks like in context.

Guided Practice

You apply the idea with scaffolding still visible.

Independent Practice

You work more freely, with less support, to prove the idea is sticking.

Assessment Advice

Use these prompts to judge whether your reasoning meets the standard.

Mastery Check

The final target tells you what successful understanding should enable you to do.

Reasoning tools and formal patterns

Rules and standards

These are the criteria the unit uses to judge whether your reasoning is actually sound.

Maximize Expected Utility

When probabilities are known and preferences are represented by a utility function, a rational agent should choose the option with the highest expected utility.

Common failures

  • Choosing the option with the highest possible payoff without weighing how likely it is.
  • Substituting the most likely outcome for the expected value and ignoring the remaining possibilities.
  • Treating expected utility as a guarantee of the preferred outcome rather than as a long-run average over repeated choices.

Transitivity of Preferences

If an agent prefers A to B and B to C, then the agent should prefer A to C; cycles of preference are irrational and expose the agent to exploitation.

Common failures

  • Preferring A to B in one framing and B to A in another because the choice context changed the salience of attributes.
  • Holding cyclical preferences (A over B, B over C, C over A) that can be pumped for arbitrary losses.

Ignore Sunk Costs

A rational decision is determined by the future consequences of available options; past investments that cannot be recovered should play no role in the choice.

Common failures

  • Continuing a failing project because of the money and time already spent on it.
  • Refusing to abandon a plan that clearly will not succeed because doing so would 'waste' prior effort.
  • Letting the size of a past commitment substitute for an analysis of current expected value.

Dominance Principle

A rational agent should never choose an option that is weakly dominated, and should always prefer an option that strictly dominates its alternatives.

Common failures

  • Selecting a dominated option because it is familiar, vivid, or emotionally salient.
  • Missing a dominance relation because the decision matrix was not laid out explicitly.
  • Treating dominance as a tiebreaker rather than as the most powerful decision rule available.

Independence Axiom

If an agent prefers A to B, then the agent should prefer any mixture (A with probability p, some outcome X with probability 1-p) to (B with probability p, X with probability 1-p); the presence of a common outcome should not flip the preference.

Common failures

  • Allen Allais-style reversals where the same underlying comparison flips depending on whether a sure thing is framed into the choice.
  • Letting certainty (a sure outcome) dominate analysis in a way that contradicts the agent's ordering over the non-sure parts.

Respect Base Rates in Decision Analysis

When decisions depend on probabilities, those probabilities must reflect background base rates and not just vivid or recent information; decision analysis inherits the base-rate discipline of Bayesian inference.

Common failures

  • Inflating the probability of a dramatic outcome because it is easy to imagine (availability bias).
  • Using a recent anecdote as if it were a reliable estimate of the underlying frequency.
  • Ignoring the actual prevalence of failures when evaluating an optimistic business forecast.

Weigh Opportunity Cost Explicitly

An option is only as good as the best alternative it displaces; a good choice must be compared against its next-best alternative, not evaluated in isolation.

Common failures

  • Accepting an option because it looks attractive on its own without asking what is being given up.
  • Treating a small positive expected value as a clear win without asking whether a better option was available for the same resources.

Patterns

Use these when you need to turn a messy passage into a cleaner logical structure before evaluating it.

Decision Matrix

Input form

practical_choice_with_uncertainty

Output form

options_by_states_table_with_payoffs

Steps

  • List the available options as rows.
  • List the possible states of the world as columns.
  • Fill in the payoff or utility for each option-state pair.
  • If probabilities are known, add a row for state probabilities.
  • Check for dominance relations first.
  • Compute expected utility for each non-dominated row.
  • Identify the option with the highest expected utility as the recommended choice, noting any assumptions made about probabilities or utility.

Watch for

  • Omitting a state of the world and thereby biasing the calculation.
  • Filling in outcomes by intuition without actually asking what would happen in each state.
  • Computing expected value across dominated options and missing that the dominance check could have eliminated them immediately.
  • Treating the chosen option as guaranteed to produce the payoff that went into the expected-value calculation.

Expected Value Calculation

Input form

option_with_probabilistic_outcomes

Output form

numerical_expected_value

Steps

  • List every possible outcome that results from the option.
  • Assign a probability to each outcome, ensuring the probabilities sum to 1.
  • Assign a payoff (in dollars, utility units, or another common scale) to each outcome.
  • Multiply each probability by its payoff.
  • Sum the products to obtain the expected value.
  • Compare the expected value against the expected values of alternative options and against any relevant reference point (the current status quo, a safe alternative).

Watch for

  • Using probabilities that do not sum to 1 because one outcome was forgotten.
  • Mixing dollar payoffs with utility values in the same calculation.
  • Reading the computed expected value as a likely outcome rather than as a long-run average.
  • Ignoring variance and tail risk when the stakes are high enough that a bad outcome would be unrecoverable.

Utility Function Application

Input form

monetary_gamble_or_prospect

Output form

expected_utility_score

Steps

  • Identify the decision maker's wealth or baseline reference level.
  • Transform each dollar payoff into a utility value using a concave function such as the square root or logarithm when risk aversion is appropriate.
  • Multiply each utility value by the probability of the corresponding outcome.
  • Sum the products to obtain expected utility.
  • Compare expected utility across options, remembering that the utility scale is meaningful only up to positive linear transformations.

Watch for

  • Using a linear utility function and then wondering why the analysis recommends obviously reckless gambles.
  • Switching utility functions between options in the same comparison.
  • Confusing utility units with dollars when reporting the conclusion.

Worked Through

Examples that model the standard before you try it

Do not skim these. A worked example earns its place when you can point to the exact move it is modeling and the mistake it is trying to prevent.

Worked Example

A Simple Expected Value Comparison

A risk-neutral agent always picks the option with the highest expected value. But this does not automatically generalize to every agent — a risk-averse agent might still prefer Option A, as the next example shows.

Content

  • Option A: Receive 500 dollars with certainty.
  • Option B: 60 percent chance of 1,000 dollars and 40 percent chance of 0 dollars.
  • Expected value of A: 1.0 times 500 = 500 dollars.
  • Expected value of B: 0.6 times 1,000 plus 0.4 times 0 = 600 dollars.
  • Risk-neutral recommendation: Take Option B, which has the higher expected value by 100 dollars.

Worked Example

The Same Choice Under Concave Utility

The same choice looks different depending on the utility function. A square-root-utility agent turns the 100-dollar expected-value advantage of Option B into a utility disadvantage, because the extra upside of 1,000 dollars versus 500 dollars is worth less in utility than the downside of losing the guaranteed 500 dollars.

Content

  • Option A: Receive 500 dollars with certainty, ending wealth 500 dollars.
  • Option B: 60 percent chance of 1,000 dollars (wealth 1,000) and 40 percent chance of 0 dollars (wealth 0).
  • Using U(x) = square root of x with starting wealth 0:
  • Utility of A = square root of 500, approximately 22.36.
  • Expected utility of B = 0.6 times square root of 1,000 plus 0.4 times square root of 0 = 0.6 times 31.62 + 0.4 times 0 = 18.97.
  • Risk-averse recommendation: Take Option A, which has higher expected utility (22.36 > 18.97) even though Option B has higher expected money value.

Worked Example

St Petersburg under Logarithmic Utility

Expected utility gives a finite, sensible answer to the St Petersburg paradox. The paradox arises only because expected value treats each unit of money as equally important, while a diminishing-marginal-utility agent correctly treats very large sums as much less special than their raw dollar amounts suggest.

Content

  • Game: Flip a fair coin repeatedly until heads appears. If heads first appears on flip n, win 2 to the n dollars.
  • Expected monetary value: (1/2)(2) + (1/4)(4) + (1/8)(8) + ... = 1 + 1 + 1 + ... (infinite).
  • Under U(x) = log base 2 of x: utility of 2 to the n is exactly n. Expected utility is (1/2)(1) + (1/4)(2) + (1/8)(3) + ... = sum of n divided by 2 to the n for n from 1 to infinity = 2.
  • A log-utility agent should pay up to 2 utils worth of wealth — which in raw money terms is a very modest amount — to play the game, not an infinite price.

Pause and Check

Questions to use before you move into practice

Self-check questions

  • Do the probabilities I used sum to exactly 1, and did I include every possible outcome?
  • Am I computing expected value of money or expected utility, and is that choice appropriate for the stakes?
  • If I changed from a risk-neutral to a risk-averse utility function, would the recommendation change, and does that change match my intuition about the case?

Practice

Now apply the idea yourself

Move into practice only after you can name the standard you are using and the structure you are trying to preserve or evaluate.

Formalization Practice

Integrated

Compute Expected Value and Expected Utility

For each gamble or decision, compute the expected value in dollars. Then compute the expected utility using the utility function U(x) = square root of x, treating x as total wealth. State which option the risk-neutral agent prefers and which option the square-root-utility agent prefers, and explain why they can differ.

Five expected-value and expected-utility problems

Show your arithmetic for each step. Where wealth matters for the utility calculation, assume the agent starts with 100 dollars unless stated otherwise.

Case 1 — Fair coin, unfair payoff

A fair coin is flipped. Heads pays 50 dollars, tails pays 0 dollars. Compute the expected value of the payoff. Then compute the expected utility of total wealth after the flip assuming a starting wealth of 100 dollars and U(x) = square root of x. Compare with a sure 25-dollar payment added to the starting wealth.

Does the risk-neutral agent prefer the gamble or the sure thing? Does the square-root-utility agent?

Case 2 — Insurance decision

A homeowner has 100,000 dollars in wealth. There is a 2 percent chance of a fire causing 60,000 dollars in damage. Insurance against the fire costs 1,500 dollars. Compute the expected value of buying insurance versus not buying insurance. Then compute expected utility under U(x) = square root of x. Does risk aversion change the recommendation?

Compare the expected-value recommendation and the expected-utility recommendation explicitly.

Case 3 — Unfair die

A weighted die lands on 6 with probability 1/3 and on each of the other faces with probability 2/15 each. Pay 5 dollars to roll. Win 20 dollars on a 6 and nothing otherwise. Compute the expected value of the net payoff. Would a risk-neutral agent play?

Check that the probabilities sum to 1 and that you are computing net rather than gross expected value.

Case 4 — Salary versus commission

A job offer gives the choice of a flat 60,000 dollar salary or a commission structure with equal chances of 40,000, 60,000, or 100,000 dollars. Compute both expected values. Then, assuming the agent has no other wealth and uses U(x) = square root of x, compute the expected utility of each offer. Which does a square-root-utility agent prefer?

Notice that the commission has a higher expected value but also higher variance, and see how the utility function treats that.

Case 5 — Small gamble on low stakes

A lottery ticket costs 1 dollar and has a 1-in-100 chance of paying 80 dollars and a 99-in-100 chance of paying nothing. Compute the net expected value. Then compute the expected utility assuming a starting wealth of 1,000 dollars and U(x) = square root of x. Explain why the utility analysis is very close to the expected-value analysis when the stakes are small relative to total wealth.

The closer the stakes are to zero relative to wealth, the closer expected utility comes to expected value.

Not saved yet.

Quiz

Integrated

Concept Check: Expected Value and Utility

Answer each question in one or two sentences, making explicit reference to the formulas and distinctions introduced in the lesson.

Short-answer concept check

Use precise decision-theoretic language. Where calculations are requested, show the key step.

Q1

State the expected-value formula for a lottery with outcomes x1, x2, x3 and probabilities p1, p2, p3. What constraint must the probabilities satisfy?

Include the sum-to-one condition on the probabilities.

Q2

A bet pays 200 dollars with probability 0.25 and 0 dollars with probability 0.75. What is the expected value of the bet? At what price would it be a fair bet?

Compute the expected value and compare it to the price.

Q3

Explain in plain language what the St Petersburg paradox shows about the limits of using expected monetary value as a decision rule.

Mention the divergence of the sum and the role of diminishing marginal utility in the resolution.

Q4

What shape of utility function corresponds to risk aversion, and why? Sketch in words how the curve looks.

Mention concavity and diminishing marginal utility explicitly.

Q5

Suppose an agent with logarithmic utility is offered a 50/50 gamble between doubling her wealth and losing half of it. Is the gamble attractive to her? Explain using the log utility function.

Note that log(2W) plus log(W/2) equals log(W squared), so the expected log wealth is exactly log(W), meaning the gamble is a wash for log-utility agents.

Not saved yet.

Analysis Practice

Integrated

Multi-Mode Analysis: Expected Value and Expected Utility

The scenario below requires multiple reasoning modes. Identify which types of reasoning (deductive, inductive, abductive, etc.) are at work and evaluate each strand.

Practice scenarios

Work through each scenario carefully. Apply the concepts from this lesson.

Case Study

A pharmaceutical company tested a new drug on 5,000 patients (inductive basis). Their hypothesis was that the drug blocks a specific receptor (abductive reasoning). If the receptor is blocked, inflammation should decrease (deductive prediction). They observed a 35% reduction in inflammation markers.

Case Study

An archaeologist found pottery fragments at a site (evidence). She reasoned: all known pottery from this region uses local clay (general rule). These fragments use local clay (observation). Therefore, they were likely made locally (deduction). Similar fragments found 200 miles away suggest a trade network (abduction).

Not saved yet.

Classification Practice

Integrated

Identify the Reasoning: Expected Value and Expected Utility

For each passage, identify the primary reasoning type being used (deductive, inductive, abductive, analogical, etc.) and justify your classification.

Practice scenarios

Work through each scenario carefully. Apply the concepts from this lesson.

Passage 1

Since the last five winters have each been warmer than the previous one, next winter will probably be warmer still.

Passage 2

The footprints lead to the window, the glass is broken outward, and the alarm was disabled from inside. The most likely explanation is that someone broke out, not in.

Passage 3

All valid syllogisms with true premises have true conclusions. This syllogism is valid and has true premises. Therefore, its conclusion is true.

Passage 4

Just as a gardener must prune dead branches to help a tree grow, a manager must sometimes cut underperforming projects to help a company thrive.

Not saved yet.

Guided Problem Solving

Integrated

Synthesis Challenge: Expected Value and Expected Utility

Construct a well-structured argument about the given topic that uses at least two distinct reasoning modes. Clearly label where each mode appears.

Practice scenarios

Work through each scenario carefully. Apply the concepts from this lesson.

Topic

Should cities invest in renewable energy over natural gas? Construct an argument that combines empirical evidence (inductive), a general principle (deductive), and an explanation of observed trends (abductive).

Topic

Is remote work more productive than office work? Build an argument using statistical evidence, logical principles, and the best explanation for conflicting findings.

Not saved yet.

Analysis Practice

Integrated

Deep Practice: Expected Value and Expected Utility

Each scenario involves multiple reasoning types working together. Identify and evaluate each reasoning strand, then assess the overall argument quality.

Multi-modal reasoning cases

Decompose each argument into its component reasoning types and evaluate each independently.

Case A

A climate scientist argues: (1) CO2 levels have risen 40% since 1850 (empirical data). (2) Lab experiments show CO2 traps infrared radiation (deductive/experimental). (3) The best explanation for observed warming patterns is greenhouse gas accumulation (abductive). (4) All greenhouse gases trap heat; CO2 is a greenhouse gas; therefore CO2 traps heat (deductive). (5) If current trends continue, temperatures will rise 2-4 degrees by 2100 (inductive projection).

Case B

A defense attorney argues: (1) DNA evidence is absent from the scene (observation). (2) The defendant's alibi is confirmed by three witnesses (testimonial evidence). (3) The best explanation for the lack of physical evidence is that the defendant was not present (abductive). (4) If the defendant was elsewhere, they could not have committed the crime (deductive). (5) In similar cases without physical evidence, conviction rates are below 10% (inductive/statistical).

Not saved yet.

Analysis Practice

Integrated

Real-World Transfer: Expected Value and Expected Utility

Apply what you have learned to these real-world contexts. Analyze each scenario using the tools and concepts from this lesson.

Transfer practice

Connect the concepts from this lesson to contexts outside the classroom.

Media literacy

A social media post claims: 'A new study proves that video games improve intelligence.' The post links to a study of 40 college students who played puzzle games for 2 weeks and showed improved scores on one type of spatial reasoning test. Evaluate this claim using what you know about arguments, evidence, and reasoning.

Everyday reasoning

A friend argues: 'I should not get vaccinated because my cousin got vaccinated and still got sick. Also, I read an article that said natural immunity is better.' Identify the types of reasoning, assess their strength, and explain what additional evidence would be relevant.

Professional context

A manager says: 'Our last three hires from University X performed well, so we should recruit exclusively from University X.' Analyze the reasoning type, identify potential problems, and suggest a better approach.

Not saved yet.

Rapid Identification

Integrated

Timed Drill: Expected Value and Expected Utility

For each passage, quickly identify all reasoning types present (deductive, inductive, abductive, analogical). Label each sentence or clause with its type.

Rapid reasoning-type identification

Tag each reasoning move in the passage. Aim for under 90 seconds per item.

Item 1

Since every tested sample contained trace metals (observation), and the factory upstream uses those metals (known fact), the contamination probably originates from the factory (best explanation). If the factory is the source, then water downstream should show higher concentrations than upstream (deduction).

Item 2

In 15 out of 16 observed cases, teams that adopted agile methodology delivered on time (data). This team adopted agile (observation). They will probably deliver on time (generalization). If they do deliver on time, the VP's prediction was wrong (conditional reasoning).

Item 3

The building's energy bill is 40% above comparable buildings (observation). The best explanation is poor insulation (abduction). All poorly insulated buildings lose heat through walls (general principle). This building will therefore lose heat through its walls (deduction). Based on three similar retrofit projects, insulation upgrades reduced bills by 25-30% (inductive projection).

Item 4

Just as antibiotics transformed medicine in the 20th century, AI may transform diagnostics in the 21st (analogy). Every major technological shift has created new job categories (inductive generalization). Therefore, AI will probably create new job categories (inductive conclusion). If new jobs emerge, retraining programs will be essential (conditional).

Item 5

The satellite images show forest cover decreased by 12% over five years (data). Logging permits increased 300% during the same period (correlation). The most likely cause is commercial logging (abduction). If deforestation continues at this rate, the watershed will be compromised within a decade (projection). All compromised watersheds eventually affect downstream water quality (deductive generalization).

Not saved yet.

Evaluation Practice

Integrated

Peer Review: Expected Value and Expected Utility

Below are sample student analyses that attempt to use multiple reasoning types. Evaluate: Did they correctly identify each reasoning type? Did they integrate them properly? What reasoning strands are missing?

Evaluate multi-modal student analyses

Each student attempted to analyze a complex argument using multiple reasoning types. Assess their work.

Student A's multi-modal analysis

Topic: Should the city build a new stadium? Student A wrote: 'Deductive: If the stadium generates more tax revenue than it costs, it is a good investment. The projected revenue exceeds costs by 20%. Therefore, it is a good investment. Inductive: 7 out of 10 cities that built stadiums saw economic growth. Our city will likely see growth too.' Missing reasoning type: abductive.

Student B's multi-modal analysis

Topic: Why are bee populations declining? Student B wrote: 'The best explanation is pesticide use (abductive). All neonicotinoids affect insect nervous systems (deductive). In 12 European studies, banning neonicotinoids correlated with bee recovery (inductive). My analysis integrates all three reasoning types for a comprehensive answer.'

Not saved yet.

Argument Building

Integrated

Construction Challenge: Expected Value and Expected Utility

Construct a multi-layered argument that deliberately combines at least three reasoning types. Label each component clearly and explain how the different reasoning modes work together.

Build integrated arguments

For each topic, construct an argument that uses at least three distinct reasoning types, clearly labeled.

Task 1

Build a comprehensive argument about whether autonomous vehicles should be allowed on public roads. Use deductive reasoning (from legal/ethical principles), inductive reasoning (from crash statistics or pilot programs), and abductive reasoning (best explanation for observed safety patterns). Label each component.

Task 2

Construct an argument about the effectiveness of universal basic income. Integrate: (1) a deductive argument from economic principles, (2) inductive evidence from pilot programs, (3) an analogical argument comparing it to existing social programs, and (4) an abductive explanation for why some pilots succeeded and others did not.

Not saved yet.

Diagnosis Practice

Integrated

Counterexample Challenge: Expected Value and Expected Utility

Each argument below uses multiple reasoning types. Find the weakest link and construct a counterexample that targets that specific reasoning strand without affecting the others.

Target the weakest reasoning strand

Find the most vulnerable reasoning type in each multi-modal argument and attack it with a specific counterexample.

Argument 1

Claim: Electric vehicles are better for the environment. Deductive strand: If EVs produce zero tailpipe emissions and zero emissions is better, EVs are better. Inductive strand: In 20 studies, regions with more EVs had lower air pollution. Construct a counterexample targeting either strand.

Argument 2

Claim: Remote work increases productivity. Abductive strand: The best explanation for higher output during COVID lockdowns is remote work. Inductive strand: In three company-wide surveys, remote workers reported higher productivity. Analogical strand: Just as flexible scheduling improved factory output, flexible location will improve knowledge-work output. Identify the weakest strand and construct a counterexample.

Argument 3

Claim: Meditation reduces anxiety. Deductive strand: All activities that lower cortisol reduce anxiety; meditation lowers cortisol; therefore, meditation reduces anxiety. Inductive strand: 8 of 10 randomized controlled trials showed reduced anxiety scores. Construct a counterexample that shows how the deductive strand might be unsound.

Not saved yet.

Analysis Practice

Integrated

Integration Exercise: Expected Value and Expected Utility

These advanced exercises require you to deploy every reasoning type you have learned. Analyze each complex scenario using deductive, inductive, abductive, and problem-solving approaches together.

Full-spectrum reasoning exercises

Each scenario demands all major reasoning types. Label and evaluate each reasoning strand.

Grand Challenge 1

A country is debating whether to implement a carbon tax. Evidence: (a) 15 countries with carbon taxes reduced emissions by an average of 12%, (b) economic models predict a 0.3% GDP reduction, (c) the country's constitution requires that tax policy must not disproportionately burden low-income households, (d) the leading explanation for why some carbon taxes failed is poor revenue recycling. Analyze using all reasoning types: evaluate the inductive evidence, check the deductive constitutional constraint, assess the abductive explanation for failures, and propose a problem-solving approach.

Grand Challenge 2

A hospital system is deciding whether to adopt a new AI diagnostic tool. Data: (a) the tool has 94% accuracy in clinical trials involving 20,000 patients, (b) human doctors average 88% accuracy for the same conditions, (c) the tool performs worse on underrepresented demographic groups, (d) hospital policy states that any diagnostic tool must meet or exceed human accuracy for all patient groups. Apply all reasoning types to analyze this decision.

Not saved yet.

Diagnosis Practice

Integrated

Misconception Clinic: Expected Value and Expected Utility

Each item presents a misconception about how different reasoning types relate to each other. Identify the error and explain the correct relationship between reasoning modes.

Misconceptions about reasoning integration

Diagnose and correct each misconception about how reasoning types work together.

Misconception 1

A student says: 'Deductive reasoning is always better than inductive reasoning because deduction gives certainty while induction only gives probability.'

Misconception 2

A student claims: 'You should never mix reasoning types in a single argument. Each argument should use only one type of reasoning to stay rigorous.'

Misconception 3

A student writes: 'Abductive reasoning is just a combination of deduction and induction. It does not have its own distinct logic.'

Misconception 4

A student argues: 'An argument is only as strong as its weakest reasoning strand. If one part is inductive (and therefore uncertain), the whole argument is uncertain.'

Misconception 5

A student says: 'Analogical reasoning is not a real form of reasoning -- it is just a rhetorical device. You cannot draw legitimate conclusions from analogies.'

Not saved yet.

Analysis Practice

Integrated

Scaffolded Multi-Modal Analysis: Expected Value and Expected Utility

Build a comprehensive analysis in stages, adding one reasoning type at a time. At each stage, explain what the new reasoning mode contributes that previous modes could not.

Layer-by-layer reasoning

Add one reasoning type per stage and explain what each uniquely contributes.

Scaffold 1

Topic: Should a city ban single-use plastics? Stage 1 (Deductive): State any logical principles or definitions that frame the issue. Stage 2 (Inductive): What empirical evidence supports or undermines a ban? Stage 3 (Abductive): What is the best explanation for why some bans succeed and others fail? Stage 4 (Problem-solving): Design an implementation approach that accounts for your findings. Stage 5 (Integration): How do the four reasoning strands combine into a coherent recommendation?

Scaffold 2

Topic: Is social media harmful to adolescents? Stage 1 (Inductive): Summarize the empirical evidence. Stage 2 (Abductive): What best explains the conflicting study results? Stage 3 (Deductive): If certain principles about child welfare are accepted, what follows logically? Stage 4 (Analogical): Compare to previous technology concerns (TV, video games). Stage 5 (Integration): Synthesize all strands into a balanced conclusion.

Not saved yet.

Analysis Practice

Integrated

Synthesis Review: Expected Value and Expected Utility

These capstone-level exercises require you to deploy every reasoning skill you have learned across all units. Analyze each complex real-world scenario using deductive, inductive, abductive, analogical, and problem-solving approaches.

Full-spectrum synthesis review

Use every reasoning tool at your disposal. Label each reasoning type clearly.

Grand Synthesis 1

A national education policy proposes replacing letter grades with narrative assessments for all K-12 students. Available evidence: (a) Three pilot programs showed improved student engagement but inconclusive effects on learning outcomes. (b) Universities say they need standardized metrics for admissions. (c) Teachers in pilot programs reported spending 3x more time on assessments. (d) Student anxiety about grades decreased in pilot schools. (e) Parents in pilot schools had mixed reactions -- 55% positive, 45% negative. Analyze this policy using all reasoning types: evaluate the inductive evidence, construct deductive arguments from educational principles, provide abductive explanations for the mixed results, draw analogies to other educational reforms, and apply problem-solving to the implementation challenges.

Grand Synthesis 2

A tech company must decide whether to open-source its AI model. Arguments for: transparency, community contributions, trust building. Arguments against: competitive advantage, safety risks, loss of revenue. Data: (a) 70% of companies that open-sourced saw increased revenue within 3 years, (b) two open-sourced AI models were misused for generating misinformation, (c) the company's terms of service prohibit misuse but enforcement is difficult. Apply every reasoning type to analyze this decision comprehensively.

Not saved yet.

Animated Explainers

Step-by-step visual walkthroughs of key concepts. Click to start.

Read the explanation carefully before jumping to activities!

Riko

Further Support

Open these only if you need extra help or context

Mistakes to avoid before submitting
  • Do not report the expected value as the amount you will actually receive on a single trial.
  • Do not use a linear utility function when the stakes are large enough that losing everything would be ruinous.
Where students usually go wrong

Computing expected value using probabilities that do not sum to 1 because one outcome was overlooked.

Interpreting a computed expected value as a forecast of the next trial rather than as a long-run average.

Mixing dollar payoffs and utility scores in the same summation.

Assuming every rational agent is risk-neutral and recommending the highest-expected-value option regardless of stakes or utility shape.

Using a linear utility function for a high-stakes decision and producing an unrealistic recommendation.

Historical context for this way of reasoning

Daniel Bernoulli

Bernoulli's 1738 paper on the St Petersburg paradox is one of the most important single documents in the history of decision theory. By proposing that people value money by its utility rather than its face amount, he laid the foundation for every later treatment of risk aversion and expected utility, including the 20th-century axiomatic work of von Neumann and Morgenstern.