PM OS
Module 4Intermediate130 min

The Art of Prioritization

Master ten frameworks and the judgment to pick the right one — with five live simulators to practice on.

RICEICEValue vs EffortKano modelMoSCoWWSJF / Cost of DelayStack RankingBuy a FeatureOpportunity Scoring (ODI)Theme ScoringEisenhower MatrixConfidence weighting

Explainer

Prioritization is not about producing the objectively correct list. It is a structured conversation about value, confidence, timing, and constraints. The framework is a forcing function for honesty — a way to make implicit assumptions explicit and to expose disagreement to the air. The wrong question is 'which framework is best'. The right question is 'which framework forces our specific team to confront what we usually avoid'.

1

Why Prioritization Frameworks Exist

Every prioritization framework is a heuristic for trading off four things: value, confidence, effort, and time. Different frameworks foreground different trade-offs. RICE foregrounds confidence; Kano foregrounds emotional payoff; WSJF foregrounds time; Value/Effort foregrounds the leverage ratio. The framework that fits your team is the one that makes you argue about the thing you were avoiding.

  • Frameworks expose disagreement, they don't resolve it.
  • A score that everyone agrees on without debate is suspect — usually nobody is being honest about confidence.
  • Use multiple frameworks for the same backlog and look for items that score divergently — those are the most interesting.
2

RICE: Reach × Impact × Confidence ÷ Effort

Intercom's RICE is the most common scoring framework in modern PM. Reach is the number of users or events affected per period. Impact is a coarse multiplier (Massive=3, High=2, Medium=1, Low=0.5, Minimal=0.25). Confidence is a percentage that discounts speculative scores. Effort is person-months. The score is comparable across initiatives in the same backlog, even when they're in very different areas of the product.

  • Reach must be a real number with a time period — '5,000 users per quarter' not 'lots'.
  • Impact is intentionally coarse — the precision is fake; the discrete buckets force conversation.
  • Confidence must reflect evidence, not enthusiasm — 50% is the honest default for unsubstantiated ideas.
  • Effort should include design, eng, QA, comms, and ramp-down support — not just code.
3

ICE: Impact × Confidence × Ease

Sean Ellis's ICE is RICE's faster cousin. Each input is a 1-10 score; the product is the rank. Use ICE for high-throughput experiment backlogs (growth experiments, conversion tests) where you'd burn a week scoring RICE on each. The trade-off: no Reach component, so ICE is biased toward easy wins and against systemic improvements with broad reach but harder construction.

  • ICE is best for experiment backlogs, not feature backlogs.
  • Ease is the inverse of effort — higher Ease = faster to ship.
  • Score from gut, calibrate against actual outcomes; recalibrate scores monthly.
  • Pair with a separate 'theme' tag so you don't accidentally optimize away from a strategic theme.
4

Value vs Effort (2×2 Matrix)

The simplest visual prioritization tool. Plot each item on a 2×2 of value (impact on the goal) versus effort. Quadrants: quick wins (high value, low effort), big bets (high value, high effort), fill-ins (low value, low effort), thankless (low value, high effort — drop). Best used for a 30-minute alignment workshop where the conversation is the deliverable, not the matrix itself.

  • Force the team to place every item; ties are a sign of unclear definitions.
  • Use sticky notes physically or a virtual board; debate the placement openly.
  • Quick wins ship first; big bets get sequenced; fill-ins fill capacity; thankless die.
  • Don't over-engineer — this is for clarity, not for tracking.
5

Kano Model: Must-Have, Performance, Delighter

Noriaki Kano's model classifies features by the relationship between feature presence and user satisfaction. Must-haves are invisible when present and catastrophic when absent (login working, data persistence). Performance features have a linear relationship — more is better (faster load times, more storage). Delighters are unexpected positives that don't hurt when absent but generate love when present. Reverse features actively annoy users when present.

  • Survey users with a feature-functional / feature-dysfunctional question pair to classify each feature.
  • Must-haves are oxygen — necessary but not differentiating.
  • Performance features are the bulk of the roadmap; pick the ones with the steepest satisfaction curve.
  • Delighters wear off; today's delighter is tomorrow's must-have (think Uber's map showing the driver coming).
6

MoSCoW: Must / Should / Could / Won't

MoSCoW is a release-scoping tool more than a roadmap tool. Must-have items are non-negotiable for the release. Should-have are important but not critical. Could-have are nice-to-haves. Won't-have are explicitly out of scope, named to prevent scope creep. Useful in fixed-deadline contexts (regulatory, launches) where the question is 'what's in this release', not 'what should we work on next'.

  • Most teams over-Must — calibrate by asking 'would we delay launch a week to ship this?'
  • Won't-have is the most underused category; explicitly listing exclusions reduces scope creep.
  • Re-categorize at every release, not as a static document.
  • Pair with a date; without a date, MoSCoW is just renamed wishlist.
7

WSJF: Weighted Shortest Job First

WSJF (from SAFe / lean) prioritizes by Cost of Delay divided by Job Size. Cost of Delay = User-Business Value + Time Criticality + Risk Reduction / Opportunity Enablement. The framework excels when timing matters — regulatory deadlines, market windows, dependencies blocking other work. Use it for a backlog where 'when' is as important as 'what'.

WSJF = (User-Business Value + Time Criticality + Risk Reduction & Opportunity Enablement) / Job Size
  • Use Fibonacci-like 1/2/3/5/8/13 scales for each component.
  • Time criticality captures decay: how much value drops the longer we wait.
  • Risk reduction / opportunity enablement captures the second-order value of unblocking other work.
  • Higher WSJF first. Surprisingly often, smaller jobs with high time criticality win over big bets.
8

Cost of Delay

Cost of Delay (Don Reinertsen) is the economic value lost per unit time by not shipping. It generalizes WSJF and is the most rigorous lens when finance is in the room. CoD per week × weeks of delay = the lost value. Useful when you need to make the case for hiring, parallelizing, or descoping in dollars rather than feelings.

  • Estimate as a range, not a point — '$50-150k per week' is honest; '$73,520' is theater.
  • Account for non-linear decay: launch windows, regulatory deadlines, viral moments.
  • Pair with team capacity to see whether parallelizing actually reduces total CoD or just stretches each project.
  • Reinertsen: 'If you only quantify one thing, quantify Cost of Delay.'
9

Stack Ranking

The simplest framework: lay out the entire backlog and force-rank from 1 to N. No ties allowed. The discipline is in the disagreement: if two items are 'tied for #3', the team avoids the trade-off. Stack ranking forces it. Useful as a final pass after RICE/ICE/Kano have produced a candidate ordering — humans cross-check the score-implied ranking against intuition and surface mis-scored items.

  • Print the backlog and physically order it on a wall — the spatial layout exposes weak rationales.
  • When stuck on a tie, ask: 'if I could only ship one of these next quarter, which?'
  • Use as the final pass, not the only pass.
10

Buy a Feature

Luke Hohmann's collaborative game: give each stakeholder a fixed budget of fake currency, present a menu of features with prices proportional to effort, and let stakeholders 'buy' the features they value most — including pooling money to buy expensive items. Reveals true preference, surfaces coalitions, and makes scarcity tangible.

  • Best for customer-facing prioritization sessions (advisory boards, design partner programs).
  • The pooling behavior is the most interesting signal — stakeholders' coalitions reveal alignment.
  • Don't take results literally as a roadmap; treat as discovery into demand intensity.
  • Variation: 'Pay for a Feature' with discounted bundles tests packaging hypotheses.
11

Opportunity Scoring (ODI)

Tony Ulwick's outcome-driven scoring: for each desired outcome (not feature), survey users for importance and current satisfaction, both 1-10. Opportunity = Importance + max(0, Importance − Satisfaction). High-opportunity outcomes (≥12) are under-served; low-opportunity outcomes (<10) are over-served or saturated. Drives investment toward outcomes the user cares about and the market hasn't satisfied.

Opportunity = Importance + max(0, Importance − Satisfaction)
  • Outcomes must be solution-agnostic statements ('minimize time to identify duplicates').
  • Survey n≥100 users in the target segment to get stable scores.
  • Avoid surveying users who don't actually do the job — irrelevant data dilutes signal.
  • Re-survey annually; saturation drifts as the market evolves.
12

Theme Scoring

When the backlog is 200 items, scoring each is a waste of energy. Theme scoring rolls items into 5-10 strategic themes (e.g. 'reduce time-to-first-value', 'expand into mid-market', 'platform reliability'), scores the themes, and allocates capacity by theme. Within a theme, the team prioritizes tactically. Best for product orgs with strategic themes and many small bets.

  • Score themes by strategic alignment + measurable outcome value, not feature-level scores.
  • Allocate capacity by theme (e.g. 50% reliability, 30% growth, 20% experiments).
  • Within each theme, use RICE or ICE for tactical ordering.
  • Re-balance themes quarterly with leadership; tactical choices stay with the team.
13

Confidence as a First-Class Score

The single most underused dimension in PM prioritization is honest confidence. A massive idea backed by weak evidence should not outrank a smaller idea with clear demand by default. Confidence prevents speculative reach and impact from dominating the roadmap. Make it explicit, evidence-tagged, and revisited.

  • Define confidence levels with evidence anchors: 100% (already validated by data), 80% (qualitative + leading indicator), 50% (interview support, no data), 30% (informed hypothesis).
  • Tag the evidence behind each confidence score.
  • Revisit and adjust confidence at every roadmap review.
  • When confidence drops, items can rank lower — which is the point.
14

Picking the Right Framework

Match the framework to the decision. New product backlog with mixed bet sizes? RICE. Growth experiment backlog? ICE. Quick alignment workshop? Value/Effort. Stakeholder demand session? Buy a Feature. Release scoping with a deadline? MoSCoW. Time-sensitive backlog with dependencies? WSJF. Large mature backlog? Theme scoring. The framework you reach for first reveals the kind of decisions you face — and the discipline you've practiced.

Framework atlas

Reference cards for each method in this mission

Expand a card for when to deploy it, misuse patterns, sequencing guidance, and (where relevant) shorthand formulas.

Quantitative scoring · Sean McBride at Intercom (2017)RICE(RICE)

Reach × Impact × Confidence / Effort. The most common cross-functional scoring framework.

RICE = (Reach × Impact × Confidence%) / Effort

When to use

  • Roadmap candidate backlog with mixed bet sizes.
  • Cross-team prioritization where transparency matters.
  • When confidence varies wildly across items.

When not to

  • Pure experiment backlogs (use ICE, faster cycle).
  • When all items are within the same theme (RICE noise > signal).

How to apply

  1. Estimate Reach as a number per period (users, events, or sessions).
  2. Pick Impact from a coarse scale: Massive=3, High=2, Medium=1, Low=0.5, Minimal=0.25.
  3. Set Confidence honestly with an evidence tag (100/80/50/30%).
  4. Estimate Effort in person-months across all functions.
  5. Compute the score; rank descending; argue about top vs bottom of ties.
Pitfalls / anti-patterns
  • Reach written as 'all users' — meaningless without a time period.
  • Confidence inflated to win the comparison.
  • Effort that excludes design, ops, comms, and ramp-down.
Quantitative scoring · Sean Ellis (popularized through GrowthHackers)ICE(ICE)

Impact × Confidence × Ease, each on a 1-10 scale. Lightweight scoring for high-volume experiment backlogs.

ICE = Impact × Confidence × Ease

When to use

  • Growth experiment backlogs.
  • When you need a score in <2 minutes per item.
  • Backlogs where Reach is roughly equal across items.

When not to

  • Roadmap-level prioritization with mixed bet sizes.
  • Cross-team trade-offs (RICE's transparency helps more here).

How to apply

  1. Score Impact (1-10): expected lift on the target metric.
  2. Score Confidence (1-10): evidence quality.
  3. Score Ease (1-10): inverse effort.
  4. Rank by product; calibrate weekly against actual experiment outcomes.
Pitfalls / anti-patterns
  • Same person scoring all three components — biased.
  • Never recalibrating against actual outcomes.
Visual prioritization · Lean / agile workshopsValue vs Effort 2×2

Simple 2×2 matrix plotting items by value and effort to surface quick wins, big bets, fill-ins, and items to drop.

When to use

  • Quick alignment in a 30-minute workshop.
  • Sanity-checking a RICE-ordered list visually.

When not to

  • When you need numerical comparability across many items.

How to apply

  1. Define the value dimension explicitly (revenue impact? activation lift? NSM movement?).
  2. Plot every item with the team co-present.
  3. Cluster into quick wins (DO), big bets (PLAN), fill-ins (FILL), thankless (DROP).
  4. Sequence quick wins immediately; scope big bets formally.
Pitfalls / anti-patterns
  • Letting 'value' stay vague — produces a useless map.
  • Skipping the DROP quadrant — it's the most valuable category.
Customer satisfaction · Noriaki Kano (1984)Kano Model

Classifies features by the relationship between presence and satisfaction: Must-Have, Performance, Delighter, Indifferent, Reverse.

When to use

  • Designing a tier or plan structure.
  • Detecting under-investment in basics or over-investment in delighters.

When not to

  • Highly novel categories where users can't classify yet.

How to apply

  1. List candidate features.
  2. For each, ask users a functional and dysfunctional question pair (how would you feel if this feature existed / didn't exist?).
  3. Classify each feature by the response pattern.
  4. Re-score annually as delighters age into must-haves.
Pitfalls / anti-patterns
  • Treating Kano as a static taxonomy instead of a moving target.
  • Surveying customers who don't yet know the feature category.
Release scoping · Dai Clegg / DSDM (1990s)MoSCoW(MoSCoW)

Must-have, Should-have, Could-have, Won't-have-this-time. Release-scoping tool for fixed-deadline contexts.

When to use

  • Fixed-date launches (regulatory, partner integrations, conferences).
  • Releases with strict scope boundaries.

When not to

  • Continuous-delivery products without releases as a unit.

How to apply

  1. List every candidate item for the release.
  2. Categorize as Must / Should / Could / Won't.
  3. Validate: would you delay the release a week to land each Must? If not, demote.
  4. Publish the Won't list to set expectations explicitly.
Pitfalls / anti-patterns
  • Over-Must-ing under stakeholder pressure.
  • Treating Won't as a deferral list instead of an exclusion list.
Time-sensitive scoring · SAFe / Don ReinertsenWSJF(WSJF)

Weighted Shortest Job First: Cost of Delay divided by Job Size. Prioritizes high-CoD short jobs first.

WSJF = (User-Business Value + Time Criticality + Risk Reduction & Opportunity Enablement) / Job Size

When to use

  • Backlogs where time-to-value or sequencing matters.
  • Dependency-heavy programs where unblocking work has compounding value.

When not to

  • Pure feature backlogs without time pressure.

How to apply

  1. Score each component on a Fibonacci-like scale: 1, 2, 3, 5, 8, 13.
  2. Sum the value components; divide by job size.
  3. Sequence highest WSJF first.
Pitfalls / anti-patterns
  • Confusing Job Size with story points — Job Size includes dependency unblocking.
  • Ignoring time criticality decay.
Economic prioritization · Don ReinertsenCost of Delay

The economic value lost per unit time by not shipping. Quantifies the urgency case in dollars.

When to use

  • Cross-team capacity allocation arguments.
  • Hiring decisions ('how much will the second engineer save us per month?').
  • When finance partners want a dollar-denominated case.

When not to

  • Early-stage products without revenue mechanics.

How to apply

  1. Identify the value the project will create (annual revenue, retention dollars, support hours saved).
  2. Estimate value-per-week if shipped today vs in 13 weeks.
  3. Account for non-linear decay (launch windows, deadlines).
  4. Express as a range; communicate to stakeholders weekly.
Forced-choice ordering · Industry-wide practiceStack Ranking

Force-rank the entire backlog 1..N with no ties allowed. Surfaces hidden disagreements that scored frameworks miss.

When to use

  • Final pass after RICE/ICE has produced a candidate order.
  • Small backlogs (<30 items).

When not to

  • Backlogs of 100+ items — use theme scoring first.

How to apply

  1. List all items.
  2. Force-rank with no ties.
  3. When stuck, ask 'if we could ship only one, which?'
  4. Reconcile against framework scores; investigate any large divergences.
Collaborative · Luke HohmannBuy a Feature

Stakeholders allocate fake currency to features priced by effort. Pooling and trading reveal demand intensity and coalitions.

When to use

  • Customer advisory boards.
  • Internal alignment workshops with multiple competing priorities.

When not to

  • Solo PM scoring — Buy a Feature requires multiple participants.

How to apply

  1. Price each feature at 1-3 times its effort proportional cost.
  2. Give each participant a budget that can buy ~1/3 of the menu.
  3. Run a 30-45 minute session; allow pooling.
  4. Debrief on why pooling happened; that's the strategic signal.
Outcome-based scoring · Tony Ulwick / StrategynOpportunity Scoring (ODI)

Score outcomes by importance and current satisfaction; opportunity = importance + max(0, importance − satisfaction).

Opportunity = Importance + max(0, Importance − Satisfaction)

When to use

  • Mature markets with well-understood jobs.
  • Roadmap arbitration in a saturated category.

When not to

  • Brand-new categories where users can't articulate the job yet.

How to apply

  1. Decompose the job into outcome statements.
  2. Survey users (n≥100) on importance and satisfaction.
  3. Compute opportunity score per outcome.
  4. Prioritize outcomes with score ≥ 12 as under-served.
Personal triage · Dwight D. EisenhowerEisenhower Matrix

Urgent vs Important 2×2 for personal task triage. Useful for managing the PM's own workload, not for product prioritization at scale.

When to use

  • Personal weekly triage of inbound asks.
  • Distinguishing reactive support from strategic work.

When not to

  • Roadmap prioritization — 'important' is too vague at scale.

How to apply

  1. List all open asks/tasks.
  2. Tag each: urgent or not; important or not.
  3. Triage: do, schedule, delegate, drop.

Product Psychology

Cognitive biases that distort product decisions

Confidence Inflation

Scorers inflate confidence to ensure their preferred items rank highly.

Product Risk

Speculative bets dominate the roadmap because nobody will admit a 30% confidence score.

Research Countermove

Tag every confidence score with the specific evidence behind it. Audit retroactively: did 80%-confidence items actually hit?

Effort Sandbagging

Engineers overestimate effort on items they don't want to do; underestimate on items they want to ship.

Product Risk

Roadmaps reflect engineering preferences disguised as effort scores.

Research Countermove

Track estimated vs actual effort over time per scorer; calibrate against history; consider a second estimator.

Reach Hallucination

Optimistic Reach estimates because nobody fact-checks the user counts.

Product Risk

Big-feeling features dominate over targeted features that are genuinely more impactful.

Research Countermove

Pull reach from analytics, not hopes. If the segment can't be quantified, set Reach = 0 and force the team to instrument first.

HiPPO Effect

Highest Paid Person's Opinion overrides scoring frameworks.

Product Risk

Frameworks become rationalization for predetermined exec preferences.

Research Countermove

Score blind first; reveal to leadership only after the team has converged. Disagreements with leadership become explicit conversations, not silent overrides.

Organizational anti-patterns

When ceremonies look like rigor but aren't

Score Then Ignore

Team produces a beautifully scored RICE sheet, then ships in the order leadership prefers anyway.

The framework is performative; real authority is elsewhere.

Fix

Either commit to the score or admit the framework is advisory. Performative scoring corrodes trust faster than honest political prioritization.

All Confidence at 80%

Every item has confidence 70-90%; the lever is never used.

Confidence below 50% feels like an admission of failure; nobody wants to write it down.

Fix

Define explicit confidence anchors (with examples). Reward calibration over confidence.

Effort Without Capacity Check

Top-ranked RICE items add up to 18 person-months in a 3-month quarter; team commits anyway.

Prioritization and capacity planning live in different docs.

Fix

After ranking, draw the line at actual capacity. Items below the line are explicit not-doing; their existence is information for next quarter.

One Framework Forever

Team uses RICE for everything regardless of decision type.

Cargo-culting; nobody on the team has tried the alternatives.

Fix

Use at least three frameworks across the year. Notice which one each kind of decision gravitates toward.

Resurrection Backlog

Items that lost prioritization three quarters ago keep returning, slightly rebranded, against the same evidence.

Stakeholders feel unheard; the team has no public record of why items were declined.

Fix

Maintain a public 'declined items log' with the reason and the evidence threshold for revisiting. Items can return only when the threshold is met.

Worked examples

Walkthroughs translated from real trade-off rooms

Walking through a real RICE comparison

A B2B SaaS team is comparing two roadmap candidates: 'Onboarding checklist' and 'Bulk admin invite flow'.

  1. Onboarding checklist: Reach 800/quarter (new accounts), Impact 2 (High — activation step), Confidence 85% (validated in 8 user interviews), Effort 3 PM. Score = (800 × 2 × 0.85) / 3 = 453.
  2. Bulk admin invite: Reach 260/quarter (mid-market signups only), Impact 3 (Massive — unlocks team activation), Confidence 60% (qualitative only, no behavioral data), Effort 5 PM. Score = (260 × 3 × 0.6) / 5 = 94.
  3. First instinct says bulk-invite is bigger ('massive impact!'), but the smaller-reach + lower-confidence + bigger-effort combination drops it 5x below the checklist.
  4. Action: ship checklist now; instrument the bulk-invite hypothesis to raise its confidence; revisit next quarter.

TakeawayWithout RICE, gut feel might have prioritized the bulk invite. With RICE, the lower-confidence bigger bet correctly defers to the higher-confidence smaller bet.

Picking ICE over RICE for a growth experiment backlog

A growth team has 28 conversion experiments queued. RICE-scoring all of them would take a full week.

  1. Reach is roughly constant (all hit the same signup funnel) — so RICE's Reach lever is dead weight.
  2. Switch to ICE; score each from gut in a 90-minute session.
  3. Top-scoring experiment: 'Add social proof on signup'. ICE = 7×6×9 = 378.
  4. Bottom-scoring: 'Animated hero on landing page'. ICE = 4×3×5 = 60.
  5. Run the top 3; recalibrate scores against actual lift; adjust scoring style.

TakeawayICE was the right tool for high-volume same-Reach experiments. RICE's transparency overhead would have starved velocity.

Using WSJF when the deadline is real

Three roadmap items are competing: 'GDPR DPA renewal automation', 'New-user activation experiment', 'Premium plan tier launch'.

  1. GDPR DPA: Job 5, User-Business Value 5, Time Criticality 13 (regulatory deadline 4 months out), Risk Reduction 8. WSJF = (5+13+8)/5 = 5.2.
  2. Activation experiment: Job 2, V 5, TC 3, RR 2. WSJF = (5+3+2)/2 = 5.0.
  3. Premium tier: Job 8, V 13, TC 5, RR 5. WSJF = (13+5+5)/8 = 2.875.
  4. Rank: GDPR > Activation experiment > Premium tier. Despite Premium having the highest dollar value, its size and lower time criticality push it to third.

TakeawayWSJF correctly highlighted that the regulatory deadline created ambient cost-of-delay invisible to RICE. The big premium bet still gets done, just sequenced after the urgent shorter jobs.

Catching a Kano misclassification

Team wants to invest heavily in custom dashboard builders, treating them as a Delighter that will differentiate.

  1. Run Kano survey on 50 enterprise users: results show 71% classify dashboard builders as Must-Have or Performance, only 12% as Delighter.
  2. Reframe: dashboard builders are table-stakes, not delight. Delight will need to come from somewhere else.
  3. Re-prioritize: ship the dashboard builder MVP fast as a competitive must-have; redirect 'delight' investment to the AI-summarization feature that scored as a true Delighter (78% Delighter response).

TakeawayKano correctly redirected investment from a feature that would only meet expectations to a feature that would actually generate satisfaction lift.

Interactive lab

These instruments implement the textbook formulas loosely—use them to stress‑test judgments, compare frameworks on the same backlog, then document evidence and decisions.

← swipe to see all frameworks →

RICE

Reach × Impact × Confidence ÷ Effort (Intercom scale)

Reach × Impact × Confidence% ÷ Effort— Reach = users/events in one consistent period; Effort in person-months

score = (reach × impact × confidence/100) ÷ effort
1
RICE Score453.33
800×2×0.85÷3
2
RICE Score93.60
260×3×0.60÷5

Resources / Case Studies

Curated reading for this mission

Framework

The original Intercom blog post introducing the RICE framework with worked examples.

Source-of-truth article for RICE; the simulator in this curriculum implements the same math.

The seminal text on Cost of Delay, queuing theory, and product development economics.

The most rigorous treatment of why time-based prioritization matters; foundational for WSJF.

ICE Scoring Method

Sean Ellis / GrowthHackers

Framework

Sean Ellis's lightweight scoring method for high-velocity experiment backlogs.

Establishes when ICE is the right tool over heavier frameworks like RICE.

Kano Model

Mind the Product

Framework

Practical guide with templates for running a Kano survey, classifying responses, and avoiding common pitfalls.

Most accessible operational guide for actually running Kano analysis.

Hohmann's collaborative game for surfacing demand intensity from stakeholders.

Strong tool for advisory boards and internal alignment workshops.

What is Cost of Delay?

Joshua J. Arnold (Black Swan Farming)

Essay

Plain-language explanation of CoD with case studies showing the dollar impact of late shipments.

Best free introduction to CoD for PMs new to the concept.

Framework

Outcome-Driven Innovation's quantitative scoring method for under-served outcomes.

The clearest reference for combining JTBD discovery with quantitative roadmap prioritization.

Roadmaps Are Dead

Janna Bastow / ProdPad

Essay

The case for Now / Next / Later format and why date-based roadmaps damage trust.

Foundational reading for moving teams off Gantt-roadmap pathology.

Gilad's Confidence Meter — a structured way to attach evidence to confidence scores in any prioritization framework.

Operationalizes confidence in a way every framework benefits from.

Newsletter / Feed

Recurring case studies from PMs at large tech companies on how they actually prioritize.

Reality-check against the canonical frameworks; current practitioners share what works and what doesn't.

John Cutler's TBM Newsletter

John Cutler / Amplitude

Newsletter / Feed

Weekly notes from a product strategy lens on prioritization, theme allocation, and team operating models.

Sharpens systems-level thinking about prioritization beyond per-item scoring.

Shape Up

Ryan Singer / Basecamp

Book

Basecamp's alternative to backlog-driven development: shaped bets, fixed appetite, no estimation theater.

Provocative counter-view to scoring-based prioritization; worth reading even if you don't adopt it.