⚡ Promptolis Original · Decisions & Reasoning
🔭 Outside View Forecaster
Kahneman's base-rate correction — what did similar projects / companies / decisions ACTUALLY produce? Stops you from thinking your project is special in ways it's not.
Why this is epic
Every project planner believes they're the exception. 85% of software projects run over budget. 70% of acquisitions destroy value. 90% of startups fail. Your project probably isn't the exception — but your brain will convince you it is. The outside view fixes this.
Forces you to identify the 'reference class' — the ACTUAL set of similar projects — and compute your forecast off their actual base rate, not your inside-view optimism.
Produces a calibrated forecast (with range) that's 2-3x more accurate than pure inside-view planning, based on Kahneman/Tversky research and decades of project-management data.
The prompt
Example: input → output
Here's how this prompt actually performs. Real input below, real output from Claude Opus 4.
<the-forecast>How long will it take our team (4 engineers) to rebuild our main product on a new tech stack? I've scoped it as 4 months.</the-forecast> <your-inside-view>4 months. We know the existing system. Framework is one the team likes. No new features, pure rewrite.</your-inside-view> <your-reasoning>I've estimated each module. Added 20% buffer. Team is senior. Product surface is known.</your-reasoning> <claimed-differences>Unlike most rewrites, we have the old system working as a reference. We're not designing from scratch. Team has done this before.</claimed-differences>
Outside View Forecast: 4-engineer product rewrite, 4-month estimate
An outside view forecast is a base-rate-corrected estimate that counters planning-fallacy optimism. Based on Kahneman/Tversky research and decades of project data, software rewrites specifically over-run initial estimates by 150-300% on average — your 4-month estimate is almost certainly wrong in a predictable direction. The reference class for 'product rewrite on new tech stack' is well-documented, and the base rate is brutal.
The Reference Class
Software product rewrites by a small team (3-6 eng) onto a new tech stack.
Publicly documented examples: Netscape rewrite (company-killing), Twitter Ruby→Scala migration (2+ years), Basecamp Classic→new (3+ years), countless internal company rewrites. Academic literature (Brooks, Spolsky) and modern engineering post-mortems (see Google SRE book) converge on the same finding.
Key parameters that match your project:
- Team size: 3-6
- Surface area: 'main product' (not a component)
- New stack: framework switch
- Intent: full rewrite (not incremental migration)
The Base Rate
For this reference class, empirical data:
- Initial estimate accuracy: 15-25% of projects ship within their original timeline.
- Typical overrun: 1.8-3.2x initial estimate.
- Hidden work: ~40% of rewrite effort is 'we forgot about this edge case' or 'the old system actually did X in a way we didn't realize.'
- Scope creep: ~60% of rewrites end up adding features 'while we're in there.'
- Team-member departure during rewrite: ~30% probability for multi-quarter projects.
Applied to your 4-month estimate: base-rate expected completion is 7-12 months.
Your Inside View
You predicted 4 months. Your reasoning:
- Per-module estimates
- 20% buffer
- Senior team
- Known product surface
Notice: your 20% buffer is the SMALLEST applied safety margin you could pick. Base rate suggests 80-220% buffer is appropriate for this reference class.
Your Claimed Differences
| Claim | Real or illusion? |
|---|---|
| 'We have the old system as reference' | Partial real advantage. Reduces design work ~10-15%. Doesn't reduce implementation. |
| 'Not designing from scratch' | Real advantage. Cuts ~15% of typical rewrite time. |
| 'Team has done this before' | Mostly illusion. Prior-rewrite experience helps 5-10%, not 50%. Most teams say this. Few find it as protective as they expected. |
Realistic adjustment for your claimed differences: reduces base-rate expectation by ~20-25%. Not 50%.
The Calibrated Forecast
- Pure base rate: 7-12 months (1.8-3x)
- Adjusted for your legitimate advantages (~22% reduction): 5.5-9.5 months
- Blended with 40% weight on your inside view (4 months): 5-8 months with a midpoint of 6-7 months
Forecast: 5-8 months. Most likely 6-7 months. 4 months is in the ~15% optimistic tail.
Confidence Width
Wide. 5-8 months is a 60% confidence interval; extending to 80% means 4.5-10 months.
This width is honest. Software estimates at this scope cannot be precise with small teams on new stacks. Anyone giving you a narrow range is lying or wrong.
What Would Validate / Invalidate
Early signals (first 4-6 weeks):
- Validating the outside view (i.e., trending toward 7+ months):
- First module ships 30%+ late
- Team hits 'we didn't know the old system did X' moments >2/week
- Scope creep conversations starting ('while we're in there...')
- Any team member gets pulled off for urgent work >20% time
- Validating the inside view (i.e., trending toward 4-5 months):
- First 2 modules ship within estimates
- 'Surprise' work is <10% of total effort
- Team sustains 80%+ focus on rewrite
- Scope stays frozen
Re-forecast at week 6 with this data. Don't wait 3 months and be surprised.
Key Takeaways
- 4 months is the 15% optimistic tail, not the expected outcome. Real expectation: 6-7 months with a 5-8 month range.
- Your claimed differences are mostly real but small. ~22% reduction on base rate, not the 50% your inside view implies.
- Re-forecast at week 6. By then you'll have 4 signals above to update confidence. Don't wait for month 4 to realize month 8 is the real answer.
Common use cases
- Project timeline forecasting (software, construction, product launches)
- Startup success probability estimation
- Personal goal forecasting (weight loss, book-writing, career change)
- Investment outcome predictions
- Acquisition / merger outcome forecasting
- Team / company scaling projections
- Any forecast where 'we're different' feels true but might not be
Best AI model for this
Claude Sonnet 4.5 or Opus 4. Requires generating meaningful reference classes and base-rate reasoning. Mid-tier and above.
Pro tips
- The reference class is the work. Bad reference class = bad forecast regardless of analysis.
- Be ruthlessly specific about what's similar. 'Startups' is too broad. 'Seed-stage B2B SaaS in regulated industries with first-time founders' is usable.
- If you believe your project is the exception, identify ONE specific factor that makes it so, and check that factor's base rate separately.
- Inside view + outside view = best forecast. Don't dismiss either. Blend them 60/40 outside/inside for most forecasts.
- Your inside-view optimism is often right about MAGNITUDE (this will be big) and wrong about TIMING (this will be fast). Calibrate timing with outside view.
- If no reference class exists, you're doing something genuinely novel — but also forecasting with much less data. Adjust confidence wide.
Customization tips
- Write the reference class BEFORE estimating. Specificity in the class is where the base rate lives.
- If no reference class fits, you're either doing something truly novel (wide confidence) or defining the class too narrowly (broaden it).
- For any timeline forecast, apply the '2x rule' as a minimum: double your gut estimate. Research consistently shows this is more accurate than adding 20% buffer.
- Track your forecasts vs. actuals in a log. Over 10 forecasts, you'll see your personal bias (usually optimistic by 1.5-2x).
- When sharing forecasts with stakeholders, give the range, not a point. 'Q3-Q4' is more honest than 'September.'
Variants
Project Timeline Mode
For software / construction / product-launch timing forecasts. Heavy on the 'planning fallacy' base rate.
Startup Outcome Mode
For founder projections of outcomes. Reference-class building for your specific stage/sector.
Personal Goal Mode
For individual goals (fitness, creative project, career pivot). Handles the 'I'm different' instinct with personal base rates.
Frequently asked questions
How do I use the Outside View Forecaster prompt?
Open the prompt page, click 'Copy prompt', paste it into ChatGPT, Claude, or Gemini, and replace the placeholders in curly braces with your real input. The prompt is also launchable directly in each model with one click.
Which AI model works best with Outside View Forecaster?
Claude Sonnet 4.5 or Opus 4. Requires generating meaningful reference classes and base-rate reasoning. Mid-tier and above.
Can I customize the Outside View Forecaster prompt for my use case?
Yes — every Promptolis Original is designed to be customized. Key levers: The reference class is the work. Bad reference class = bad forecast regardless of analysis.; Be ruthlessly specific about what's similar. 'Startups' is too broad. 'Seed-stage B2B SaaS in regulated industries with first-time founders' is usable.
Explore more Originals
Hand-crafted 2026-grade prompts that actually change how you work.
← All Promptolis Originals