Back to skills

Assumption Challenger

Identify and challenge implicit assumptions in plans, proposals, and technical decisions. Use when strategic-cto-mentor needs to surface hidden assumptions and wishful thinking before they become costly mistakes.

32 stars
0 votes
0 copies
0 views
Added 12/19/2025
data-aigotestingapidatabasedevopsperformancedocumentation

Works with

api

Install via CLI

$openskills install alirezarezvani/claude-cto-team
Download Zip
Files
SKILL.md
---
name: assumption-challenger
description: Identify and challenge implicit assumptions in plans, proposals, and technical decisions. Use when strategic-cto-mentor needs to surface hidden assumptions and wishful thinking before they become costly mistakes.
---

# Assumption Challenger

Systematically identifies and stress-tests assumptions that are treated as facts but may not be validated.

## When to Use

- Validating roadmaps and project plans
- Reviewing architecture proposals
- Assessing build vs buy decisions
- Evaluating timelines and budgets
- Challenging strategic initiatives

## Why Assumptions Matter

Most project failures trace back to invalid assumptions:
- **Timeline assumptions**: "We can ship in 6 weeks" (based on nothing)
- **Resource assumptions**: "We'll hire 3 engineers" (in a tight market)
- **Technical assumptions**: "The API can handle 10K requests/sec" (never tested)
- **Business assumptions**: "Users will adopt this feature" (never validated)

**The cost of invalid assumptions compounds over time.** A bad assumption in week 1 can waste months of work.

---

## Assumption Categories

### 1. Timeline Assumptions
Assumptions about how long things will take.

**Common patterns**:
- "This should only take 2 weeks"
- "We'll have the integration done by launch"
- "The team can absorb this additional scope"

**Challenge questions**:
- What's this estimate based on? Past experience or hope?
- What similar work have we done? How long did it actually take?
- What's not included in this estimate? (Testing, documentation, deployment)
- What happens if this takes 2x longer?

### 2. Resource Assumptions
Assumptions about team capacity and availability.

**Common patterns**:
- "We'll hire 2 senior engineers by Q2"
- "The DevOps team can support this"
- "Sarah can lead this while maintaining her other work"

**Challenge questions**:
- What's the hiring timeline? What if we can't find the right people?
- What's the team's current utilization? Where does time come from?
- Who's the backup if the key person is unavailable?
- What happens if the team is 50% of expected capacity?

### 3. Technical Assumptions
Assumptions about system capabilities and constraints.

**Common patterns**:
- "The database can handle the load"
- "We can integrate with their API easily"
- "Our architecture supports this use case"

**Challenge questions**:
- Has this been tested at the required scale?
- What are the documented limits? What happens at those limits?
- What's the failure mode? How do we recover?
- Have we talked to the API provider about our usage patterns?

### 4. Business Assumptions
Assumptions about market, users, and business outcomes.

**Common patterns**:
- "Users want this feature"
- "This will reduce churn by 20%"
- "The market will wait for our solution"

**Challenge questions**:
- What evidence supports this? User research? Data?
- What if users don't adopt this? What's the fallback?
- What are competitors doing in this space?
- How will we know if this assumption is wrong?

### 5. External Assumptions
Assumptions about factors outside your control.

**Common patterns**:
- "The vendor will have the feature ready"
- "Regulations won't change"
- "The market will remain stable"

**Challenge questions**:
- What's our contingency if this doesn't happen?
- What's the vendor's track record on commitments?
- What early warning signs would indicate this is wrong?
- What's the cost of being wrong?

---

## Assumption Identification Process

### Step 1: Surface Assumptions

Read the proposal and identify statements that are treated as facts but aren't validated:

**Flag statements containing**:
- "We will..." (without evidence)
- "We can..." (without proof)
- "Users want..." (without data)
- "It should..." (without testing)
- "We expect..." (without basis)
- "We assume..." (at least they're honest)

### Step 2: Categorize Assumptions

For each assumption, categorize:

| Category | Risk if Wrong | Validation Difficulty |
|----------|---------------|----------------------|
| Timeline | Project delay | Medium (compare to past) |
| Resource | Execution failure | Medium (check market) |
| Technical | System failure | High (requires testing) |
| Business | Wasted investment | High (requires market validation) |
| External | Plans disrupted | Variable |

### Step 3: Assess Each Assumption

For each significant assumption:

```markdown
### Assumption: [Statement]

**Category**: [Type]
**Stated or Implicit**: [Was it stated or hidden?]

**Evidence Supporting**:
- [Evidence 1]
- [Evidence 2]

**Evidence Against**:
- [Counter-evidence 1]
- [Counter-evidence 2]

**Risk if Wrong**:
- [Impact on timeline]
- [Impact on cost]
- [Impact on success]

**Validation Method**:
- [How to test this assumption]
- [Cost/time to validate]

**Verdict**:
[ ] Valid - Evidence supports
[ ] Questionable - Needs validation
[ ] Invalid - Evidence contradicts
[ ] Unknown - Cannot assess
```

### Step 4: Prioritize Challenges

Focus on assumptions that are:
1. **High impact** - Project fails if wrong
2. **Low evidence** - Based on hope, not data
3. **Testable** - Can be validated before commitment

---

## Challenge Patterns

### The Reality Check
Compare assumption to external data.

**Template**:
> "You assume [X]. Industry data shows [Y]. What makes you different?"

**Example**:
> "You assume you can hire 3 senior engineers in 2 months. The average time-to-hire for senior engineers in your market is 4-6 months. What's your strategy to beat that?"

### The History Test
Compare to organization's past performance.

**Template**:
> "You assume [X]. Last time you attempted [similar thing], it took [Y]. What's changed?"

**Example**:
> "You assume 8 weeks for the microservices migration. Your last infrastructure migration took 5 months. What's different this time?"

### The Stress Test
Push assumption to failure point.

**Template**:
> "You assume [X]. What happens when [stress scenario]?"

**Example**:
> "You assume the system handles 10K concurrent users. What happens during a flash sale with 50K? What's the failure mode?"

### The Dependency Audit
Trace assumption to its dependencies.

**Template**:
> "For [assumption] to be true, what else must be true?"

**Example**:
> "For your 6-week timeline to work, you need: (1) API specs finalized by week 1, (2) no scope changes, (3) 100% team availability, (4) no production incidents. How realistic is that?"

### The Inverse Test
Consider what happens if assumption is wrong.

**Template**:
> "If [assumption] is wrong, what's the impact? What's your Plan B?"

**Example**:
> "If users don't adopt this feature, what's the fallback? Have you defined failure criteria?"

---

## Wishful Thinking Indicators

Red flags that suggest assumption is based on hope rather than evidence:

### 1. The Optimistic Timeline
- "Should only take..."
- "If everything goes well..."
- "We can do it if we're focused..."

**Reality**: Things rarely go perfectly. Add 30-50% buffer.

### 2. The Magical Hiring
- "We'll just hire..."
- "Once we have the team..."
- "We're planning to bring on..."

**Reality**: Hiring takes 3-6 months for senior roles. Onboarding adds another 2-3 months to productivity.

### 3. The Simple Integration
- "It's just an API call..."
- "Should be straightforward..."
- "They have good documentation..."

**Reality**: Integrations always have edge cases, rate limits, and unexpected behaviors.

### 4. The Obvious Market
- "Everyone needs this..."
- "Users have been asking for..."
- "It's clear that..."

**Reality**: "Everyone" is not a market segment. Validate with actual user research.

### 5. The Linear Scaling
- "If we can do X, we can do 10X..."
- "We'll scale as needed..."
- "Growth shouldn't be a problem..."

**Reality**: Scaling is non-linear. What works at 10K users may fail at 100K.

---

## Output Format

When challenging assumptions, provide:

```markdown
# Assumption Analysis: [Plan/Proposal Name]

## Summary
- **Total Assumptions Identified**: [Count]
- **High-Risk Assumptions**: [Count]
- **Requires Immediate Validation**: [Count]

## Critical Assumptions (Must Validate Before Proceeding)

### Assumption 1: [Statement]

**Category**: Timeline / Resource / Technical / Business / External
**Stated or Implicit**: Stated / Implicit

**The Problem**:
[Why this assumption is questionable]

**Evidence For**:
- [Supporting evidence]

**Evidence Against**:
- [Counter-evidence]

**If Wrong, Impact**:
- Timeline: [Impact]
- Budget: [Impact]
- Success: [Impact]

**How to Validate**:
- [Validation method]
- [Time required]
- [Cost]

**Verdict**: Valid / Questionable / Invalid / Unknown

---

### Assumption 2: [Statement]
[Same format]

---

## Medium-Risk Assumptions (Should Validate)
[List with brief analysis]

## Low-Risk Assumptions (Monitor)
[List]

## Recommendations

### Before Proceeding
1. [Validation action 1]
2. [Validation action 2]

### Risk Mitigation
1. [Mitigation for assumption 1]
2. [Mitigation for assumption 2]

### Contingency Plans Needed
1. [Plan B for assumption 1]
2. [Plan B for assumption 2]
```

---

## Integration with Validation

The assumption-challenger skill feeds into the broader validation workflow:

```
Proposal/Plan
     │
     ▼
[assumption-challenger] → List of assumptions with verdicts
     │
     ▼
[antipattern-detector] → Identified anti-patterns
     │
     ▼
[validation-report-generator] → Final 8-section report
```

---

## References

- [Wishful Thinking Patterns](wishful-thinking-patterns.md) - Common patterns to watch for
- [Challenge Questions](challenge-questions.md) - Ready-to-use challenge questions by category

Comments (0)

No comments yet. Be the first to comment!