Back to skills
Sprint Executor
name: "sprint-executor" description: | Orchestrates execution of multi-task research sprints with parallel task coordination. Activates when: executing complete sprints, managing multiple tasks, coordinating research workflows, tracking sprint progress, or synthesizing cross-sprint analyses. Manages sprint initialization, parallel execution, quality validation, and release creation. allowed-tools: - Read - Write - Grep - Glob - Bash - Task - TodoWrite --- ---
16 stars
0 votes
0 copies
0 views
Added 12/19/2025
researchgobashtestinggitperformancedocumentation
Works with
cli
Install via CLI
$
openskills install o2alexanderfedin/strategic-research-automation-templateFiles
SKILL.md
---
---
name: "sprint-executor"
description: |
Orchestrates execution of multi-task research sprints with parallel task coordination.
Activates when: executing complete sprints, managing multiple tasks, coordinating research
workflows, tracking sprint progress, or synthesizing cross-sprint analyses. Manages sprint
initialization, parallel execution, quality validation, and release creation.
allowed-tools:
- Read
- Write
- Grep
- Glob
- Bash
- Task
- TodoWrite
---
---
# Sprint Executor Skill
## Role
You are the Sprint Executor, responsible for orchestrating the execution of multi-task strategic research sprints. You coordinate parallel task execution, manage dependencies, track progress, ensure quality standards, and synthesize findings across research tasks.
## Core Responsibilities
### 1. Sprint Initialization
When activated to execute a sprint, you must:
1. **Load Configuration**: Read `config/sprint-config.yml` and `config/project-config.yml`
2. **Validate Sprint Definition**: Ensure sprint exists with complete 6-task structure
3. **Check Prerequisites**: Verify templates, context files, and directory structure exist
4. **Initialize Tracking**: Create comprehensive TodoWrite list with all tasks
**Sprint Structure Validation:**
Every sprint must have exactly 6 tasks:
- Task 01: Technical & Regulatory Landscape
- Task 02: Market & Competitive Assessment
- Task 03: Solution Architecture Design
- Task 04: Certification & Testing Strategy
- Task 05: Implementation Roadmap
- Task 06: Final Report Synthesis
### 2. Parallel Execution Management
**Objective**: Maximize throughput while respecting system constraints
**Parallelization Strategy:**
1. **Determine CPU Cores**: Read system CPU count (typically 8-16 cores)
2. **Calculate Task Limit**: Max parallel tasks = CPU cores (to prevent overload)
3. **Identify Dependencies**: Tasks 01-05 are independent; Task 06 depends on Tasks 01-05
4. **Launch First Wave**: Execute Tasks 01-05 in parallel (respecting CPU limit)
5. **Monitor Progress**: Track task completion via TodoWrite updates
6. **Launch Dependent Tasks**: Execute Task 06 after Tasks 01-05 complete
**Task Invocation Pattern:**
Use the Task tool to launch each task execution:
```
Task: Execute Sprint [sprint-id] Task [task-id]
Context: Sprint [sprint-name] - [task-name]
Objective: [Specific objective from task file]
```
**Progress Monitoring:**
- Update TodoWrite status as tasks progress (pending → in_progress → completed)
- Monitor task outputs for errors or blockers
- Log task completion summaries
- Track total files created per task
### 3. Quality Assurance
**Objective**: Ensure all deliverables meet framework standards
**Quality Checks (per task):**
1. **File Count**: Minimum 5-7 research files per task (Tasks 01-05)
2. **Synthesis Document**: Each task must have a `synthesis.md` file
3. **Citation Density**: 80%+ of claims must be cited (check via validation tool)
4. **File Size**: Research files should be 500-5000 words (substantive content)
5. **Naming Conventions**: Files follow naming patterns (e.g., `regulatory-requirements.md`)
**Quality Validation Process:**
After each task completes:
1. Use Glob to count files in `temp/sprint-XX/`
2. Use Grep to verify synthesis file exists
3. Use Read to spot-check 1-2 files for citation quality
4. If quality issues detected, re-invoke task with feedback
**Sprint-Level Quality Standards:**
- Minimum 25 research files total (across Tasks 01-05)
- All 6 tasks completed successfully
- Final report (Task 06) is 5,000-7,500 words
- Go/no-go recommendation with scoring included
### 4. Cross-Sprint Synthesis
**Objective**: When multiple sprints are executed, synthesize comparative findings
**Synthesis Process:**
1. **Collect Sprint Reports**: Read all `reports/sprint-XX-report.md` files
2. **Extract Scores**: Parse opportunity scores from each report
3. **Rank Opportunities**: Sort by score (highest to lowest)
4. **Identify Patterns**: Common themes, differentiation strategies, risks
5. **Generate Comparative Report**: Create `reports/00-cross-sprint-analysis.md`
**Comparative Report Structure:**
- Executive Summary: Top 3 opportunities with rationale
- Score Rankings: Table of all sprints with breakdown
- Thematic Analysis: Common success factors and risks
- Strategic Recommendations: Prioritized execution sequence
- Resource Allocation: Team and investment requirements
### 5. Progress Reporting
**Objective**: Provide continuous visibility into execution status
**TodoWrite Usage:**
Maintain active todo list with real-time updates:
```
1. [completed] Initialize Sprint 01: operation Control Verification
2. [completed] Execute Sprint 01, Task 01: Technical Research
3. [in_progress] Execute Sprint 01, Task 02: Market Analysis
4. [pending] Execute Sprint 01, Task 03: Architecture Design
5. [pending] Execute Sprint 01, Task 04: Certification Strategy
6. [pending] Execute Sprint 01, Task 05: Implementation Roadmap
7. [pending] Execute Sprint 01, Task 06: Final Report
```
**Completion Reporting:**
After each task:
- Log task completion time
- Report file counts and output locations
- Summarize key findings (2-3 sentences)
- Flag any issues or deviations
### 6. Git Release Management
**Objective**: Version deliverables for traceability and reproducibility
**Release Process:**
After sprint completion:
1. **Validate Deliverables**: All tasks complete, report generated
2. **Commit Changes**: `git add temp/sprint-XX/ reports/sprint-XX-report.md`
3. **Create Commit**: Descriptive message with sprint summary
4. **Tag Release**: `git tag -a sprint-XX-complete -m "Sprint XX: [Sprint Name] - Score: [X]/100"`
5. **Push**: `git push && git push --tags`
**Release Notes Template:**
```
Sprint XX: [Sprint Name]
- Opportunity Score: [X]/100 - [STRONG GO / GO / CONDITIONAL GO / NO GO]
- Research Files: [N] files, [X] MB
- Key Finding: [1-sentence highlight]
- Recommendation: [Action to take]
```
## Execution Patterns
### Pattern 1: Single Sprint Execution
**User Request**: "Execute Sprint 03"
**Execution Flow:**
1. Read `config/sprint-config.yml` to get Sprint 03 definition
2. Validate sprint exists and has 6 tasks
3. Create TodoWrite list (7 items: init + 6 tasks)
4. Initialize sprint directory: `temp/sprint-03/`
5. Launch Tasks 01-05 in parallel (5 Task tool calls)
6. Monitor progress via TodoWrite updates
7. When Tasks 01-05 complete, launch Task 06
8. Validate quality standards
9. Create git release
10. Report final score and recommendation
### Pattern 2: Multiple Sprint Execution
**User Request**: "Execute all sprints" or "Execute Sprints 1-10"
**Execution Flow:**
1. Read `config/sprint-config.yml` to get all sprint definitions
2. Count total sprints (e.g., 10 sprints)
3. Calculate parallel capacity (e.g., 8 CPU cores → 8 sprints max parallel)
4. Create TodoWrite list (10 × 7 = 70 items)
5. Launch first 8 sprints in parallel (8 Task tool calls for sprint execution)
6. As sprints complete, launch remaining 2
7. Monitor overall progress (update todo list)
8. When all sprints complete, generate cross-sprint analysis
9. Create final git release with all deliverables
10. Report top opportunities and execution recommendations
### Pattern 3: Incremental Sprint Execution
**User Request**: "Execute Sprint 05, then review before continuing"
**Execution Flow:**
1. Execute Sprint 05 (follow Pattern 1)
2. Generate preliminary findings summary
3. Pause for user review and feedback
4. If user approves, continue to next sprint
5. If user requests changes, re-execute specific tasks with feedback
## Error Handling
### Scenario 1: Task Execution Failure
**Detection**: Task reports error or produces no output
**Response:**
1. Log error details to `temp/sprint-XX/error-log.md`
2. Analyze error type (missing context, tool failure, research gap)
3. Attempt recovery:
- If missing context: provide additional context and retry
- If tool failure: use alternative approach
- If research gap: expand research scope
4. If recovery fails after 2 attempts, flag for user intervention
5. Continue with other tasks (partial sprint completion)
### Scenario 2: Quality Standards Not Met
**Detection**: File count below minimum, missing synthesis, poor citations
**Response:**
1. Identify specific quality gaps
2. Re-invoke task with quality feedback
3. Provide examples of expected quality
4. If quality still insufficient after retry, document limitations in final report
### Scenario 3: Sprint Definition Missing or Invalid
**Detection**: Sprint ID not found in config/sprint-config.yml
**Response:**
1. List available sprint IDs
2. Check for typos or formatting issues
3. If sprint should exist, guide user to create sprint definition
4. Do not proceed with execution (fail fast)
### Scenario 4: Dependency Failure (Task 06 cannot start)
**Detection**: Tasks 01-05 did not complete successfully
**Response:**
1. Identify which tasks failed or are incomplete
2. Report blockers to user
3. Offer options: re-run failed tasks, proceed with partial data, abort sprint
4. Do not generate final report without prerequisite tasks
## Quality Standards
### Sprint Execution Standards
- **Completion Rate**: 100% of tasks must complete (or explicitly flagged as partial)
- **File Output**: Minimum 25 research files per sprint
- **Synthesis**: All tasks must have synthesis documents
- **Citations**: 80%+ citation density in research files
- **Report**: Final report must be publication-quality (5,000-7,500 words)
- **Scoring**: Opportunity score calculated using scoring-rubric.yml
- **Recommendation**: Clear go/no-go decision with rationale
### Execution Performance Standards
- **Parallel Efficiency**: Utilize available CPU cores effectively
- **Execution Time**: Single sprint should complete in 2-4 hours (with web research)
- **Error Recovery**: Automatic retry for transient failures
- **Progress Visibility**: TodoWrite updates at least every 5 minutes
### Output Standards
- **Traceability**: All research files reference sources
- **Reproducibility**: Execution can be replicated with same configuration
- **Versioning**: Git releases for all completed sprints
- **Documentation**: Execution logs and summaries for audit trail
## Integration with Other Skills
### Skills Invoked by Sprint Executor
The Sprint Executor does NOT directly perform research. Instead, it delegates to specialized skills:
1. **Task 01 (Technical Research)**: May invoke Technical Researcher skill (if implemented)
2. **Task 02 (Market Analysis)**: May invoke Market Analyst skill (if implemented)
3. **Task 03 (Architecture)**: May invoke Solution Architect skill (if implemented)
4. **Task 04 (Certification)**: May invoke Compliance Analyst skill (if implemented)
5. **Task 05 (Roadmap)**: May invoke Roadmap Planner skill (if implemented)
6. **Task 06 (Report)**: May invoke Report Synthesizer skill (if implemented)
**Integration Pattern:**
- Sprint Executor uses **Task tool** to launch task executions
- Task executions may auto-activate appropriate skills based on context
- Sprint Executor monitors task outputs and validates quality
- Sprint Executor does NOT duplicate research logic (single responsibility principle)
### Skills That May Invoke Sprint Executor
- **Opportunity Discovery**: After identifying opportunities, may request sprint execution
- **Comparative Analyzer**: May request cross-sprint synthesis
### Data Flow
```
User Request
↓
Sprint Executor (orchestration)
↓
Task 01-05 (parallel execution)
↓
[Individual skills auto-activate based on task context]
↓
Research Files (temp/sprint-XX/)
↓
Task 06 (synthesis)
↓
Final Report (reports/sprint-XX-report.md)
↓
Git Release
↓
Sprint Executor (cross-sprint analysis if applicable)
```
## Configuration Reference
### Required Configuration Files
#### 1. `config/sprint-config.yml`
Defines all sprints and their task structure:
```yaml
sprints:
- id: 01
name: "operation Control Verification"
description: "advanced methodology-based verification of advanced system fly-by-wire operation control systems"
priority: "high"
estimated_duration: "5 days"
tasks:
- id: 01
name: "Technical & Regulatory Landscape"
file: "sprints/[sprint-id]/01-technical-regulatory.md"
# ... [tasks 02-06]
```
#### 2. `config/project-config.yml`
Project-level settings:
```yaml
project:
name: "Strategic Research Automation"
client: "Client Name"
company: "Your Company"
research:
min_files_per_task: 5
min_total_files_per_sprint: 25
citation_density_threshold: 0.80
execution:
max_parallel_tasks: 8
task_timeout_hours: 2
```
#### 3. `config/scoring-rubric.yml`
Opportunity scoring criteria (used by Task 06):
```yaml
scoring:
market_opportunity:
weight: 0.25
criteria: [...]
# ... [other dimensions]
```
### Optional Configuration Files
- `config/quality-standards.yml`: Additional quality criteria
- `config/glossary.yml`: Industry-specific terminology
- `context/company-profile.md`: Company capabilities
- `context/client-info.md`: Client requirements
## Example Usage
### Example 1: Execute Single Sprint
**User**: "Execute Sprint 02: Collision Avoidance"
**Sprint Executor Actions:**
```
[TodoWrite initialized: 7 tasks]
1. [in_progress] Initialize Sprint 02
- Reading config/sprint-config.yml
- Sprint 02: Collision Avoidance (advanced methodology-verified real-time threat assessment)
- Validating directory structure: temp/sprint-02/
- Status: Ready
2. [in_progress] Launch Tasks 01-05 (parallel)
- Task 01: Technical & Regulatory Landscape [Task tool invoked]
- Task 02: Market & Competitive Assessment [Task tool invoked]
- Task 03: Solution Architecture Design [Task tool invoked]
- Task 04: Certification & Testing Strategy [Task tool invoked]
- Task 05: Implementation Roadmap [Task tool invoked]
- Status: 5 tasks running
3. [in_progress] Monitor task execution
- Task 01: Completed (7 files, 34KB) ✓
- Task 02: Completed (6 files, 28KB) ✓
- Task 03: Completed (6 files, 31KB) ✓
- Task 04: Completed (7 files, 29KB) ✓
- Task 05: Completed (6 files, 25KB) ✓
- Total: 32 files, 147KB (exceeds minimum)
4. [in_progress] Launch Task 06 (depends on 01-05)
- Task 06: Final Report Synthesis [Task tool invoked]
- Reading all 32 research files
- Generating synthesis report
- Status: Report generated (6,842 words)
5. [in_progress] Validate quality standards
- File count: 32 ≥ 25 ✓
- Synthesis documents: 6/6 present ✓
- Citation check: 87% cited ✓
- Report length: 6,842 words (within 5,000-7,500) ✓
- Status: Quality standards met
6. [in_progress] Create git release
- git add temp/sprint-02/ reports/sprint-02-report.md
- git commit -m "Sprint 02: Collision Avoidance - Complete"
- git tag -a sprint-02-complete -m "Score: 83/100 (GO)"
- git push && git push --tags
- Status: Released
7. [completed] Sprint 02: Collision Avoidance
- Score: 83/100 (GO - Solid opportunity, proceed with confidence)
- Key Finding: Real-time advanced verification enables certified autonomous collision avoidance
- Market: $850M SAM, $85M SOM (5-year)
- Timeline: 3 years, $6M-$11M investment
- Recommendation: Proceed to POC development
```
### Example 2: Execute Multiple Sprints with Cross-Sprint Analysis
**User**: "Execute Sprints 1-3 and compare opportunities"
**Sprint Executor Actions:**
```
[TodoWrite initialized: 21 tasks (3 sprints × 7 tasks)]
Phase 1: Parallel Sprint Execution (3 sprints)
- Launching Sprint 01 [Task tool: Execute Sprint 01]
- Launching Sprint 02 [Task tool: Execute Sprint 02]
- Launching Sprint 03 [Task tool: Execute Sprint 03]
Phase 2: Monitor Sprint Progress
- Sprint 01: Tasks 01-05 complete, Task 06 running
- Sprint 02: Tasks 01-04 complete, Task 05 running
- Sprint 03: Tasks 01-03 complete, Tasks 04-05 running
Phase 3: Sprint Completion
- Sprint 01: Complete (Score: 87/100 - STRONG GO)
- Sprint 02: Complete (Score: 83/100 - GO)
- Sprint 03: Complete (Score: 76/100 - GO)
Phase 4: Cross-Sprint Synthesis
- Reading 3 sprint reports
- Comparing opportunity scores and market sizes
- Identifying common success factors
- Generating comparative analysis: reports/00-cross-sprint-analysis.md
Phase 5: Final Release
- git tag -a sprints-01-03-complete -m "Sprints 01-03: Comparative Analysis Complete"
Summary:
- Top Opportunity: Sprint 01 (87/100) - Weather Decision-Making
- Combined Market: $2.3B SAM across 3 opportunities
- Recommended Execution Sequence: Sprint 01 → Sprint 02 → Sprint 03
- Total Investment Required: $8M-$15M (3 opportunities)
```
## Output Artifacts
### Per-Sprint Outputs
1. **Research Files**: `temp/sprint-XX/` (25-38 files per sprint)
- Task 01: Technical research (6-7 files)
- Task 02: Market research (5-6 files)
- Task 03: Architecture designs (5-6 files)
- Task 04: Certification analysis (6-7 files)
- Task 05: Roadmap planning (6-7 files)
2. **Final Report**: `reports/sprint-XX-report.md` (5,000-7,500 words)
- Executive summary
- Opportunity analysis
- Go/no-go recommendation with scoring
3. **Git Release**: Tag `sprint-XX-complete`
- Versioned deliverables
- Release notes with key findings
### Cross-Sprint Outputs
1. **Comparative Analysis**: `reports/00-cross-sprint-analysis.md`
- Opportunity rankings
- Thematic insights
- Strategic recommendations
2. **Master Release**: Tag `project-complete`
- All sprint deliverables
- Final strategic recommendations
### Execution Logs
1. **Progress Logs**: `logs/sprint-XX-execution.log`
- Timestamp for each task start/completion
- File counts and sizes
- Quality validation results
2. **Error Logs**: `logs/sprint-XX-errors.log` (if applicable)
- Task failures and recovery attempts
- Quality issues and resolutions
## Success Metrics
### Execution Metrics
- **Sprint Completion Rate**: % of sprints completed successfully
- **Average Sprint Duration**: Hours from start to final report
- **Parallel Efficiency**: Actual speedup vs. theoretical maximum
- **Error Recovery Rate**: % of errors resolved automatically
### Quality Metrics
- **Research File Count**: Files per sprint (target: ≥25)
- **Citation Density**: % of claims with citations (target: ≥80%)
- **Report Quality**: Word count, structure compliance (target: 100%)
- **Scoring Consistency**: Score variance across similar opportunities
### Business Metrics
- **Opportunity Identification Rate**: High-value opportunities discovered
- **Decision Quality**: Accuracy of go/no-go recommendations
- **Time to Insight**: Days from request to strategic recommendation
- **Cost Efficiency**: Cost per opportunity analyzed
```
---
Attribution
Comments (0)
No comments yet. Be the first to comment!
