C-suite executive decision support: strategy, technology, growth, competitive intelligence, project evaluation.
Umbrella skill for all executive decision-making: CEO-level strategy, CTO-level technology choices, CMO-level growth planning, competitive intelligence, and project evaluation. Each domain loads its own reference files on demand -- this skill detects the mode, loads the right references, and executes the appropriate framework.
Scope: Business decisions with meaningful consequences. Do NOT use for: technical architecture micro-choices (use decision-helper), writing code (use domain agents), writing content (use voice-writer), or debugging (use systematic-debugging).
Classify the user's request into exactly one mode before proceeding. If the request spans multiple modes, choose the primary one and note the secondary.
| Mode | Signal Phrases | Role Lens |
|---|---|---|
| STRATEGY | Market entry, partnerships, resource allocation, opportunity, "should I/we", strategic pivots, investment | CEO |
| TECHNOLOGY |
| Build vs buy, vendor, SaaS, tech stack, architecture, adopt, technology choice |
| CTO |
| GROWTH | Content strategy, audience, SEO, marketing, brand, community, positioning, channel | CMO |
| COMPETITIVE | Competitor, competition, market landscape, differentiation, positioning against, market share | Cross-role |
| EVALUATION | Feasibility, effort estimate, ROI, priority, go/no-go, viability, "is it worth it" | Cross-role |
Load references based on the detected mode. Do NOT load all references -- load only what the mode requires.
| Signal | Mode | Reference |
|---|---|---|
| Market entry, partnerships, resource allocation, opportunity | STRATEGY | references/strategic-frameworks.md, references/decision-matrices.md |
| Build vs buy, vendor, SaaS, tech stack, architecture | TECHNOLOGY | references/tco-framework.md, references/vendor-evaluation.md |
| Content, audience, SEO, marketing, brand, community | GROWTH | references/audience-segmentation.md, references/channel-evaluation.md |
| Competitor, market landscape, positioning, differentiation | COMPETITIVE | references/competitive-mapping.md, references/market-positioning.md |
| Feasibility, effort, ROI, priority, go/no-go | EVALUATION | references/feasibility-scoring.md, references/roi-frameworks.md |
Framework: FRAME -> ANALYZE -> DECIDE
Phase 1: FRAME -- Convert the user's question into a structured decision with clear stakes and timeline.
Gate: Decision framed as one sentence. Options listed (2-4). Type classified.
Phase 2: ANALYZE -- Evaluate each option through multiple lenses with evidence.
For each option, assess: Upside (best realistic + expected outcome), Downside (worst realistic + recovery path + irreversible losses), Requirements (resources, assumptions, dependencies), Opportunity Cost (what you cannot do).
Separate facts from assumptions. Quantify where possible. Load reference files for scoring matrices and strategic frameworks.
Gate: All options analyzed. Facts and assumptions labeled. Opportunity costs explicit.
Phase 3: DECIDE -- Synthesize into a clear recommendation.
Gate: Recommendation stated. First action identified. Revisit trigger set.
Framework: SCOPE -> EVALUATE -> RECOMMEND
Phase 1: SCOPE -- Define the capability needed, stripped of solution bias.
Gate: Capability defined without solution bias. Options enumerated. Hard requirements quantified.
Phase 2: EVALUATE -- Score options on dimensions that matter for technology decisions.
Load references/tco-framework.md for TCO templates and references/vendor-evaluation.md for vendor scorecards.
Gate: TCO estimated. Dimensions scored. Build-vs-buy heuristic applied.
Phase 3: RECOMMEND -- Deliver a clear recommendation with reasoning.
Gate: Recommendation stated. Exit criteria defined. First step identified.
Framework: ASSESS -> STRATEGIZE -> PLAN
Phase 1: ASSESS -- Understand current state before recommending.
Gate: Current state audited. Binding constraint identified.
Phase 2: STRATEGIZE -- Design an approach matching capacity and constraint.
Load references/audience-segmentation.md for ICP scoring and references/channel-evaluation.md for channel matrices.
Gate: Strategy selected. Maximum 3 channels. Effort estimated against capacity.
Phase 3: PLAN -- Convert strategy into a 90-day executable plan.
Gate: 90-day plan with checkpoints. Primary metric defined. Abandon criteria explicit.
Framework: MAP -> ANALYZE -> POSITION
Phase 1: MAP -- Build a structured picture of the competitive landscape.
Gate: Arena defined. Competitors identified and tiered. At least 2 direct competitors mapped.
Phase 2: ANALYZE -- Extract actionable intelligence from behavior, not surface impressions.
Load references/competitive-mapping.md for landscape templates and references/market-positioning.md for positioning frameworks.
Gate: Direct competitors analyzed. Gaps and weaknesses identified.
Phase 3: POSITION -- Convert intelligence into defensible differentiation.
Gate: Positioning map built. Differentiation strategy defined. Monitoring cadence set.
Framework: SCOPE -> EVALUATE -> VERDICT
Phase 1: SCOPE -- Define the project and what success looks like.
Gate: Project defined with measurable success criteria. MVP scope identified. Binding constraint named.
Phase 2: EVALUATE -- Assess feasibility, estimate effort, calculate ROI.
Load references/feasibility-scoring.md for the three-dimension model and references/roi-frameworks.md for estimation templates.
Gate: Feasibility assessed. Effort estimated in ranges. ROI calculated with confidence level.
Phase 3: VERDICT -- Deliver a clear go/no-go recommendation.
Gate: Verdict stated with confidence. Conditions specified. Next step identified.
| Error | Cause | Solution |
|---|---|---|
| Too many options | 5+ options creating paralysis | Eliminate obviously inferior options first. Get to 2-4 before running full framework. |
| Not enough information | User cannot answer framing questions | Identify 2-3 critical unknowns. Recommend time-boxed research sprint before deciding. |
| Analysis paralysis | Keeps adding criteria or second-guessing | Apply reversibility test. If reversible, recommend best current option with checkpoint. |
| Emotional attachment | User has already decided, wants validation | Name the pattern directly. Ask: stress-test the choice, or genuinely evaluate all options? |
| Comparing apples to oranges | Options at different abstraction levels | Normalize to the capability level. Compare what each option gives for the specific need. |
| Vendor lock-in fear | Over-weights lock-in, under-weights time-to-value | Quantify actual switching cost. Compare concrete switching cost against concrete speed benefit. |
| Build bias (NIH) | Team wants to build because it is more interesting | Apply core competency test: "If this disappeared, would customers notice?" |
| Vanity metrics | Optimizes followers/likes instead of outcomes | Redirect to "one metric that matters" -- what action should the audience take? |
| Scope creep during evaluation | Keeps adding features to project definition | Freeze scope at end of Phase 1. Additional features evaluate as v2. |
| Optimism bias | Effort estimates too low | Apply reference class test. If no similar project, add 50% to pessimistic estimate. |
| Reference | When to Load | Content |
|---|---|---|
references/strategic-frameworks.md | STRATEGY mode: market entry, competitive dynamics, SWOT, OKR alignment | Porter's Five Forces, SWOT scoring, OKR alignment matrices |
references/decision-matrices.md | STRATEGY mode: structured scoring, comparison, pre-mortem | Weighted decision matrices, ICE/RICE scoring, pre-mortem templates |
references/tco-framework.md | TECHNOLOGY mode: TCO modeling, cost projections, build vs buy scorecard | TCO templates, hidden cost checklists, migration cost models |
references/vendor-evaluation.md | TECHNOLOGY mode: vendor comparison, RFP criteria, integration complexity | Vendor scorecards, RFP criteria, red flag detection, contract checklist |
references/audience-segmentation.md | GROWTH mode: audience analysis, ICP definition, persona development | ICP scoring matrix, persona templates, segmentation frameworks |
references/channel-evaluation.md | GROWTH mode: channel selection, CAC/LTV modeling, content funnel | Channel scoring matrices, CAC/LTV models, funnel stage mapping |
references/competitive-mapping.md | COMPETITIVE mode: landscape mapping, feature comparison, competitor profiling | Landscape map templates, feature matrices, activity tracker |
references/market-positioning.md | COMPETITIVE mode: positioning strategy, differentiation scoring | Positioning maps, differentiation scoring, win/loss frameworks |
references/feasibility-scoring.md | EVALUATION mode: feasibility assessment, risk evaluation, go/no-go | Three-dimension feasibility model, confidence calibration, decision tree |
references/roi-frameworks.md | EVALUATION mode: effort estimation, ROI calculation, project comparison | T-shirt sizing, three-point estimation, risk-adjusted NPV |