AI Cold Outreach workflow skill. Use this skill when the user needs When the user wants to build an AI-powered outreach system, write cold emails, improve deliverability, or scale personalized outreach. Also use when the user mentions 'cold email,' 'cold outreach,' 'outreach automation,' 'Instantly,' 'Smartlead,' 'Clay,' 'email sequences,' 'deliverability,' 'personalization at scale,' 'reply rate,' or 'outreach stack.' This skill covers the complete AI cold outreach system from signal detection through conversion. Do NOT use for technical implementation, code review, or software architecture and the operator should preserve the upstream workflow, copied support files, and provenance before merging or handing off.
This public intake copy packages packages/skills-catalog/skills/(gtm)/ai-cold-outreach from https://github.com/tech-leads-club/agent-skills into the native Omni Skills editorial shape without hiding its origin.
Use it when the operator needs the upstream workflow, support files, and repository context to stay intact while the public validator and private enhancer continue their normal downstream flow.
This intake keeps the copied upstream files intact and uses EXTERNAL_SOURCE.json plus ORIGIN.md as the provenance anchor for review.
Imported source sections that did not map cleanly to the public headings are still preserved below or in the support files. Notable imported sections: Before Starting, The AI Outreach Stack, The 3-Line Cold Email Framework.
Use this section as the trigger filter. It should make the activation boundary explicit before the operator loads files, runs commands, or opens a pull request.
| Situation | Start here | Why it matters |
|---|---|---|
| First-time use | EXTERNAL_SOURCE.json | Confirms repository, branch, commit, and imported path before touching the copied workflow |
| Provenance review | ORIGIN.md | Gives reviewers a plain-language audit trail for the imported source |
| Workflow execution | references/benchmarks-deliverability-tactics.md | Starts with the smallest copied file that materially changes execution |
| Supporting context | references/quick-reference.md | Adds the next most relevant copied source file without loading the entire package |
| Handoff decision | ## Related Skills | Helps the operator switch to a stronger native skill when the task drifts |
This workflow is intentionally editorial and operational at the same time. It keeps the imported source useful to the operator while still satisfying the public intake standards that feed the downstream enhancer flow.
Before building or optimizing any cold outreach system, gather:
If the user skips these, ask. Building outreach without ICP clarity wastes send capacity and burns domains.
Use @ai-cold-outreach to handle <task>. Start from the copied upstream workflow, load only the files that change the outcome, and keep provenance visible in the answer.
Explanation: This is the safest starting point when the operator needs the imported workflow, but not the entire repository.
Review @ai-cold-outreach against EXTERNAL_SOURCE.json and ORIGIN.md, then explain which copied upstream files you would load first and why.
Explanation: Use this before review or troubleshooting when you need a precise, auditable explanation of origin and file selection.
Use @ai-cold-outreach for <task>. Load only the copied references, examples, or scripts that change the outcome, and name the files explicitly before proceeding.
Explanation: This keeps the skill aligned with progressive disclosure instead of loading the whole copied package by default.
Review @ai-cold-outreach using the copied upstream files plus provenance, then summarize any gaps before merge.
Explanation: This is useful when the PR is waiting for human review and you want a repeatable audit packet.
Treat the generated public skill as a reviewable packaging layer around the upstream repository. The goal is to keep provenance explicit and load only the copied source material that materially improves execution.
Symptoms: The result ignores the upstream workflow in packages/skills-catalog/skills/(gtm)/ai-cold-outreach, fails to mention provenance, or does not use any copied source files at all.
Solution: Re-open EXTERNAL_SOURCE.json, ORIGIN.md, and the most relevant copied upstream files. Load only the files that materially change the answer, then restate the provenance before continuing.
Symptoms: Reviewers can see the generated SKILL.md, but they cannot quickly tell which references, examples, or scripts matter for the current task.
Solution: Point at the exact copied references, examples, scripts, or assets that justify the path you took. If the gap is still real, record it in the PR instead of hiding it.
Symptoms: The imported skill starts in the right place, but the work turns into debugging, architecture, design, security, or release orchestration that a native skill handles better. Solution: Use the related skills section to hand off deliberately. Keep the imported provenance visible so the next skill inherits the right context instead of starting blind.
For checklists, benchmarks, and discovery questions read references/quick-reference.md when you need detailed reference.
@accessibility - Use when the work is better handled by that native specialization after this imported skill establishes context.@ai-pricing - Use when the work is better handled by that native specialization after this imported skill establishes context.@ai-sdr - Use when the work is better handled by that native specialization after this imported skill establishes context.@ai-seo - Use when the work is better handled by that native specialization after this imported skill establishes context.Use this support matrix and the linked files below as the operator packet for this imported skill. They should reflect real copied source material, not generic scaffolding.
| Resource family | What it gives the reviewer | Example path |
|---|---|---|
references | copied reference notes, guides, or background material from upstream | references/benchmarks-deliverability-tactics.md |
examples | worked examples or reusable prompts copied from upstream | examples/n/a |
scripts | upstream helper scripts that change execution or validation | scripts/n/a |
agents | routing or delegation notes that are genuinely part of the imported package | agents/n/a |
assets | supporting assets or schemas copied from the source package | assets/n/a |
The modern cold outreach system is a six-stage pipeline. Each stage has specific tools, metrics, and failure modes.
+------------------+ +------------------+ +---------------------+
| 1. SIGNAL |---->| 2. ENRICHMENT |---->| 3. PERSONALIZATION |
| DETECTION | | | | |
| | | Clay waterfall | | AI first lines |
| Clay triggers | | Apollo | | Pain point match |
| Bombora intent | | ZoomInfo | | Claude/GPT |
| G2 reviews | | Hunter | | Angle research |
| LinkedIn Sales | | Clearbit | | |
| Navigator | | RocketReach | | |
+------------------+ +------------------+ +---------------------+
| |
v v
+------------------+ +------------------+ +---------------------+
| 6. FOLLOW-UP |<----| 5. SENDING |<----| 4. SEQUENCING |
| | | | | |
| AI contextual | | Instantly | | Multi-step |
| replies | | Smartlead | | Conditional logic |
| Objection | | Multi-mailbox | | A/B variants |
| handling | | rotation | | Channel mixing |
| Meeting booking | | IP sharding | | Timing rules |
+------------------+ +------------------+ +---------------------+
Signals tell you WHO to reach out to and WHEN. Cold email without signals is spam with extra steps.
Signal types ranked by conversion intent:
| Signal Type | Source | Intent Level | Timing Window |
|---|---|---|---|
| Category page view on G2 | G2 Buyer Intent | Very High | 7-14 days |
| Competitor evaluation | Bombora + G2 | Very High | 7-21 days |
| Job posting for your category | LinkedIn, Indeed | High | 14-30 days |
| Funding announcement | Crunchbase, Clay | High | 30-60 days |
| Tech stack change | BuiltWith, HG Data | Medium-High | 14-30 days |
| Leadership hire | LinkedIn Sales Nav | Medium | 30-45 days |
| Content engagement | Bombora cooperative | Medium | 7-14 days |
| Company growth spike | Clay, LinkedIn | Medium-Low | 30-60 days |
Signal layering strategy: Single signals produce 3-5% reply rates. Layer two or more signals and reply rates jump to 8-15%. Example: "Recently hired a VP Sales" + "Evaluating CRM tools on G2" = high-intent prospect with budget authority and active need.
Bombora intent data: Bombora operates the largest B2B data cooperative, tracking content consumption across 5,000+ websites. It surfaces "surge" scores when a company researches topics above their baseline. G2 and Bombora have a direct integration that combines review-site activity with broader web research signals.
Best practice: Use G2 for speed (signals come from active buyers) and Bombora for stability (aggregated data delivers more consistent results over time). Layer both for full coverage.
Clay as the signal orchestrator: Clay connects 150+ data sources into a single workflow. Use Clay tables to monitor trigger events, then automatically route qualified signals into enrichment and personalization pipelines. Clay's HTTP request action lets you connect any API as a signal source.
Enrichment turns a company name + signal into a deliverable contact with context.
The waterfall enrichment model:
Lead enters Clay table
|
v
[Provider A: Apollo]
Found email? ----YES----> Verified? --YES--> Done
| |
NO NO
| |
v v
[Provider B: Hunter] [Provider C: ZoomInfo]
Found email? ----YES----> Verified? --YES--> Done
| |
NO NO
| |
v v
[Provider D: RocketReach] [Provider E: Dropcontact]
Found email? ----YES----> Verified? --YES--> Done
|
NO
|
v
Skip or manual research
Why waterfall beats single-provider: No single provider covers more than 60-70% of B2B contacts. Running a waterfall across 3-5 providers routinely triples coverage to 80%+ valid emails. Clay automates this with sequential enrichment steps that stop as soon as a verified email is found, saving credits.
Enrichment data to collect (in priority order):
Email verification is non-negotiable: Run every email through verification (ZeroBounce, NeverBounce, or MillionVerifier) before sending. A bounce rate above 2% triggers spam filters at Google and Microsoft. One bad list can burn a domain in a day.
Generic cold emails get 1-2% reply rates. AI-personalized emails get 8-12%. The difference is the first two lines.
The AI personalization pipeline:
Enriched lead data (company news, tech stack, hiring, social)
|
v
[AI Agent: Claude or GPT]
|
+---> Research summary (2-3 key findings)
+---> Personalization angle (why NOW, why THEM)
+---> Custom first line (specific observation)
+---> Pain hypothesis (inferred from signals)
|
v
Merge into email template via {{variables}}
First line frameworks that work:
| Framework | Example | Best For |
|---|---|---|
| Observation + Implication | "Saw you just opened a London office - scaling support across time zones gets messy fast." | Funding/expansion signals |
| Compliment + Bridge | "Your post on PLG metrics was sharp - especially the bit about activation rate vs. NPS." | Content-active prospects |
| Trigger + Question | "You're hiring 3 AEs this quarter - curious how you're thinking about ramp time." | Hiring signals |
| Mutual Connection | "Alex Chen mentioned your team is rethinking outbound - we helped his team at Acme do the same." | Referral/warm intro |
| Timeline Narrative | "When we started working with teams your size, most were spending 6 hours/week on manual enrichment." | Timeline hooks (highest reply rate) |
Timeline hooks outperform everything else: Data from 2025 shows timeline-based hooks achieve 10% reply rates vs. 4.4% for problem-based hooks - a 2.3x gap. Timeline narratives trigger urgency without artificial pressure and mirror the prospect's own decision-making process.
AI model selection for personalization:
| Model | Strength | Best Use |
|---|---|---|
| Claude Sonnet | Natural tone, avoids corporate speak | First lines, full email drafts |
| Claude Opus | Deep research synthesis | Complex enterprise personalization |
| GPT-4o | Speed, structured output | Batch processing at scale |
| Claude Haiku | Cost-efficient | Simple variable generation |
Claude models produce the most natural-sounding cold emails. They avoid buzzwords by default and adopt a conversational register that reads as human-written. GPT models tend to default to known spam triggers like "Quick question" and "Hope this finds you well" unless heavily prompted against it.
Scaling AI personalization with Clay:
A sequence is the multi-step campaign structure. It defines how many emails, when they send, and what each email does.
The anatomy of a high-performing sequence:
Day 0: Email 1 - The opener (personalized, carries the hook)
|
Day 3: Email 2 - Value add (case study, data point, or insight)
|
Day 7: Email 3 - Social proof (specific result for similar company)
|
Day 12: Email 4 - Breakup/new angle (shift approach entirely)
|
Day 18: Email 5 - Permission-based close ("Should I close this out?")
Sequence length and timing rules:
| Factor | Recommendation | Why |
|---|---|---|
| Total emails | 4-7 | First email captures 58% of replies. Diminishing returns after 7. |
| Gap between emails | 2-4 business days | 3 days is the sweet spot. Less feels pushy, more loses momentum. |
| Total sequence duration | 14-25 days | Beyond 25 days, leads go stale. |
| SMB sequences | 5-8 touches over 30 days | Shorter decision cycles. |
| Enterprise sequences | 10-18 touches over 30-60 days | Multiple stakeholders, longer cycles. |
Conditional branching logic: Modern sequences are not linear. Build branches based on:
A/B testing framework: Test ONE variable at a time across minimum 200 sends per variant:
| Priority | Variable | Impact on Reply Rate |
|---|---|---|
| 1 | Subject line | 20-40% swing in open rate |
| 2 | First line / hook | 2-3x reply rate difference |
| 3 | CTA style | 1.5-2x reply rate difference |
| 4 | Email length | Moderate impact |
| 5 | Send time | Marginal impact |
Infrastructure is where most outreach systems break. Perfect copy with bad deliverability lands in spam.
Domain and mailbox architecture:
Primary Domain: yourcompany.com
(NEVER use for cold outreach)
Secondary Domains (for outreach only):
yourcompany-team.com --> mailbox1@, mailbox2@
tryyourcompany.com --> mailbox1@, mailbox2@
getyourcompany.com --> mailbox1@, mailbox2@
yourcompanyhq.com --> mailbox1@, mailbox2@
Formula:
Daily volume target / 150 = domains needed (round up)
Add 30-50% for rotation reserve
Example: 600 emails/day
600 / 150 = 4 domains minimum
+ 50% reserve = 6 domains total
x 2 mailboxes each = 12 mailboxes
Infrastructure sizing guide:
| Daily Volume | Domains Needed | Mailboxes | Monthly Domain Cost |
|---|---|---|---|
| 100-200 | 2-3 | 4-6 | $20-30 |
| 300-500 | 3-5 | 6-10 | $30-50 |
| 500-1,000 | 5-8 | 10-16 | $50-80 |
| 1,000-2,000 | 8-15 | 16-30 | $80-150 |
| 2,000+ | 15+ | 30+ | $150+ |
Per-mailbox sending limits:
| Type | Daily Limit | Notes |
|---|---|---|
| Warmup emails | 15-20/day | Run for 14-21 days before cold sends |
| Cold emails | 25-30/day | Never exceed 40 |
| Combined total | 40-50/day | Stay under provider thresholds |
Domain warmup protocol:
| Week | Daily Volume/Mailbox | Activity |
|---|---|---|
| Week 1 | 10-15 | Warmup only, no cold sends |
| Week 2 | 20-30 | Warmup + 5-10 cold sends |
| Week 3 | 30-40 | Warmup + 15-20 cold sends |
| Week 4 | 40-50 | Full cold sending capacity |
Authentication setup checklist (do this on Day 1):
Authenticated senders are 2.7x more likely to reach the inbox vs. unauthenticated.
DMARC rollout sequence:
p=none with reporting (rua=mailto:[email protected])p=quarantine (soft enforcement)p=reject (full enforcement)Never jump straight to p=reject before inventorying all legitimate senders.
Sending platform comparison: Instantly vs. Smartlead
| Feature | Instantly | Smartlead |
|---|---|---|
| Best for | Solo founders, small teams | Agencies, high-volume senders |
| Pricing (entry) | $37/mo | $33/mo |
| Pricing (scale) | $97-358/mo | $94-174/mo |
| Email accounts | Unlimited (Growth+) | Unlimited (all plans) |
| Built-in lead database | Yes (SuperSearch, 450M+) | No (import only) |
| Warmup network | 4.2M+ accounts | Smaller network |
| AI reply agent | Yes (responds in <5 min) | Limited |
| Deliverability approach | IP sharding + rotation (SISR) | Human-mimicking variable volume |
| Sending behavior | Exact daily volume | Variable (sends 22 when set to 25) |
| API and webhook support | Good | Excellent (API-first) |
| White-label | Limited | Full white-label |
| CRM integration | Built-in basic CRM | Via integrations |
| Clay integration | Native | Native |
| Inbox rotation | Automatic | Automatic |
| Campaign analytics | Detailed dashboards | Detailed dashboards |
| Multi-channel | Email + LinkedIn (beta) | Email focused |
Decision framework:
Need built-in lead database?
YES --> Instantly
NO --> Continue
Running an agency or white-labeling?
YES --> Smartlead
NO --> Continue
Need AI auto-replies?
YES --> Instantly
NO --> Continue
Sending 1,000+/day and need API control?
YES --> Smartlead
NO --> Continue
Want simplest setup and UI?
YES --> Instantly
NO --> Smartlead
Most replies are not "Yes, let's meet." They are questions, objections, or soft interest. AI follow-up handles these at scale.
Reply categories and handling:
| Reply Type | % of Replies | AI Action |
|---|---|---|
| Positive interest | 25-35% | Book meeting link, confirm time |
| Question about offer | 20-30% | Answer with specifics, re-CTA |
| Objection (timing) | 15-20% | Acknowledge, offer future follow-up |
| Objection (budget) | 5-10% | Share ROI data, offer smaller entry |
| Referral to colleague | 10-15% | Thank, ask for intro or direct email |
| Not interested | 10-15% | Thank, remove from sequence |
| Auto-reply/OOO | 5-10% | Pause, re-send after return date |
AI reply handling setup:
Instantly's AI Reply Agent handles this natively and responds in under 5 minutes. Smartlead users typically build this with Clay + webhook integrations.
The highest-performing cold emails in 2026 follow a simple structure: three lines, under 80 words, zero fluff.
Line 1 (PAIN): A specific observation about their situation.
Derived from signal data + AI research.
NOT "Are you struggling with X?" (everyone sends this).
Line 2 (PROOF): One sentence of credibility.
A specific result for a similar company.
NOT "We're the leading platform for..."
Line 3 (CTA): A low-friction ask.
NOT "Book 30 minutes on my calendar."
YES "Worth a quick look?" or "Open to hearing more?"
Example (good):
Noticed you just raised your Series B and are hiring 4 AEs - ramping that many reps without standardized outbound playbooks usually means 2-3 months of wasted pipeline.
We helped Acme's team cut AE ramp from 90 to 45 days after their Series B.
Worth a 10-minute look at how?
Example (bad):
Hi [Name], I hope this email finds you well. I'm reaching out because I noticed your company is growing. We're the leading sales enablement platform trusted by 500+ companies. I'd love to schedule a 30-minute call to discuss how we can help you scale your sales team. Would Tuesday at 2pm work?
Why the bad example fails:
Cold email anatomy rules:
| Element | Rule | Why |
|---|---|---|
| Subject line | 2-5 words, lowercase, no punctuation | Looks like an internal email |
| Preview text | First 40 chars of body visible in inbox | Make the hook visible |
| Word count | 50-125 words | Under 50 feels incomplete, over 125 loses attention |
| Paragraphs | 1-2 sentences each | Mobile-friendly whitespace |
| Links | Zero in first email | Links trigger spam filters |
| Images | Zero in first email | Images trigger spam filters |
| Attachments | Zero in first email | Attachments trigger spam filters |
| Signature | Name + title + company only | Minimal, no banners or social icons |
| CTA | One per email | Multiple CTAs reduce response rate |
| Personalization | First 1-2 lines | Generic everything else is fine if the hook lands |
For benchmarks, deliverability playbook, week-by-week build, cost analysis, failure modes, and advanced tactics read references/benchmarks-deliverability-tactics.md.