The Dark Funnel
All posts

How to Build Your First Outbound Program (Without Wasting Six Months)

A step-by-step playbook for standing up AI-native outbound from scratch — account scoring, experimentation, channel selection, and the feedback loops that make it compound.

March 25, 2026 | The Dark Funnel

Most first outbound programs fail the same way.

Someone hires three SDRs, buys a list, drops everyone into a sequence, and waits. Six months later, pipeline is thin, morale is lower, and leadership is debating whether “outbound even works for our business.”

It works. The program was just built wrong.

Here’s how to build it right.


Step 1: Know who you’re going after before you send a single email

The most common mistake is starting with messaging before you’ve done account selection.

Your ICP isn’t “mid-market SaaS companies.” That’s a category. You need to know: which companies in that category have the problem you solve, have budget to act on it, and are showing signals right now that they’re in a buying window.

Install a tool like Keyplay before you do anything else. Use it for two things:

TAM mapping. Get a real count of how many companies actually fit your ICP — firmographics, tech stack, growth signals. Most teams think they have 50,000 accounts in their TAM. It’s usually 3,000 real fits and 47,000 noise.

Account scoring. Weight your list so reps are always working the highest-fit, highest-signal accounts first. Not alphabetically. Not by list upload date. By actual buying potential.

This is table stakes now. If your reps are working unscored lists, they’re burning cycles on accounts that will never buy.


Step 2: Run experiments, not campaigns

The difference between a campaign and an experiment is intentionality.

A campaign is: “We’re going to send 500 emails about feature X to VPs of Sales.”

An experiment is: “We’re testing whether VP of Sales at Series B SaaS companies with 50–150 headcount respond better to pain-led messaging about rep ramp time versus ROI-led messaging about pipeline coverage — and we’ll know in three weeks.”

In the first 60 days, you should be running 3–5 concurrent experiments across:

  • Persona. Who are you talking to — economic buyer, technical buyer, champion? Same company, different results.
  • Messaging angle. Pain-led vs. ROI-led vs. competitive vs. status quo disruption.
  • Company profile. Funded vs. bootstrapped. High growth vs. steady state. Tech-forward vs. traditional. Your product doesn’t resonate the same way across all of them.

The goal isn’t to win. The goal is to learn fast enough that by week eight, you’re doubling down on what works and killing what doesn’t.

What “not working” looks like is always more obvious than what “working” looks like. Pay attention to the floors, not just the ceilings.


Step 3: Don’t pick your channels — let the data pick them

There’s a persistent myth that every B2B company should be doing email + LinkedIn + phone in a coordinated multi-touch sequence.

Some businesses live on cold calls. Some get all their replies from LinkedIn DMs. Some have reply rates on cold email that blow everything else out of the water. It is not the same for every company, every persona, or every geography.

The right call is to run your experiments across all three channels simultaneously, with enough volume to get statistical signal, and let the results tell you where to concentrate.

What you’re looking for:

  • Email: open rate, reply rate, positive reply rate (not just total replies — “unsubscribe” is a reply)
  • LinkedIn: acceptance rate, response rate, meeting booked rate
  • Phone: connect rate, conversation rate, conversion to meeting

Once you have 30–60 days of data, you’ll see it clearly. Double down on the channel that’s converting. Don’t spread effort evenly across channels just because it feels more thorough.


Step 4: Put sharp BDRs on this, not warm bodies

This is the part most people get wrong about AI-native outbound.

AI handles the leverage — personalization at scale, signal routing, sequence automation. But the people in the loop matter more than ever, not less.

You need BDRs who are:

  • Genuinely curious about what’s working. Not just hitting their dial count, but asking why one message crushed it while a nearly identical one got nothing.
  • Fast feedback loops. They’re on the front lines. They hear objections, they sense when a sequence feels off, they notice that a certain company profile is ghosting. If they’re not surfacing that insight, you’re flying blind.
  • Willing to contribute ideas. The best experiments often come from a BDR who says “I feel like we should be leading with X instead of Y.” Run it. Find out.

Unmotivated BDRs turn an AI-native program into an expensive autodialer. Sharp BDRs turn it into a compounding system.

Hire fewer. Hire better. Pay them more. It’s not close.


Step 5: Build the feedback loop into the system from day one

This is what separates programs that plateau from programs that compound.

Every signal that comes out of your outbound — reply rates, meeting rates, objections, win/loss data — needs to flow back into the system in a structured way. Not as a Slack message or a weekly all-hands. As structured data that improves how your AI picks accounts, writes messages, and routes sequences.

In practice, this means:

Tagging your outcomes. Every reply gets categorized: interested, not now, wrong person, already have a solution, price objection. This isn’t busywork — it’s training data.

Closing the loop on booked meetings. When a sequence converts, what was the message, the persona, the signal that triggered outreach, the company profile? Capture it. Replicate it.

Killing losers fast. If a variant isn’t converting after 100 touchpoints, it’s dead. Don’t run it for another month to be sure. Kill it and reallocate to a new test.

The programs that compound are the ones where the system in month six is meaningfully smarter than it was in month one — because the feedback loop is tight enough to learn.


The secret weapon: chained agents

Here’s where most AI outbound implementations are leaving serious performance on the table.

Teams treat AI as a single step — “write me a cold email for this prospect.” That’s like hiring David Ogilvy and asking him to cold-call without a brief. The output is generic because the input is generic.

The breakthrough is chaining agents together so each one does one thing extraordinarily well, and hands off a richer context to the next.

A proper agent chain for outbound looks like this:

Agent 1: Company researcher. Given a target account, this agent digs into everything publicly available — recent press, job postings, product announcements, tech stack signals, executive LinkedIn activity, earnings calls if they’re public, G2 reviews, Glassdoor patterns. The output isn’t a summary. It’s a structured brief: what is this company doing right now, what problems does that imply, what’s changing.

Agent 2: Fit analyst. Takes that brief and runs it against your product. Not “do we fit” — that’s too broad. It answers: where specifically do we fit, what is the most credible problem we solve for this company given what’s happening right now, and what’s the angle that will land. It also flags where fit is weak, so you don’t send embarrassing pitches into accounts that will never close.

Agent 3: Persona writer. This is where the Mad Men principles come in. Takes the company brief, the fit analysis, and the specific person you’re reaching — their title, their apparent priorities, what they’ve published or posted, what their role actually cares about day-to-day — and writes the message.

Not a template with variables swapped in. An actual piece of copy built around one idea, one tension, one reason this person should care today.


Why 1950s copywriting still beats 2024 AI slop

David Ogilvy didn’t write ads. He wrote arguments. Every headline was a claim. Every body was proof. Every call to action was earned.

The principles haven’t changed:

Lead with the reader, not yourself. The worst cold emails open with “We at [Company] are excited to…” Nobody cares. The best ones open with something the reader recognizes as true about their own situation.

One idea per message. Not five features, not three use cases, not a paragraph about your Series B. One tension. One resolution. One ask.

Specificity is credibility. “We help companies like yours grow faster” is noise. “I noticed you posted two VP of Sales roles last month — that’s usually the point where pipeline coverage becomes the constraint, not headcount” is signal.

The offer has to be worth the cost. The cost of replying to a cold email is small but real — time, attention, the risk of getting into a sales cycle you don’t want. The offer has to clear that bar. A generic demo link doesn’t. A specific insight or a sharp question sometimes does.

The difference between AI-generated slop and AI-generated copy that converts is whether you’ve given the model the brief to actually do the job. That’s what the agent chain delivers — a brief good enough that the writing agent can execute like a professional.


What this looks like end to end

A prospect gets scored and flagged by Keyplay. That triggers Agent 1 to run a company research pass — job postings, news, LinkedIn activity, product announcements. Agent 2 takes that output and runs fit analysis against your ICP criteria and product positioning, outputs a structured brief with the top two or three angles ranked by strength. Agent 3 takes the brief plus the specific contact’s profile and writes three message variants — one for email, one for LinkedIn, one as a call talk track opener.

The whole chain runs in under two minutes. The BDR reviews, picks the best variant or edits lightly, and sends.

What used to take a good SDR 45 minutes of research and writing per account now takes two minutes and produces better output — because the research is more thorough than any human would do at scale, and the writing is built on a real brief instead of a gut feeling.

That’s the leverage. That’s what compounds.


What this looks like in practice

Week 1–2: TAM mapping and account scoring in Keyplay. Clean list, scored list, prioritized list.

Week 3–4: First experiments live. Three to five tests across persona and message angle. Multi-channel from day one.

Week 5–8: Data starts coming in. Kill the clear losers. Start doubling down on early signals.

Week 8–12: Channel concentration becomes clear. Winning persona and message angle emerges. BDRs contributing refinements.

Month 4+: You’re running a machine, not a campaign. Feedback loop is live. The system is learning faster than your competitors are moving.


The companies that will have the best outbound programs in 2028 aren’t the ones that bought the most tools or hired the most SDRs. They’re the ones that ran the most experiments, built the tightest feedback loops, and compounded their learnings the earliest.

Start that clock now.

More Intelligence

The Brief

Get the AI GTM brief before your portfolio figures it out.

AI outbound teardowns, portfolio GTM patterns, stack analysis.

1–2x/month. Only when something's worth reading.