Table of contents

Seed Lists Explained: How to Run Inbox Tests Without Getting False Positives

Hugo Pochet
Co-Founder @Mailpool and Cold Email Expert

If you’ve ever run an inbox placement test and thought, “Nice — 90% inbox!”, only to watch your real campaign land in spam… you’ve met the #1 problem with deliverability testing: false positives.
Seed lists are one of the fastest ways to spot deliverability issues before you burn a domain, tank a sender reputation, or waste a week “optimizing” the wrong thing. But seed tests are only useful if you understand what they can (and can’t) tell you and how to run them without misleading results.
This guide breaks down how seed lists work, how to run accurate inbox placement tests, and how to interpret results like a deliverability pro.

What is a seed list (and what is it used for)?

A seed list is a set of test email addresses across multiple mailbox providers (MBPs)  like Gmail, Outlook, Yahoo, iCloud, and regional providers  that you send to in order to measure where your email lands:

  • Inbox (Primary/Focused)
  • Promotions/Updates (for Gmail)
  • Spam/Junk
  • Missing (not delivered/filtered/blocked)

A seed list is primarily used for:

  • Inbox placement testing (where did the email land?)
  • Spam filtering detection (which providers are filtering you?)
  • Authentication validation (SPF/DKIM/DMARC alignment issues often show up here)
  • Monitoring deliverability over time (after DNS changes, warming, new domains, new copy, etc.)
Seed list testing vs. real-world deliverability

Seed lists are a proxy, not a perfect mirror of real campaigns. They’re best for directional signals, not absolute truth.

Think of it like a smoke alarm:

  • It won’t tell you exactly where the fire is.
  • But it will tell you something is wrong, early.

What is an inbox placement test?

An inbox placement test is the process of sending an email to a seed list and measuring:

  1. Delivery rate (did it arrive at all?)
  2. Placement (inbox vs spam vs tab)
  3. Provider-specific behavior (Gmail ok, Outlook spam, Yahoo missing, etc.)

Inbox placement tests help you answer questions like:

  • “Did my new domain setup work?”
  • “Did this new subject line trigger spam filters?”
  • “Is Outlook blocking my IP?”
  • “Did my warmup tool actually help?”

But here’s the catch…

Why seed list tests often produce false positives

A false positive is when your seed test looks good (inbox placement is high), but real sends perform poorly.
This happens because mailbox providers don’t treat seed addresses exactly like normal recipients.

Common causes of false positives
1. Seed addresses behave differently from real recipients

Real inbox placement depends heavily on engagement signals:

  • Opens
  • Replies
  • Deletes without reading
  • Marked as spam
  • Moving messages between folders

Seed addresses often have little or no natural engagement, so providers may treat them differently, sometimes more leniently, sometimes more aggressively.

2. Your sample size is too small

If you send 20 seeds and 18 hit the inbox, that “90% inbox” can be statistically meaningless. One provider’s filtering can skew results massively.

3. You’re only testing one provider “type.”

Many seed lists are heavy on Gmail-style inboxes and light on Outlook variants (which behave very differently). If your real list is 60% Outlook, your seed test can mislead you.

4. You’re testing a “clean” path that doesn’t match the real sending

Example: you test from one inbox, one domain, one IP, one tool, but your real campaign sends:

  • from multiple inboxes
  • across multiple domains
  • through a different sending platform
  • at higher volume

Your seed test becomes a best-case scenario, not a real simulation.

5. You accidentally trained filters to like your test

If you repeatedly send similar emails to the same seeds, providers may classify that pattern as “expected” and stop filtering it the same way they would a new campaign.

6. You’re measuring the wrong thing (tabs vs spam vs inbox)

For cold email, landing in Gmail Promotions might still be “fine” depending on your goals,  but for some teams, it’s a warning sign. If your tool counts Promotions as “inbox,” your results can look better than reality.

How seed lists work behind the scenes

When you send an email to a seed list, the testing system typically:

  1. Receives the email at each seed mailbox
  2. Logs whether it arrived
  3. Checks the folder placement (Inbox/Spam/Tab)
  4. Reports results by provider

Some tools also:

  • parse headers
  • check authentication alignment
  • flag common spam signals

Important: seed list tools do not control mailbox provider decisions, they only observe outcomes.

How to run an inbox placement test the right way (step-by-step)

Here’s a practical process that reduces false positives and makes your results actionable.

Step 1: Test the exact sending setup you’ll use in production

Sounds obvious, but most teams don’t do it.

Match:

  • Sending domain
  • From-name format
  • Mailbox provider (Google Workspace vs Outlook vs shared IP)
  • Sending tool (Instantly, Smartlead, Lemlist, etc.)
  • Tracking settings (open tracking can change outcomes)
  • Volume pattern (don’t test 1 email if you’ll send 500/day)

If your production setup uses multiple inboxes/domains, test a representative sample.

Step 2: Use multiple test variants (don’t test one email)

Run at least 2–4 variations:

  • Subject line A vs B
  • Plain text vs light HTML
  • With vs without links
  • With vs without personalization tokens

This helps you separate:

  • infrastructure issues (domain/IP/auth) from
  • content issues (copy/links/spam triggers)
Step 3: Send in realistic batches (not all at once)

Mailbox providers notice patterns.

Instead of blasting 100 seeds in 30 seconds, send in a cadence closer to real outreach:

  • stagger sends
  • spread across inboxes
  • avoid identical timestamps
Step 4: Include enough seeds per provider to be meaningful

If you only have 1 Gmail seed and it lands in inbox, that tells you almost nothing.

Aim for:

  • multiple Gmail seeds
  • multiple Outlook/Hotmail seeds
  • Yahoo + iCloud
  • plus any regional providers relevant to your audience
Step 5: Run tests at different times

Deliverability can fluctuate by:

  • day of week
  • time of day
  • provider incidents
  • IP reputation changes

Run at least:

  • one baseline test
  • one repeat test 24–48 hours later
Step 6: Track “missing” emails separately from spam

Spam is bad, but missing is worse.

Missing can indicate:

  • hard blocks
  • throttling
  • temporary deferrals
  • silent filtering

If a provider shows “missing,” you likely have an infrastructure or reputation issue.

How to interpret your seed test results (what they actually mean)

Scenario A: Gmail inbox is great, Outlook is junk

This is extremely common.

Likely causes:

  • IP reputation issues affecting Microsoft
  • content patterns Microsoft dislikes
  • too many similar sends
  • misaligned authentication (especially DMARC alignment)

What to do:

  • verify SPF/DKIM/DMARC alignment
  • reduce links/tracking
  • slow down send rate
  • consider provider mix (Google vs Microsoft inbox types)
Scenario B: Everything is inbox… but real campaigns still spam

Classic false positive.

Likely causes:

  • real recipients aren’t engaging
  • list quality is poor (bounces, traps, old leads)
  • your real sends are higher volume
  • your real copy includes different links/personalization

What to do:

  • validate list hygiene
  • compare production headers to test headers
  • run a test that matches production volume and cadence
  • monitor engagement metrics (replies matter a lot)
Scenario C: High spam across all providers

This usually points to:

  • authentication problems
  • domain reputation issues
  • new domain not warmed
  • aggressive content (spammy phrasing, heavy links, tracking)

What to do:

  • check SPF/DKIM/DMARC
  • reduce send volume
  • simplify content (plain text, fewer links)
  • warm for 3–4 weeks before scaling

The biggest “spam” misconceptions in seed testing

Misconception 1: “Spam words” are the main reason you land in spam

Words matter, but reputation and behavior matter more.
If your domain is new, your sending pattern is aggressive, or your list is weak,  removing the word “free” won’t save you.

Misconception 2: “Inbox placement is a single score”

Deliverability is provider-specific and dynamic.
A single percentage hides the truth:

  • Gmail 100% inbox
  • Outlook 20% inbox
  • Yahoo 60% inbox

That’s not “80% inbox.” That’s “Microsoft hates you.”

Misconception 3: Seed tests replace real-world monitoring

Seed tests are proactive. Real-world monitoring is reactive.
You need both.

Best practices to avoid false positives (quick checklist)

Use this before trusting any seed test:

  • Test using the exact production sending stack
  • Run multiple variants (subject/copy/link structure)
  • Use enough seeds per provider
  • Send in realistic cadence, not a blast
  • Repeat tests over multiple days
  • Separate spam vs missing
  • Compare results against:
    • bounce rate
    • reply rate
    • spam complaints (if available)
    • provider-specific deliverability trends

When you should run seed list tests and when you shouldn’t

Run a seed test when:
  • You change DNS/authentication
  • You increase volume
  • You change copy structure (links, HTML, tracking)
  • You see a sudden drop in replies
Don’t rely on seed tests when:
  • Your list is unvalidated or low quality
  • You’re sending at a completely different volume than the test
  • You’re trying to predict exact campaign performance
  • You’re only testing one provider or tiny sample sizes

A smarter approach: seed tests + real deliverability signals

If you want accuracy, combine seed tests with:

  • Warmup progress (are inboxes behaving normally?)
  • Bounce rate (list quality + domain health)
  • Reply rate (strong proxy for inbox placement in cold outreach)
  • Provider breakdown (Gmail vs Outlook performance)
  • Header/auth checks (SPF/DKIM/DMARC alignment)

Seed lists tell you where you land. Real metrics tell you why it matters.

Seed lists are powerful if you run them like a scientist

Seed list testing is one of the fastest ways to catch deliverability issues early, but it’s also one of the easiest ways to fool yourself.
If you:

  • test realistically,
  • use enough seeds,
  • repeat over time,
  • and interpret provider-by-provider,

…you’ll stop getting “great” test results that don’t match reality and start spotting problems before campaigns suffer.

Want to test inbox placement without guesswork?

Mailpool makes it easy to set up inboxes, configure DNS correctly, and scale cold outreach while protecting deliverability, without spending days piecing together infrastructure.
Book a demo and we’ll show you how to build a deliverability-safe sending setup in minutes.

Blog

More articles

Everything about cold email, outreach & deliverability

Get started now

You're just one click away from a top-notch email infrastructure with Mailpool.