Partner Tech Talk: AI, MCP, and Operator Workflows That Scale

Published on March 2026
Expert advice from Sean Harris (Sr. Director of Product, PartnerStack) and Justin Zimmerman (Founder, Partnerplaybooks)

Table of Contents

Snapshot

We are at the point where AI will stop being a collection of tools you try and start becoming a layer inside how partnerships operate. You are either going to use AI to compress the boring parts of your workflow and push time toward strategy, or you are going to watch the operational overhead pile up as partner programs scale.

The stakes are simple: partner discovery, recruitment, activation, and reporting already consume too much time across teams, and the “all or nothing” integration approach is too slow for modern go-to-market cycles. AI-centered product thinking changes that by moving from search to recommendations, from message personalization to AI-driven outreach, and from manual lead logging to automated capture from the channels your partners already use.

Below, you will get a framework for thinking about AI maturity in partnerships, plus concrete examples of how product and partnerships teams are using agent-style workflows to reduce manual effort in partner discovery, recruitment outreach, and lead capture. You will also see where MCP and API-first designs fit, and how to avoid token-limits and “UI-only” AI that never reaches the outcome.

“AI is transforming the way that we work, and the people who apply it to existing workflows will separate from those who just dabble.” -Justin Zimmerman

The big shift: stop starting with tools

Stop collecting AI tool names. Start by identifying the problem you are trying to solve and the workflow you are trying to improve.

Justin Zimmerman puts it bluntly: AI is here to solve old problems faster. The risk is that you get pulled into the “new shiny tool” cycle, where everyone has opinions, but nobody has outcomes.

So your starting point should look like this:

  1. List your recurring workflows across partnerships and go-to-market (discovery, recruitment, partner messaging, lead capture, reporting).
  2. Pinpoint the manual work you wish you could remove (translating call insights, searching for partners, chasing forms, logging referrals, building status reports).
  3. Map inputs to outputs: What do you need to start the process, and what is the final artifact or decision?
  4. Choose the AI layer that can compress the bottleneck without breaking trust or accuracy.

Once you think this way, AI becomes an accelerator rather than a distraction.

Justin Zimmerman slide titled “Justin’s predictions about 2026” showing a quote about operators building AI into workflows

“Don’t start with the tools. Start with the problems and the workflows you already have, then choose what fits into them.” -Justin Zimmerman

Skills: people skills and AI skills

Sean Harris introduces a powerful distinction that helps you avoid confusion when people talk about “skills” in AI. He describes two kinds of skills you must build in parallel.

1) People skills (the operator layer)

This is your ability to operate with AI effectively. It includes knowing how to prompt, how to verify outputs, how to structure tasks, and how to iterate on workflows until the result is usable. In other words, your own professional fluency.

2) AI skills (the workflow layer)

This is the actual, reusable skill artifact you teach to AI systems. In Sean’s framing, AI skills are standardized “operating guides” made to repeat steps and access tools when needed. They reduce the “starting fresh” problem you face every time a new context window is opened.

The key insight: if you only build people skills, you keep doing everything manually with better prompts. If you only build AI skills, you may automate steps that still do not match how your org actually works. You need both.

“AI-specific skills are directives you give your AI access to, so it can repeat the same steps with more context.” -Sean Harris

How AI skills work (and why they feel like “operating guides”)

In practice, an AI skill is a repeatable procedure. Sean describes it as a standard format where you provide directives, and the system follows an established set of steps. That matters because it turns AI from “a chat you talk to” into “a process you can run.”

He offers a concrete internal example: a skill connected to Linear (their Jira-like system) and Slack updates on releases. Instead of someone spending a day collecting data and writing a report, the AI skill can pull data and generate a weekly report across product, design, and engineering.

This is where the concept becomes operational. You are not asking for one-off answers. You are building an automation loop that gets better through iteration.

Why the improvement loop matters

Sean also highlights that your skills can improve when you provide feedback. You can mark recommendations as bad fit or save for later, and the next iteration gets smarter because your actions become signal. That is how you move from “interesting demo” to “workflow you trust.”

A simple skills example: “Grill me”

Sean shares a skill example called “Grill me,” designed to stress test a plan or design. It asks about context in a structured way. Then, when you ask for a plan, the output is far more grounded because the skill forced the necessary questions upfront.

What you should take from that: the best skills do not just answer. They ask the questions that prevent bad assumptions.

Screenshot of an AI skill document named “grill-me” with instructions to interview users relentlessly about a plan or design

“Interview the user relentlessly about a plan or design until you reach a shared understanding.” -Sean Harris

MCP vs. APIs: how systems will connect in the near term

Partnership workflows are rarely inside a single system. They cross PRM platforms, CRMs, ticketing tools, communication channels, spreadsheets, reporting layers, and marketing stacks.

This is where MCP enters the conversation, along with a more grounded view that you should not over-rotate on any single approach.

What MCP is meant to do

MCP is positioned as a way for AI tools to connect to external systems through standardized servers. The promise is faster integration and easier tool-to-tool communication.

The hot take: MCP is likely to mix with APIs

Justin predicts demand for MCP-powered features, especially when you want to connect in-house AI workflows to your core systems. Sean agrees, but adds a key practical caution: MCP can be constrained by usage limits and token consumption. He describes a situation where you can run a few MCP server calls, then you are out of tokens and cannot run more for hours.

So his directionally strong advice is to evolve toward an API-centered approach for many use cases, while using MCP as the easiest path for connecting apps quickly.

Why reporting is the deciding factor

Sean points out that many complex workflows involve reporting and data processing. If an MCP server pulls all data without “reporting-specific apis,” you can waste time and resources. That pushes you toward designing APIs that support targeted reporting queries so the agent does not brute force everything.

Takeaway: if you are building or selecting partnership tooling, think in terms of architecture. You want the option to connect with AI, but you also need reliable, purpose-built API endpoints for high-frequency reporting and performance metrics.

Operator workflows: recruitment, referrals, and lead capture

It is easy to get lost in AI terminology. The fastest way to cut through is to look at how agent workflows show up inside the daily work of partner managers.

The conversation centers around three stages:

  • Recruitment (discover and invite partners)
  • Activation (help partners execute and engage)
  • Reporting and analytics (answer questions quickly without UI friction)

Justin and Sean both emphasize that you should prioritize high-frequency, admin-heavy tasks. The goal is to reduce time spent being “an administrator” and increase time spent on strategy and partner relationships.

Recruitment as the first high-frequency use case

Sean describes recruitment as a universal pressure point. As you scale, it becomes harder to do consistently. So the AI layer should help you find good-fit partners and create personalized outreach without requiring weeks of manual research and drafting.

Activation as the next frontier

Recruitment is not enough. Once partners join, you need them to take meaningful actions. Sean says they plan to build full co-pilot and agent support for activation in Q2, so partner managers can guide partner actions without becoming gatekeepers.

Reporting as the “UI friction” problem

Even if AI can generate text, the real pain is navigating dashboards. Sean’s direction is: provide a co-pilot that answers questions about common partner metrics. Then, connect it to reporting APIs so you can get instant answers to “chart questions” rather than clicking through complex reporting views.

Real examples from PartnerStack-style workflows

Sean walks through a sequence of product capabilities that map directly to partnership work. Use these as a pattern library you can adapt to your own tools.

PartnerStack-style partner program dashboard showing program metrics, tasks, and segments

“Convert partner discovery from search into recommendations.” -Sean Harris

Example 1: Discover network partners with recommendation signals

Instead of showing a huge list of partners and hoping users can filter well, PartnerStack’s product processes partner data and builds a recommendation engine. Each week, it evaluates company attributes and suggests top partners. It even includes “AI out of match” style badges as signals for best-fit.

Then it incorporates human feedback: marking a partner as bad fit or saving for later improves the algorithm for future recommendations.

PartnerStack recruit agent screen showing recommended network partners and agent status

Discovering the right partners stopped being a manual search for us—it became recommendations driven by partner-fit signals, and we learned fast that feedback like ‘bad fit’ and ‘save for later’ makes the network smarter every week.” – Sean Harris

Example 2: Recruit agent for automated outreach with pause control

They add an agent that automates outreach, selects AI templates, and adapts messaging for each partner. The key part is not full automation without oversight. It includes user control to pause the process at any time.

The workflow also starts with targeted inputs: partner type and locations. For example, if someone wants resellers in Japan, the agent can generate personalized outreach emails that explain why each partner fits.

That is the difference between “personalization as a promise” and “personalization as a system.”

Slack referral messages routed into a partnership PRM workflow via an AI submit-lead flow

“Treat [automated outreach] like an operator workflow—templates plus partner-fit context, with pause and control—so the agent does the boring work, and your team stays in charge.” -Sean Harris

Example 3: Slack connector to submit referrals automatically

Many partner teams exchange lead and referral information inside Slack, email, and other channels before it ever reaches the PRM. Sean describes a real friction loop: partners send referrals in Slack, but partnership teams still have to ask partners to log everything in the PRM.

So they build a Slack connector plus an AI submit-lead flow. When a referral arrives in Slack, the connector can capture the content and submit the referral automatically to the PRM workflow. The user sees the referral appear in the vendor dashboard with details and activity logs, including that it was referred from Slack.

What this unlocks for you

  • Less awkward chasing for form completion
  • Higher data hygiene because capture happens earlier
  • Faster pipeline motion because leads enter the system without manual transcription

Sean also mentions that there is a parallel email flow where referrals forwarded to a specific email address are processed similarly. The principle is consistent: meet partners in their existing communication patterns.

“Automated referral submission only works when you meet partners where they already are—like Slack—and then wire that signal straight into the PRM so your team isn’t chasing forms or doing manual transcription.” -Sean Harris

Example 4: Reporting co-pilot with MCP server alongside

Next, they shift to reporting and analytics. The co-pilot can answer questions about common partner metrics across the product experience. The idea is to avoid adding too many UI layers and instead tune reporting APIs, then launch an MCP server so the agent can access the data needed for answers.

“Reporting is where teams feel the UI friction most—so the win is answering partner questions instantly by wiring agents to purpose-built reporting data, not by making people click through dashboards.” -Sean Harris

Why this matters for partner leaders

If you frequently ask questions like “Which partners performed best this month?” “How much recurring revenue did we drive?” or “What is our conversion rate from referrals to leads?” you should not need to click through complex dashboard paths every time.

You need an outcome-first system where you can ask and get an answer that is grounded in the same reporting truth your business uses.

If you want to apply these ideas without copying a vendor product exactly, use these stack patterns. They are designed to reduce admin load and create repeatable outcomes.

Pattern 1: Build skills around high-frequency artifacts

  • Weekly partner performance reports
  • Call-insight translation into structured notes
  • Partner discovery recommendations
  • Referral capture from Slack or email

Then connect skills to data sources: tickets, CRM, PRM, and comms.

Pattern 2: Pair AI skills with operator guardrails

Do not fully remove control. Use pause and review points, especially for outbound outreach. You want the agent to do the boring work, not to surprise you with risk.

Pattern 3: Design your reporting APIs for AI access

Instead of expecting AI to retrieve every row and then summarize, implement targeted reporting endpoints. Sean’s caution about MCP pulling too much data is a warning: your system should return just what the agent needs.

Pattern 4: Move from “marketplace integration” to “outcome connections”

Traditional integration work can be slow: build, align, execute, then ship. MCP-style connections plus AI skills can compress time for long-tail use cases by enabling micro-workflows. Over time, deeper integrations can follow once you validate demand.

FAQs

What does it mean that AI skills are “operating guides”?

It means the AI skill is a reusable set of directives and steps you give the AI so it can repeat the same workflow. Instead of starting from scratch in every context window, the skill provides structure and can also connect to tools to fetch the right information.

Should you start by learning AI tools or mapping workflows first?

Map workflows first. Identify recurring bottlenecks and manual work, then choose AI capabilities that fit those workflows. This prevents you from getting stuck in tool-chasing without measurable outcomes.

Is MCP the future, or will APIs replace it?

MCP is likely to remain useful because it is an easy way to connect systems. But many teams will transition toward APIs for efficiency, especially in reporting-heavy workflows. The practical outcome is a mix: MCP for quick connections, APIs for reliable data access at scale.

Where do AI agents actually help partner managers day to day?

They help with high-frequency tasks like partner discovery recommendations, drafting and personalizing outreach, capturing referrals from Slack or email, and generating reporting answers without navigating complex dashboards.

How do you keep AI from creating risk in outbound partner outreach?

Add human-in-the-loop checkpoints. Use pause controls, templates with guardrails, and review points before sending. The agent can draft and tailor, while you confirm final messaging for compliance and brand safety.

What is the quickest “first project” for implementing these ideas?

Start with one workflow that is repetitive and time-consuming: automated lead/referral capture from Slack or email, or a weekly report generation skill. Pick something measurable so you can prove time savings and data quality improvements.

Conclusion

Your AI advantage in partnerships will not come from owning the newest model or the flashiest app. It will come from turning AI into repeatable, outcome-focused workflows. Start with the old problems you already know: partner discovery, recruitment outreach, referral capture, activation support, and reporting. Then build or adopt AI skills that compress the boring work, pair them with operator guardrails, and connect them to systems through a practical mix of MCP-style connections and API-first reporting endpoints. If you do that, you will move from “dabbling” to becoming a 10X operator.

Partner Playbooks

Your trusted source for B2B SaaS playbooks

© 2026 Partner Playbooks. All rights reserved.