What you'll learn: The nine dimensions of a high-performing PM operation, what bad vs. good looks like for each, the four maturity tiers, and how to calculate what your gaps actually cost.
Discovery
Research, interviews, JTBD
Strategy
Positioning, market sizing
Competitive
Intel, battlecards, win/loss
Planning
OKRs, priorities, roadmaps
Specs
PRDs, user stories, tech specs
Data
Metrics, experiments, funnels
Communication
Updates, decks, announcements
Launch
Go-to-market, checklists, rollout
Operations
Retros, process docs, team health
Most companies evaluate individual PMs. Performance reviews, 360s, career ladders — all scoped to the person. Almost nobody evaluates whether the PM function is working.
You can have three strong PMs and still have a weak PM operation. One writes great PRDs. Another runs discovery. A third handles launches. But none of them do it the same way, nothing connects, and when one of them leaves, their piece of the operation disappears with them.
The problem isn't talent. It's that there's no shared system underneath the people. And without a way to measure the system, you can't fix it.
This product management maturity model gives you that measurement. Nine dimensions, each scored on a 0–3 scale, that tell you whether your PM team runs on individual effort or on infrastructure that compounds.
Why PM Team Maturity Matters
If you run a product management team, you probably don't have a PM ops function. You don't have someone whose job it is to make sure the PM machine works. That job falls to the Head of Product, who is also shipping features, managing stakeholders, and hiring.
The result: PM operations grow organically. Whoever joins the team brings their own process. Some of it sticks, most of it doesn't, and over time you end up with a patchwork of individual habits instead of a team system.
A product management maturity model gives you a diagnostic. It shows you where the operation depends on specific people and where it runs on shared infrastructure. That distinction matters because people leave, get promoted, or get pulled onto other problems. Systems stay.
The nine dimensions below cover the full surface area of a PM operation — from how you learn about customers to how you transfer knowledge when someone joins or leaves the team.
The dimension I see teams neglect the most is Operations. Discovery and Strategy get attention because they're exciting — they're about customers and direction. But Operations is the dimension that makes everything durable. I've worked with teams where their Discovery score was strong because one PM ran great research. Then that PM left, and the score dropped to zero. If Operations isn't there, your strengths are just one departure away from becoming gaps.
The 9 Dimensions
1. Discovery
Discovery measures how your team turns customer conversations into product decisions.
What bad looks like: PMs talk to customers, but the insights stay in their notebooks. Someone had a great interview last week — nobody else knows what they learned. Customer understanding is high for individual PMs and low for the team.
What good looks like: Every customer conversation feeds a shared insight system that directly informs what gets built. Insights don't live in one person's head. They're synthesized, searchable, and connected to the backlog.
Discovery is the foundation. If your team doesn't have a shared system for customer insight, every other dimension is built on incomplete information.
2. Strategy
Strategy measures whether your team has documented positioning that shows up in daily decisions.
What bad looks like: The founder knows the positioning. Maybe there's a pitch deck from the last fundraise. But ask three PMs who the product is for and how it wins, and you'll get three different answers.
What good looks like: Every PM can articulate who you serve, how you win, and why — and it shows up in their decisions. Positioning isn't a slide. It's a shared mental model that shapes what gets prioritized, what gets cut, and how features get framed.
Without shared strategy, PMs make locally rational decisions that don't add up to a coherent product. The roadmap becomes a collection of good ideas that don't compound.
3. Competitive
Competitive measures whether your team tracks the market systematically.
What bad looks like: If a competitor launched a major feature or changed their pricing tomorrow, your team would find out... eventually. Maybe someone on sales mentions it. Maybe a customer brings it up. There's no system for knowing what the market is doing.
What good looks like: Competitive intel is current, shared, and factored into product and positioning decisions. Your team doesn't get surprised. They know what competitors are shipping, where they're weak, and where the market is heading.
Competitive intelligence isn't about copying competitors. It's about making decisions with full context instead of partial context.
4. Planning
Planning measures whether your team has a shared prioritization framework with real data inputs.
What bad looks like: Two PMs disagree on what to build next. What decides it? Whoever argues loudest. Whoever has more exec access. Whoever framed their idea better in the last meeting. Prioritization is political, not systematic.
What good looks like: A shared framework with real data inputs turns prioritization debates into productive conversations. The framework doesn't eliminate disagreement — it channels it. PMs can argue about inputs (data, customer evidence, strategic fit) instead of arguing about conclusions.
When prioritization runs on a shared framework, the team moves faster because decisions don't require re-litigating the process every time.
5. Specs
Specs measures whether PRDs are consistent across the team — in structure, depth, and quality.
What bad looks like: Pull up the last three PRDs written by different PMs. Do they look like they came from the same organization? Different formats, different depths, different assumptions about what the audience needs. Engineering has to decode each PM's style.
What good looks like: Any PM can pick up any spec and know exactly what to build, regardless of who wrote it. Templates exist. They're actually used. Quality is consistent enough that the author doesn't matter.
Inconsistent specs create inconsistent execution. If engineering has to guess what a PM means, they'll guess wrong some percentage of the time, and you'll pay for it in rework.
6. Data
Data measures whether product recommendations are backed by evidence or by conviction.
What bad looks like: A PM presents a feature recommendation to leadership. What's backing it up? Gut feel. Maybe a few customer quotes. Maybe a screenshot from a dashboard that tells part of the story. Leadership has to trust the PM's judgment because there's nothing else to trust.
What good looks like: Every recommendation comes with a full evidence stack that leadership can trust. Usage data, customer research, market context — assembled into a case that stands on its own. The PM's conviction adds weight, but the evidence does the heavy lifting.
Evidence-based decisions scale. Conviction-based decisions don't. When your team grows from 3 to 5 PMs, you need the evidence stack because the Head of Product can't deeply evaluate every recommendation on gut feel alone.
7. Communication
Communication measures whether product status lives in a system or in people's heads.
What bad looks like: Your CEO asks for a product update for the board. In two hours. The response: scramble. Pull from Slack threads, Jira backlogs, and memory. Piece together a narrative that's accurate enough. Hope nothing important gets missed.
What good looks like: Board-ready updates can be pulled together in minutes because the system stays current. Status isn't something PMs compile — it's something they maintain. The update is always 80% done because the system of record is always current.
Communication gaps don't just waste time. They erode trust. When leadership has to ask for updates, they start wondering what else they're not seeing.
8. Launch
Launch measures whether cross-functional teams are enabled before a release or surprised by it.
What bad looks like: Your team ships a feature. Sales finds out when a customer mentions it. Support finds out when a ticket comes in. Marketing finds out when someone asks why there's no blog post. The feature is live, but the organization isn't ready.
What good looks like: Every launch includes coordinated enablement so sales, support, and marketing are ready on day one. Launch isn't a PM handing off code to engineering. It's a cross-functional event with materials, talking points, and timelines.
Bad launches waste the feature. You spent weeks building something, but the people who talk to customers can't explain it. The impact is delayed by weeks or months — if it ever catches up at all.
9. Operations
Operations measures whether institutional knowledge lives in the system or in individual people.
What bad looks like: A PM leaves the team. What happens to everything they knew? The context behind decisions, the relationships with stakeholders, the nuances of their product area — most of it walks out the door. The next person starts from scratch.
What good looks like: When someone leaves or joins, the system carries institutional knowledge — not individual memory. Onboarding is fast because the system is documented. Offboarding is smooth because the knowledge was never locked in one person's head.
Watch out — Operations is the dimension that makes all the others durable. You can score a 3 on discovery today, but if that capability lives in one person instead of a system, you're one resignation away from a zero.
The 4 AI PM Maturity Tiers
Each dimension is scored 0 to 3, for a maximum of 27 across all nine. Your total score maps to one of four tiers.
Ad Hoc (0--6): Your PM team is running on individual heroics. Every PM invents their own process. Quality depends entirely on who does the work, and when someone leaves, their knowledge leaves with them. This is common at early-stage startups where the PM function is new.
Assisted (7--13): AI helps individual PMs, but not the team. You have started using AI tools, but each PM has their own prompts, their own workflows, their own way of getting output. The gaps between PMs are where things fall through the cracks.
PM OS (14--20): Your PM operation runs on shared infrastructure. Shared context, shared skills, consistent output across the team. Specific gaps remain, but targeted investment compounds fast. This is where most teams that care about PM operations land.
Autonomous PM OS (21--27): Your PM operation runs and improves itself. Context enriches automatically, workflows run on schedules, new PMs ramp in weeks. The system gets smarter with every interaction.
Most PM teams at growing companies fall in the Assisted or PM OS tiers. That's normal. The goal isn't perfection — it's knowing where your gaps are so you can close them deliberately instead of discovering them during a crisis.
The Cost of Gaps
PM maturity gaps aren't abstract. They cost hours and dollars.
Each gap point — the distance between your score and a 3 on any dimension — costs roughly 50 hours per PM per year in wasted effort. That's about an hour per week spent on workarounds, rework, miscommunication, and duplicated effort for each dimension where the system is weak.
Example — A team of 3 PMs scoring an average of 2 on each dimension has 9 total gap points. That's 9 x 50 hours x 3 PMs = 1,350 hours per year. At a loaded PM salary, that's north of $95,000 in productivity lost to operational gaps.
Those hours don't show up as a line item. They show up as features that ship late, launches that fall flat, decisions that get relitigated, and new PMs who take months to ramp. The cost is real. It's just spread across enough small moments that nobody tracks it.
Watch out — The 50-hours-per-gap-point estimate is based on typical PM teams at growing companies. Your actual cost varies with team size, product complexity, and organizational structure. Use it as a directional indicator for prioritizing investment, not a precise financial forecast.
The PM Team Maturity Assessment calculates this number for your specific team size and scores. It takes about five minutes and shows you exactly where the hours are going.
Find Your Gaps
If you manage a PM team and you've never assessed the operation itself — not the people, the system — this is where to start.
Take the PM Team Maturity Assessment. Nine questions, five minutes, and you'll get your score across all nine dimensions, your maturity tier, and the estimated dollar cost of your gaps.
Once you know where the gaps are, you have options. Read what your score actually means for tier-specific guidance and radar chart interpretation. Browse the skills library to find tools mapped to each dimension. See what those gaps cost in dollars or learn how to standardize PM quality across your team. Look at the PM Operating System for a complete system that covers all nine. Or talk to us about services if you want help building the operation, not just the tools.
Build this for your team → We help Heads of Product close the maturity gaps that cost the most — shared skills, context files, and the infrastructure to raise every PM's output to the same bar. See how it works →
Either way, start by measuring. You can't fix what you can't see.
About the Author
Ron Yang is the founder of mySecond — he builds and manages PM Operating Systems for product teams. Prior to mySecond, he led product at Aha! and is a product advisor to 25+ companies.