Cybersecurity Capability Maturity Model for Pentesters

A lot of pentest reports still die the same death.
The test is solid. The screenshots are clear. The findings are accurate. The client gets a long document with criticals, highs, mediums, and a closing meeting where someone on their side asks the only question that really matters: What should we do first, and how do we avoid being back in the same place next year?
That question sits outside a normal finding list.
A list of vulnerabilities tells a client what is broken today. It rarely tells them whether they have a weak process, an immature operating model, poor ownership, or a structural gap that will keep producing the same class of issues. When clients ask for budget, board support, or headcount, they usually need that wider story. “We found weak MFA coverage” lands differently from “your identity controls are still operating at a low maturity level, and that is now affecting detection, response, and resilience”.
That’s where a cybersecurity capability maturity model becomes useful for pentesters, consultancies, and MSSPs. It turns testing output into a roadmap. Instead of handing over defects and hoping the client can sort them into a plan, you can tie findings to capability domains, score current-state maturity, and show what better looks like in practical terms.
For smaller teams, that used to sound good in theory and painful in delivery. The primary blocker wasn’t understanding the model. It was the reporting overhead. Mapping findings to maturity domains, writing the same explanations repeatedly, and rebuilding roadmap pages in Word made the work hard to sell and even harder to standardise.
That’s changed. With a more disciplined workflow and efficient reporting, maturity-led pentesting is realistic even for boutique firms.
Introduction Beyond the Finding List
One of the most familiar client reactions after a penetration test isn’t panic. It’s uncertainty.
They’ve paid for a serious assessment and received a serious report, but the result can still feel like a pile of disconnected problems. A weak password policy sits next to exposed admin interfaces, stale software, missing logging, and privilege issues. Every finding is valid. The client still doesn’t know what sequence makes operational sense.

That gap matters more now than it did a few years ago. Clients don’t just want confirmation that vulnerabilities exist. They want a defensible way to prioritise them, explain them internally, and show progress over time. The pentester who can connect a technical issue to a broader capability gap becomes harder to replace.
Why finding lists stop short
A finding list answers questions like these:
- What is vulnerable: Which asset, workflow, or control failed.
- How it can be exploited: The attacker path, proof of concept, and likely impact.
- How to remediate it: A technical fix or compensating control.
Useful, but incomplete.
It usually doesn’t answer who owns the control area, whether the issue is a one-off or systemic, whether related practices are documented and repeatable, or what maturity target is realistic for the client’s size and sector.
A vulnerability report tells you what to fix. A maturity view tells you what to build.
That distinction changes the engagement. If repeated findings cluster around access control, asset visibility, incident handling, or change management, you’re no longer looking at isolated flaws. You’re looking at an immature capability.
What stronger client conversations look like
A more strategic debrief sounds different. Instead of saying “you have several authentication weaknesses”, you can say:
- Identity controls are inconsistent: The issue isn’t one login page. The issue is that access standards aren’t being applied uniformly.
- Vulnerability management is reactive: Patching happens after urgent requests, not through a predictable process.
- Detection is underdeveloped: The team may prevent some attacks, but they’ll struggle to spot misuse quickly.
- Response depends on individuals: The organisation can react, but not reliably under pressure.
That’s a better conversation because it gives the client a path. It also gives your work a longer shelf life. You’re no longer the person who found bugs in March. You’re the person who showed them how to improve security capability in a way leadership can understand.
For pentesters, that shift is commercial as well as technical. Maturity framing supports retests, quarterly reviews, programme-level reporting, and more useful advisory follow-up.
What Is a Cybersecurity Capability Maturity Model
A cybersecurity capability maturity model is a structured way to assess how well an organisation performs security practices across different domains. It doesn’t just ask whether a control exists. It asks whether that control is ad hoc, repeatable, documented, measured, and continuously improved.
The easiest way to explain it to clients is with a fitness analogy. Counting vulnerabilities is like counting calories for one day. It tells you something, but not much about long-term health. A maturity model is closer to a full training assessment. It looks at endurance, strength, mobility, recovery, and consistency. An organisation can pass one audit and still be unfit.

The five common maturity levels
Different frameworks use different names, but most clients understand a five-level progression like this:
| Level | Practical meaning |
|---|---|
| Initial | Security work is reactive. People do the right thing when they remember or when an incident forces action. |
| Repeatable | Some practices happen consistently. They may rely on a few capable staff members rather than broad organisational discipline. |
| Defined | Processes are documented and expected across teams. Security stops being personal habit and becomes standard practice. |
| Managed | The organisation measures performance, tracks exceptions, and uses evidence to steer decisions. |
| Optimised | Security improves continuously. Teams refine controls using lessons learned, metrics, and changing threat conditions. |
These levels matter because clients often mistake compliance for capability.
A company may have a policy, annual training, and a control spreadsheet. On paper, that looks respectable. But if nobody follows the access review process, logging isn’t reviewed, and incident roles are vague, the capability is still immature. The model helps you expose that gap without turning the conversation into an argument over one failed control.
Why maturity models work in practice
A good model gives you three things a standard pentest report often lacks:
- Context: It groups individual findings into capability areas.
- Sequencing: It shows what should come before more advanced work.
- Continuity: It allows the client to track progress across engagements.
Practical rule: Don’t score maturity based only on what the environment technically allows. Score it on whether the practice is reliable, owned, and repeatable.
That matters when you’re dealing with edge cases. A client may have MFA enabled in several places, but if exceptions are informal and local admin accounts are unmanaged, IAM maturity is still lower than the tooling alone suggests.
A real example from the UK context
This isn’t just a consultancy convenience. At national level, maturity models have been used to benchmark cybersecurity capacity in a way that drives investment and prioritisation. The Cybersecurity Capacity Maturity Model for Nations, developed in 2014, has been used in over 80 countries, and in the UK it highlighted strengths in legal frameworks but a gap in societal cybersecurity culture, where 65% of organisations reported thorough employee training. That benchmarking informed programmes including the UK’s £2.6 billion National Cyber Security Programme, as described in the CMM 2021 edition from the University of Oxford GCSCC.
For client work, the lesson is simple. The model isn’t paperwork. It’s a way to turn scattered evidence into priorities that decision-makers can act on.
Comparing Common Maturity Models C2M2 vs CMMC
When practitioners talk about a cybersecurity capability maturity model, they often blur together frameworks that were built for different jobs. That’s a mistake in delivery. You don’t want to force a defence-supply-chain style model onto an energy client, and you don’t want to treat a regulated contractor’s certification problem like a general improvement exercise.
For consultancies and MSSPs, two names come up often: C2M2 and CMMC.
Where C2M2 fits
C2M2 is usually the more practical model when you’re working with critical infrastructure, operational technology, or clients that need a capability view across multiple security domains. In the UK, it has been used in critical infrastructure contexts, and a 2023 NCSC assessment found only 28% of UK energy firms achieved MIL2 in Threat and Vulnerability Management. Organisations at the lower MIL1 level experienced a 2.3x higher rate of ransomware incidents, according to the cited C2M2 reference material.
That’s useful to a pentester because it links technical hygiene directly to maturity progression. You’re not just saying, “scan and patch more consistently”. You’re saying, “this domain is underdeveloped, and that immaturity has operational consequences”.
C2M2 also works well when clients need a domain-by-domain discussion. It supports reporting that separates identity, risk, vulnerability management, incident response, and governance into distinct capability conversations.
Where CMMC fits
CMMC is a better fit when the client’s primary concern is contractual eligibility and formal assurance, especially in defence-related supply chains. It is more compliance-driven in how buyers experience it. The outcome isn’t just a clearer roadmap. It can affect whether the organisation qualifies for certain work.
For firms dealing with defence-adjacent clients, it helps to understand the wider impact of CMMC certification before you shape the assessment or report. That context changes the tone of the engagement. Clients in that space often care less about a broad maturity narrative and more about proving they can meet externally imposed requirements.
The practical difference for pentesters
The distinction is not academic. It changes how you scope, test, and report.
With C2M2, you can start with technical findings and map them into capability domains. A weak asset inventory process, absent vulnerability review cadence, and inconsistent exception handling all support a broader maturity judgement.
With CMMC, the evidence burden is often more control-centric. The client may need defensible proof that required practices are in place and maintained, not just an advisory view that they should improve over time.
If a client asks, “Which model should we use?”, the better response is “What decision are you trying to support?” before you mention any framework name.
Side-by-side comparison
| Attribute | C2M2 (Cybersecurity Capability Maturity Model) | CMMC (Cybersecurity Maturity Model Certification) |
|---|---|---|
| Primary use | Capability improvement across security domains | Certification and compliance readiness |
| Typical fit | Critical infrastructure, energy, OT-heavy environments | Defence supply chain and contract-driven environments |
| Assessment style | Domain maturity review | Requirement and control validation |
| Reporting value for pentesters | Strong for roadmap building and advisory reporting | Strong for gap-to-requirement mapping |
| Client conversation | “How mature are our practices?” | “Can we meet the required level?” |
| Best use in an MSSP workflow | Ongoing programme reviews and maturity tracking | Pre-assessment readiness and remediation planning |
Choosing without overcomplicating it
Most smaller consultancies don’t need to become framework purists. They need a method that improves client outcomes without bloating delivery.
A practical way to choose is this:
- Use C2M2-style thinking when the client wants to improve operational capability and prioritise investments sensibly.
- Use CMMC-oriented reporting when the client has certification pressure, contractual scrutiny, or procurement risk.
- Blend carefully when a client needs both. Keep the maturity narrative separate from the requirement checklist so the report stays readable.
If you want a parallel example of how maturity frameworks can be integrated into consulting delivery beyond cybersecurity alone, the discussion around capability maturity model integration is worth reviewing. The same delivery principle applies: the framework only becomes useful when it shapes decisions, not when it becomes another document to maintain.
What doesn’t work is dropping framework terminology into a pentest report without changing the logic of the report itself. If every recommendation still reads as a standalone fix, the maturity label is just decoration.
How to Conduct a Maturity Assessment
A maturity assessment works best when you run it like an evidence-led engagement, not a workshop full of vague opinions. The workflow is straightforward. Scope the review properly, collect evidence from people and systems, analyse the evidence against the chosen model, then score each domain with enough justification that the client can defend the result internally.

Scope the assessment before you touch a control
Bad maturity work usually starts with a fuzzy scope.
If the client says “assess our cybersecurity maturity”, pin down the environment, business units, and domains that matter. A startup with one cloud environment, outsourced IT, and no SOC does not need the same scope as a mid-sized financial firm with separate infrastructure, application, and identity teams.
Start by fixing four points:
Assessment boundary
Decide whether you are scoring the whole organisation, a business unit, or a specific function such as IAM or vulnerability management.Chosen framework
Pick one model for the score. You can cross-reference others later, but mixed scoring causes confusion.Evidence standard
Agree what counts as proof. Policy documents, tickets, screenshots, exports, logs, interview notes, and sampled control records are all valid if handled consistently.Target state Ask what maturity level the client needs. There is no point pushing for an advanced target if their risk profile and staffing don’t support it.
Collect evidence from three places
A useful maturity assessment blends qualitative and technical evidence. If you only interview stakeholders, you’ll over-score. If you only inspect systems, you’ll miss governance and ownership problems.
Use three evidence streams:
- Interviews: Speak to the people who run identity, patching, logging, change control, and incident response. Ask how work gets done under normal pressure, not how it is meant to work on paper.
- Document review: Read standards, playbooks, escalation procedures, access review records, exception registers, and audit artefacts.
- Technical validation: Sample the environment. Verify controls in tooling, not just in policy.
Questions should be concrete. “Do you have an incident response process?” gets polished answers. “Show me the last time you used it, who approved actions, and where the timeline was recorded” gets useful ones.
Field note: Maturity scoring becomes credible when every score can be traced back to evidence that a second reviewer could inspect.
Analyse the gap between stated and lived practice
This is the stage many pentesters will recognise, because it mirrors the gap analysis mindset already used in technical testing.
You compare what the model expects with what the client can demonstrate. Then you separate:
- Designed controls from working controls
- One-off good practice from repeatable practice
- Tool ownership from process ownership
- Compliance artefacts from operational reality
A client may own a capable IAM platform and still have low IAM maturity if joiner-mover-leaver workflows are inconsistent, local exceptions bypass review, and access decisions are not monitored centrally.
That distinction matters in regulated sectors. In the UK financial sector, regulations mandate C2M2 MIL2+ in the IAM domain, and a Bank of England assessment found that firms at MIL3 for IAM, with MFA across 95% of assets, had 3.7x faster incident response times (MTTR<4 hours) and 74% fewer phishing-led compromises than firms at MIL1, according to the cited UK financial-sector C2M2 discussion. That’s why IAM maturity can’t be scored from policy language alone.
Score by domain, not by gut feel
A final maturity score should be domain-based and explained in plain English. Avoid forcing a single headline level too early. Clients get more value from “IAM is repeatable, vulnerability management is ad hoc, incident coordination is defined but not measured” than from one flattened overall score.
A practical scoring approach:
| Domain | Evidence pattern | Likely interpretation |
|---|---|---|
| IAM | Access standards exist, but reviews and exceptions are inconsistent | Lower-middle maturity |
| Vulnerability management | Findings are fixed, but there is no reliable cadence or ownership | Low maturity |
| Incident response | Playbook exists and has been used, but lessons learned are weak | Mid maturity |
| Logging and monitoring | Telemetry exists, but triage and escalation are informal | Mid maturity at best |
Keep the narrative honest. A lower score with evidence is more useful than a flattering score that collapses on follow-up.
Building an Improvement Roadmap from Your Assessment
A maturity assessment without a roadmap is just a better-organised disappointment.
Once you’ve scored the client, the next job is to turn those scores into a practical sequence of work. That sequence matters because clients rarely fail on awareness. They fail on order. Too many teams try to implement advanced controls on top of weak ownership, poor asset visibility, and unclear operational responsibilities.
Sort recommendations by dependency, not just severity
Severity still matters for urgent technical risk, but roadmap planning needs another lens. Some improvements enable many others. Some are expensive and look impressive but won’t hold if the basics are unstable.
A clean roadmap usually falls into three layers.
Quick wins
These are changes that reduce friction and tighten control reliability without major restructuring.
- Clarify ownership: Assign named responsibility for domains like IAM, patching, and incident coordination.
- Standardise evidence collection: Define what tickets, screenshots, approvals, and logs prove a control is functioning.
- Remove obvious inconsistency: Fix repeated exceptions, stale accounts, or undocumented processes that keep generating similar findings.
Medium-term capability work
At this stage, the client starts moving from isolated remediation into repeatable practice.
- Formalise review cycles: Access reviews, vulnerability review meetings, and exception handling need cadence.
- Document operating procedures: Not glossy policies. Short, usable runbooks that people can follow under pressure.
- Improve visibility: Bring assets, findings, and control ownership into a single operating picture.
Longer-term strategic changes
These efforts usually need budget, sponsorship, or cross-team coordination.
- Metrics and measurement: The client starts tracking whether controls are effective, not just present.
- Integrated response workflows: Detection, triage, and business escalation become connected rather than improvised.
- Continuous improvement loops: Lessons learned feed back into design, standards, and testing.
Clients often overestimate the value of advanced maturity and underestimate the value of disciplined middle maturity.
Build the business case honestly
At this point, practitioners need to be more candid than many sales decks.
Not every step up the maturity ladder produces the same return. A 2025 SANS UK Forum analysis found that moving to C2M2 MIL2 yields 2.1x faster incident recovery, while advancing from there to MIL3 offers only 15% extra efficiency due to compliance overhead, according to the cited analysis on maturity trade-offs. That is exactly the kind of trade-off clients need help understanding.
For smaller consultancies, this is useful because it supports a more credible recommendation. You don’t have to imply that every client should pursue the highest maturity level available. In many environments, a well-run mid-level programme is the right target.
What a usable roadmap includes
A roadmap shouldn’t read like a rewritten findings appendix. It should look like an operating plan.
Include:
- Current maturity by domain
- Target maturity by domain
- Why the gap matters
- Dependencies
- Owner or owning function
- Evidence expected at reassessment
- Suggested timing
A simple format beats a dense one. If leadership can’t scan it quickly, they won’t use it in budget conversations.
For teams that already think in programme governance terms, this broader view maps neatly onto how organisations assess process capability in adjacent disciplines. The same planning logic appears in discussions of a PMO maturity model, where progress comes from sequencing governance and delivery improvements rather than chasing isolated fixes.
What doesn’t work
Three patterns repeatedly sink maturity roadmaps:
Over-scoping the first phase
If the first quarter includes policy rewrites, tooling replacement, and an incident response overhaul, nothing will land cleanly.Treating all domains equally
Clients don’t need balanced maturity. They need risk-aligned maturity.Ignoring delivery overhead
Recommendations that require heavy process admin without clear operational gain usually stall.
A good roadmap feels achievable. It gives technical teams enough specificity to act and gives leadership enough structure to sponsor the next step.
Integrating Maturity into Pentest Reporting with Vulnsy
The hardest part of maturity-led pentesting isn’t the model. It’s the deliverable.
Most reporting workflows were built to describe exploitation, evidence, impact, and remediation. They were not built to connect repeated findings into capability themes, track those themes over time, and present them in a way that clients can reuse for planning. That’s why many firms talk about maturity in workshops but drop it from the final report.

The fix is to treat maturity as part of the reporting model, not an afterthought added at the end.
Map technical findings to capability domains
A practical approach is to tag each finding to one or more capability areas during report drafting.
Examples:
- Weak password policy maps to IAM
- Missing asset ownership maps to asset management or governance
- Unsupported software and slow remediation map to threat and vulnerability management
- Poor alert triage evidence maps to detection and response
This gives you a second layer of reporting. The finding still stands on its own, but it also contributes to a maturity narrative. Over time, repeated tags show where the client’s operating model is weakest.
That is much more useful than adding a generic “maturity observations” page at the end. The narrative emerges from the evidence already gathered.
Build reusable reporting components
For smaller teams, this only works if the writing overhead stays under control.
A good reporting workflow should let you maintain reusable content for:
- Domain descriptions: Short plain-English explanations of what good looks like in IAM, incident response, vulnerability management, and similar areas.
- Maturity observations: Reusable text that explains the difference between ad hoc, repeatable, and defined practice.
- Roadmap recommendations: Standard recommendations for moving from one maturity state to the next, then tailoring them to the client.
- Evidence expectations: Clear notes on what would demonstrate improvement during a retest or future review.
The most efficient maturity reports are built from tested components, then edited for context. They are not written from scratch every time.
Keep risk assessment and maturity assessment distinct
Many reports often become unclear at this juncture. A pentest still needs to communicate exploitation risk, affected assets, and remediation urgency. Maturity content should support that, not replace it.
A clean report structure usually separates:
| Reporting layer | What it answers |
|---|---|
| Technical findings | What failed and how to fix it |
| Capability observations | Why similar issues keep appearing |
| Maturity assessment | How developed the underlying domain is |
| Improvement roadmap | What the client should do in sequence |
Clients often need help interpreting that distinction. A plain-language explainer like understanding your cyber risk assessment can help frame the difference between immediate risk and the broader work that follows.
Track progress across engagements
Value appears on the second and third engagement.
If your reporting process keeps maturity tags, domain notes, and roadmap recommendations consistent, you can compare current findings to prior capability gaps. Then the report can answer questions clients ask:
- Did IAM maturity improve after the remediation project?
- Are the same control failures appearing in a different form?
- Has the client moved from reactive remediation to a repeatable process?
- Which recommendations still haven’t been operationalised?
That kind of longitudinal reporting is where pentesters start acting more like strategic advisers.
For practitioners who want a simple way to explain the progression itself, a concise reference on maturity model levels is useful background material to align client language before the debrief.
What works and what doesn’t
What works:
- Tagging findings to domains while drafting
- Using standard language for maturity states
- Keeping technical evidence and strategic commentary in the same report
- Tracking domain-level progress across repeat engagements
What doesn’t:
- Adding maturity labels after the report is already finished
- Scoring domains with no evidence trail
- Mixing framework language carelessly
- Producing roadmaps that no smaller client can staff or fund
Done well, maturity-led reporting changes the client relationship. The report stops being a one-off deliverable and becomes a baseline for ongoing improvement. For MSSPs and boutique consultancies, that’s what makes the service durable.
Conclusion From Tactical Fixes to Strategic Resilience
A penetration test still needs to do the fundamentals well. It needs accurate findings, credible evidence, clear impact statements, and remediation advice that engineers can act on. None of that changes.
What changes is the value clients can extract from the work.
When you add a cybersecurity capability maturity model lens, the report stops being just a record of weaknesses found during one engagement. It becomes a picture of how the client operates, where their security capability is thin, and what sequence of improvements makes practical sense. That shift helps clients prioritise better, explain security investment more clearly, and measure whether they are becoming more resilient.
For pentesters, consultancies, and MSSPs, this is also a delivery discipline. The maturity angle only works when it is tied to evidence, scored objectively, and turned into a roadmap the client can follow. It fails when it becomes abstract framework language pasted onto a conventional report.
The strongest practitioners already do part of this instinctively. They notice recurring patterns, weak ownership, inconsistent processes, and controls that exist only on paper. A maturity model gives structure to that judgement. It gives you a language for moving from “here are the bugs” to “here is the operating gap behind them”.
That is a better service for the client and a stronger position for the tester.
If you want to turn pentest findings into cleaner, repeatable, maturity-aware deliverables without drowning in manual formatting, Vulnsy gives solo consultants, consultancies, and MSSPs a practical way to standardise reporting, reuse findings, and produce polished client reports faster.
Written by
Luke Turvey
Security professional at Vulnsy, focused on helping penetration testers deliver better reports with less effort.


