Vulnsy
Guide

PCI Pen Test: Your Guide to DSS 4.0 Compliance & Reporting

By Luke Turvey24 April 202620 min read
PCI Pen Test: Your Guide to DSS 4.0 Compliance & Reporting

If you're preparing for a pci pen test for the first time, the usual pattern is familiar. The audit date is fixed. The cardholder data environment has grown in awkward ways over time. Someone says, “We had a pentest last year,” but no one can tell you whether it covered the actual CDE boundary, the internal paths into it, or the segmentation controls that keep scope manageable.

That’s where teams get into trouble. A PCI penetration test isn’t just a generic security exercise with a compliance label attached. It’s a targeted attempt to prove whether the systems that store, process, or transmit cardholder data can be reached, abused, or pivoted into by a realistic attacker. If the answer is yes, the issue isn’t paperwork. It’s that your controls don’t hold up under pressure.

For solo testers and small consultancies, the hard part usually isn’t only the testing. It’s getting scope right, separating the required test types properly, and producing reporting that will survive audit scrutiny. The testing may take days. The arguments about evidence, methodology, and retest proof can drag on much longer if the report is weak.

Preparing for Your First PCI Pen Test

The first PCI engagement usually feels more administrative than technical at the start. You’re chasing diagrams, trying to confirm where card data flows, and asking basic questions that should have been answered months earlier. Which systems are in scope? Which VLANs are meant to be isolated? Which application talks directly to the payment processor, and which one only touches tokens?

A good pci pen test starts by correcting that mindset. You’re not there to “get Requirement 11.4 done”. You’re there to challenge assumptions before an attacker does. That changes how you plan the work and how you speak to stakeholders.

A security manager might think the goal is a clean report. A junior tester might think the goal is finding high severity issues. In practice, the goal is simpler. Show whether the controls around cardholder data work in the way the organisation believes they work.

PCI testing exposes gaps between documented architecture and real architecture. That gap is where most painful findings live.

When teams approach the engagement as a box-ticking exercise, they rush scoping, flatten technical detail in the report, and treat remediation as a separate problem for later. That approach nearly always creates rework. The better approach is to treat the test, the evidence, and the retest path as one continuous engagement from day one.

Decoding PCI DSS Pen Testing Requirements

The rule set is narrower than many teams expect, but it’s also less forgiving. PCI DSS penetration testing is mandated under Requirement 11.4 of the PCI DSS 4.0 standard, and organisations must test at least annually and after significant changes. The scope covers the full Cardholder Data Environment (CDE), including external perimeters, internal systems, and segmentation controls, with documented methodologies and risk assessments, as set out in the PCI SSC penetration testing guidance.

A professional analyzing digital data visualizations on a transparent screen in a modern server room environment.

What the standard actually expects

The easiest way to explain PCI pen testing to non-security stakeholders is to compare it to building regulations. A drawing may show fire doors and compartment walls, but you still inspect whether the doors shut properly and whether the walls separate the areas they claim to separate. PCI works the same way.

The standard expects more than “an external pentest once a year”. In practice, you need:

  • A documented methodology that maps to recognised approaches such as NIST SP 800-115, OSSTMM, or OWASP.
  • Defined scope documentation that shows what was in scope and why.
  • Testing after significant change so the organisation doesn’t rely on assumptions from an environment that no longer exists.
  • Risk-based output that supports remediation and later retesting.

That last point matters. Auditors don’t only care that testing happened. They care whether the organisation can show what was tested, how it was tested, what was found, and what was done afterwards.

Why PCI is stricter than a generic pentest

A generic pentest may be broad and exploratory. A pci pen test is narrower, but the documentation standard is higher. If a normal infrastructure test is like checking an entire estate for weakness, PCI is checking the fortified room that holds the cash and proving that the surrounding walls really separate it from the rest of the building.

That’s why web applications can’t be ignored just because the project is called a network test. If a public-facing payment application is part of the path into the CDE, it becomes part of the actual attack surface. Teams that need a clearer view of how app-layer work fits into broader assessments often benefit from reading about application security testing alongside PCI-specific requirements.

What auditors usually care about most

The auditors and QSAs I’ve worked with usually focus on three practical questions:

  1. Was the scope defensible
  2. Was the methodology credible
  3. Was remediation verified, not merely promised

If any of those are weak, the engagement becomes a document review problem instead of a security validation exercise.

Practical rule: If you can’t explain why a system is in scope or out of scope in one clear sentence, the scoping work isn’t finished.

Scoping Your PCI Pen Test Correctly

Scoping is where the whole engagement is either saved or ruined. Most bad PCI outcomes don’t start with a clever exploit. They start with an incomplete system list, a stale network diagram, or a segmentation assumption that no one has tested properly.

Think of the CDE as a fortress wall. Your job is to draw that wall accurately. Every system that stores, processes, or transmits cardholder data belongs inside it. Systems that can affect the security of that environment may also pull into scope even if they don’t handle card data directly. If the wall is drawn too narrowly, the test misses real exposure. If it’s drawn too broadly, cost and remediation effort increase fast.

What to collect before testing starts

Before any meaningful testing begins, gather the operational evidence that defines the environment:

  • Network diagrams that show where the CDE sits and how traffic reaches it.
  • Data flow diagrams that show where cardholder data is stored, processed, or transmitted.
  • Component inventories covering servers, applications, APIs, endpoints, and network devices relevant to the CDE.
  • Segmentation design notes explaining what controls are meant to isolate non-CDE systems.

If those artefacts are missing or contradictory, stop and reconcile them. A tester who charges ahead with guessed scope creates a report that looks busy but doesn’t hold up.

The real danger of scope creep

Scope creep in PCI isn’t just an administrative nuisance. It usually means one of two things. Either segmentation isn’t doing what the organisation thinks it’s doing, or no one can prove that it is. Both are bad outcomes.

A common example is a supposedly isolated payment zone that still trusts a management subnet, shares administrative pathways with broader infrastructure, or exposes routes through misconfigured firewall rules. Once that happens, systems assumed to be out of scope may no longer be defensibly excluded.

That drives two consequences:

  • Security impact because lateral movement paths into the CDE may exist
  • Compliance impact because the effective PCI boundary becomes larger than expected

How small teams should define scope

Solo testers and small firms need discipline here because there’s less room for hidden rework. Use a written scope statement that names the environment, attack origins, included applications, and segmentation assumptions. Then test against that statement, not against memory.

A defensible scope statement should answer these questions:

  1. Where does cardholder data enter?
  2. Where can it be stored, even temporarily?
  3. Which internal systems can reach those locations?
  4. Which controls are relied on to keep adjacent systems out of scope?
  5. What changed since the last test?

If the client can’t answer the last question clearly, assume the previous report isn’t enough.

The fastest way to waste time on a pci pen test is to discover halfway through that the “payment environment” actually includes another application stack, another support subnet, and another trust path no one mentioned in kickoff.

Internal External and Segmentation Tests Explained

PCI doesn’t recognise “one pentest” as a useful description. PCI DSS 4.0 mandates three distinct penetration testing types: internal (11.4.2), external (11.4.3), and segmentation (11.4.5). The critical distinction lies in segmentation testing, which validates network isolation to reduce PCI scope. Weak segmentation can remain undetected without this specific test, inadvertently expanding compliance exposure and breach risk, as described in VikingCloud’s PCI penetration testing guide.

PCI Penetration Test Types at a Glance

Test Type Objective Simulates Attacker... Common Targets
Internal Assess what an attacker can do after gaining a foothold inside the network A malicious insider or a threat actor with a compromised workstation Internal applications, directory trust relationships, administrative access paths, file shares, management services
External Assess exposure from internet-facing assets An unauthorised outsider attacking from the public internet Public web applications, remote access services, exposed authentication points, internet-facing hosts
Segmentation Validate that non-CDE systems cannot reach the CDE across supposed isolation boundaries An attacker starting outside the CDE but inside connected business infrastructure Firewall rules, VLAN boundaries, ACLs, routing controls, jump paths into payment systems

External testing asks one question

External testing asks whether someone with no legitimate foothold can break through the public-facing layer. That usually means looking hard at web applications, remote access points, authentication controls, and exposed services. This is the part most non-technical stakeholders imagine when they hear “pentest”.

It’s important, but it’s also the part many organisations overvalue. Plenty of environments have decent perimeter controls and still fail badly once an attacker gets onto an internal system.

Internal testing tells you what happens next

Internal testing assumes the perimeter has already failed. That’s realistic. Attackers don’t always come through a front door. They inherit a VPN session, compromise an endpoint, abuse weak internal authentication, or exploit trust relationships no one thought mattered.

In a PCI context, internal testing often surfaces privilege escalation chains and lateral movement opportunities that put the CDE within reach from places that were assumed to be harmless. Those paths are often more dangerous than a loud external exploit because they align with how real environments are administered.

Segmentation testing is where scope lives or dies

Segmentation testing isn’t a lighter version of an internal test. It has a specific purpose. You are trying to prove that systems outside the CDE cannot cross into it through the isolation controls the organisation relies on.

If you need a concrete example of the kind of weakness this work is meant to uncover, review insufficient network segmentation. That category captures the exact sort of control failure that turns a tidy, limited PCI scope into a much larger exposure.

If a client says, “We reduced scope through segmentation,” the next question is simple. “Show me how you proved it.”

For junior consultants, the key lesson is this. Don’t merge these three activities into one blurred engagement note. Keep the objectives, attack assumptions, evidence, and findings distinct. Auditors and clients both need to see that separation.

The PCI Pen Test Lifecycle and Methodology

A proper pci pen test follows a method, not a mood. A compliant PCI penetration testing methodology follows a multi-phase cycle including discovery, vulnerability identification, and a mandatory re-testing loop. Following remediation based on risk ranking, re-tests are conducted on the specific fixes. This iterative cycle continues until all high-risk vulnerabilities are resolved, according to Scytale’s explanation of PCI penetration testing methodology.

A seven-step flowchart illustrating the PCI penetration testing lifecycle, from initial planning to final remediation and retesting.

The phases that matter in practice

The usual frameworks such as NIST SP 800-115, PTES, OSSTMM, and OWASP give you structure. What matters on a live engagement is how you turn that structure into evidence that can survive technical scrutiny and audit review.

A practical lifecycle looks like this:

  1. Planning and rules of engagement. Confirm in-scope assets, test windows, constraints, points of contact, and what “significant change” has occurred since prior testing.
  2. Reconnaissance and discovery. Identify hosts, services, applications, trust boundaries, and undeclared assets that may force scope reconciliation.
  3. Vulnerability analysis. Combine automated output with manual review. PCI work suffers when teams mistake scanner output for actual attack logic.
  4. Controlled exploitation. Validate exploitability, not just existence. Chain issues where appropriate, but stay within agreed boundaries.
  5. Post-exploitation analysis. Determine whether access enables cardholder data exposure, lateral movement, or administrative compromise.
  6. Reporting. Record findings, impact, methodology, and evidence cleanly enough that another practitioner could follow the narrative.
  7. Remediation and retest. Verify fixes on the affected paths and document the results.

For a broader operational walkthrough, CloudOrbis has a useful guide to penetration testing services that helps frame how structured engagements are typically delivered beyond compliance-only contexts.

Why discovery is more important than many juniors expect

The discovery phase is where experienced testers safeguard the engagement's success. If you find undeclared hosts, forgotten interfaces, or a payment-supporting system that wasn’t in the original asset list, that’s not a side note. It affects scope validity.

Many junior testers want to rush towards exploitation because that feels like the “real work”. In PCI, disciplined discovery often matters more. Testing outside documented rules of engagement can invalidate findings. Failing to reconcile newly discovered in-scope assets can do the same in a different way.

Retesting isn’t an appendix

A lot of teams treat retesting as a small follow-up exercise. PCI doesn’t. It’s part of the lifecycle. A fix that exists only in a change ticket or an engineer’s message isn’t validated. The test has to return to the remediated issue and show whether the original path is gone.

That’s why I advise small teams to read retesting requirements into their working process from the start. If you need a concise breakdown of how the stages fit together operationally, this overview of the phases of penetration testing is a useful reference point.

Good PCI work produces two things at once. A credible attack narrative and a clear remediation ledger.

Mastering PCI Pen Test Reporting and Evidence

The report is where many otherwise competent engagements go weak. The testing may be sound. The screenshots may exist. The consultant may know exactly what happened. But if the final deliverable doesn’t present scope, methodology, findings, and retest evidence in a way a QSA or internal reviewer can follow, the engagement turns into a credibility argument.

That problem is bigger than many firms admit. A 2025 UK ISSA survey of 200 pentesters found that practitioners spend up to 40% of their project time on manual Word formatting, and a 2025 UK ICO report noted that 68% of PCI-related fines stemmed from inadequate documentation, as cited in Secureframe’s article on PCI penetration testing.

A professional holding a tablet displaying business compliance charts and data reports in a modern office environment.

What an audit-ready report needs

An audit-ready PCI report should be readable by more than one audience. Leadership needs a short risk summary. Engineers need exact technical detail. Auditors need traceability. If you optimise for only one of those groups, the document usually fails the others.

At minimum, a strong report contains:

  • Executive summary that explains business impact without flattening technical reality
  • Scope statement covering what was tested, from where, and under which assumptions
  • Methodology section naming the framework and how it was applied
  • Detailed findings with affected assets, exploit path, impact, severity, and remediation guidance
  • Evidence such as screenshots, command output, application responses, and notes that prove what the tester observed
  • Retest section showing which findings were reassessed and whether the original condition still exists

The issue isn’t only completeness. It’s coherence. If the scope says one subnet was tested, the findings and evidence shouldn’t imply another. If the methodology says grey-box testing, the narrative should reflect that level of information access.

Where small teams lose time

Manual reporting usually breaks down in predictable places:

  • Evidence handling because screenshots end up spread across folders and renamed inconsistently
  • Finding reuse because consultants rewrite the same access control or crypto issue repeatedly
  • Formatting drift because one report section uses one style and another uses a different one
  • Retest tracking because the original finding, remediation note, and validation result aren’t tied together cleanly

These are not glamorous problems, but they are expensive ones. They slow delivery, introduce mistakes, and create friction during audit review.

What works better than manual assembly

The fix is standardisation, not generic templates alone. Small teams need a reporting workflow that keeps finding libraries, evidence, and export quality under control without forcing every consultant to become a document production specialist.

A good workflow usually includes:

  1. Predefined report sections aligned to PCI expectations
  2. Reusable finding language that can be adapted without starting from zero
  3. Consistent evidence placement so screenshots and proof-of-concept material appear where reviewers expect them
  4. Retest states that clearly separate open, remediated, and verified conditions

If your current process is still built around stitching screenshots into Word by hand, it’s worth reviewing examples of structured test report templates and comparing them against your current deliverables.

The report should let a reviewer answer four questions quickly. What was tested, how it was tested, what failed, and whether the fix was proven.

Reporting discipline changes technical outcomes

There’s another reason reporting quality matters. Good reporting improves remediation. When engineers get a finding that names the vulnerable path, shows the impact, and includes evidence that aligns to the environment they know, they fix the right thing faster. When they get vague prose and badly cropped screenshots, they patch around the issue and the retest fails.

That’s why experienced PCI practitioners treat documentation as part of the test itself. It isn’t a post-engagement admin task. It’s the evidence trail that proves the work happened properly.

Common Findings and Validating Remediation

Most PCI environments don’t fail because of one dramatic weakness. They fail because ordinary control problems line up in useful ways for an attacker. In the UK, the most common PCI pen test findings are improper access control (CWE-284, 25% of findings) and broken cryptographic algorithms (CWE-327, 30%). The same analysis notes that 18% of tests fail segmentation validation, and QSA firms report a 35% initial failure rate on re-tests due to incomplete fixes, according to Blaze Information Security’s review of PCI DSS penetration testing findings.

A person coding in front of a computer monitor displaying code and a graph visualization.

Findings that appear again and again

Improper access control is common because it hides in normal administration. Shared privileges, weak role separation, over-permissive service access, and legacy trust decisions all create routes an attacker can use once inside. In a PCI environment, those routes matter because they can turn a compromised user context into a path towards systems that should be tightly controlled.

Broken cryptographic implementations are equally persistent. Sometimes the issue is an outdated algorithm. Sometimes it’s poor implementation around key handling or protocol use. The practical consequence is the same. Controls that look compliant on a design diagram fail to protect cardholder data in operation.

Segmentation failures deserve special attention. A client may believe the CDE is isolated because the network was designed that way. The test often shows whether the current rule base, routing, administrative exceptions, or hidden trust paths still support that claim.

Why remediation often fails the first time

Retests fail because teams fix the visible symptom instead of the exploit path. If a finding showed that a management interface was reachable from a non-CDE segment, hardening the interface may not remove the route. If the issue was privilege chaining, resetting one account may not remove the underlying delegation problem.

A good remediation response should include:

  • Control correction that removes the vulnerable condition
  • Path validation that checks adjacent routes, not just the exact exploit used in the report
  • Change evidence such as updated rules, hardened configuration, or corrected application behaviour
  • Retest readiness so the tester can verify the original path efficiently

What a successful retest looks like

A successful retest doesn’t mean “the service is patched”. It means the original attack path no longer works, the control now behaves as intended, and the report records that result clearly.

That distinction matters because PCI is looking for validated remediation, not optimistic status updates.

A passed retest proves more than closure. It proves the organisation understood the failure well enough to fix the actual control, not just the immediate symptom.

For junior consultants, this is one of the most useful habits to build. Write findings so the retest target is obvious. If the original issue can’t be retested cleanly from your evidence and steps, the first report wasn’t precise enough.

Frequently Asked Questions about PCI Pen Testing

Is a vulnerability scan the same as a pci pen test

No. A vulnerability scan identifies known weaknesses, usually through automated checks. A pci pen test goes further and validates whether those weaknesses are exploitable in the specific environment, whether they can be chained, and whether they expose the CDE in practice.

That distinction matters because PCI wants evidence about real attack paths, not only lists of possible issues.

Who should perform a PCI penetration test

The tester needs to be qualified, independent enough for the engagement, and comfortable with PCI scoping, segmentation logic, and audit-grade reporting. In practice, organisations usually want a provider or consultant who understands how QSAs read reports, even if the tester isn’t acting as the assessor.

The wrong choice here is a capable generalist with no PCI discipline around scope and evidence.

How do cloud environments change the engagement

Cloud doesn’t remove PCI obligations. It changes where the boundaries are and who controls them. In AWS or Azure, you still need to identify the systems and services that store, process, or transmit cardholder data, document the boundaries properly, and validate segmentation and access paths around them.

The common mistake is assuming a cloud security group or virtual network design is self-proving. It isn’t. The test still has to validate whether those controls hold.

What happens if a critical issue can’t be fixed immediately

Record it clearly, assess the risk, define compensating measures where appropriate, and make the retest plan explicit. Hiding the issue or describing it vaguely only creates a bigger problem later. A transparent report with clear interim controls is far easier to defend than a weak report that tries to soften the finding.

What should a junior consultant focus on first

Focus on four things:

  • Scope accuracy before tool execution
  • Clean evidence collection while testing, not after
  • Clear attack narratives that explain how a finding was reached
  • Retest-ready writing so remediation can be verified without guesswork

What usually goes wrong in small-team PCI engagements

Small teams tend to under-estimate reporting effort, merge the three required test types together in one loose narrative, and rely too heavily on inherited diagrams. They also leave retest planning until the end, which creates avoidable delays.

The fix is process discipline. Keep scope records current, separate test objectives properly, and build the report as you test rather than after the fact.


If you run PCI engagements regularly, Vulnsy is worth a serious look. It’s built for pentesters and security teams that need audit-ready, brandable reports without losing hours to Word formatting, evidence wrangling, and repetitive copy-paste. For solo consultants, boutique firms, and MSSPs, that means faster delivery, cleaner retest workflows, and more time spent on testing instead of document assembly.

pci pen testpci dss 4.0penetration testingpci compliancesecurity testing
Share:
LT

Written by

Luke Turvey

Security professional at Vulnsy, focused on helping penetration testers deliver better reports with less effort.

Ready to streamline your pentest reporting?

Start your 14-day trial today and see why security teams love Vulnsy.

Start Your Trial — $13

Full access to all features. Cancel anytime.